I think taking logs is an unnecessary indirection in the given proof. If n^m = m^n then raising both sides to the power of 1/(nm) gives us n^(1/n) = m^(1/m). So we are looking for two distinct positive integers at which the function x ↦ x^(1/x) takes the same value. The rest of the proof then goes as before.
Differentiating the above function yields (1/x^2)(1-log(x))x^(1/x), which is positive when log(x) < 1 and negative when log(x) > 1. So the function has a maximum at e and decreases on either side of it. Therefore one of our integers must be less than e, and the other greater than it. For the smaller integer there are only two possibilities, 1 and 2. Using 1 doesn't give a solution since the equation x^(1/x) = 1 only has the solution 1. So the only remaining possibility for the smaller number is 2, which does yield the solution 2^4 = 4^2. Since x^(1/x) is strictly decreasing when x > e, there can't be any other solutions with the same value.
Logs will appear no matter what, at some point in the line of reasoning. They are only held back until differentiating in this approach. I think the expression for the derivative of logn/n was much nicer to grapple with.
Seeing as we're in the integer world, I'm not sure logs need to show up ever, there's probably a proof path around unique prime factorisations. E.g. it's more or less immediately obvious that n and m would need to share the same prime factors in the same ratios.
The key thing you need is that for n >= 3 the sequence of (n^1/n) is decreasing. You can get that pretty easily without logs.
Look at two consecutive terms: n^(1/n) and (n+1)^[1/(n+1)]. The ratio of the first to the second -- we want to prove that this is >1 -- is n^(1/n) / (n+1)^[1/(n+1)]. This is bigger than 1 iff its n(n+1)th power is; that is, iff n^(n+1) / (n+1)^n > 1. We can write this as n (n/(n+1))^n; so what we need is that (n/(n+1))^n > 1/n.
(Handwavily: we know that the LHS is about 1/e, so for n>=3 this should be good. But we want a proof and we're trying to do it without nontrivial analytic machinery.)
Actually, I prefer to write this as ((n+1)/n)^n < n, or (1+1/n)^n < n. Expand this with the binomial theorem: we get sum {0<=k<=n} of (n choose k) n^-k. And we have (n choose k) = n(n-1)...(n-k+1) / k! < n^k/k! so this is strictly less than sum {0<=k<=n} of 1/k!. And for k>0 we easily have k! >= 2^(k-1) so this sum is no bigger than 1 + 1 + 1/2 + 1/4 + 1/8 + etc. = 3.
So, the function on integers n -> n^1/n is decreasing for n >= 3. Now the proof goes through as before.
(Maybe there's a more thoroughly number-theory-ish way to do it by looking at prime factorizations, but when I try it that way it always seems to end up rather a mess.)
[EDITED to add:] But elsewhere in the discussion users bustermellotron and diffeomorphism give (very similar) neat number-theory-ish proofs, either of which is definitely a better proof than the one using the calculations above.
This back-and-forth makes me wonder, is it possible to write proofs about proofs? e.g., you cannot prove this result without logarithms or there must exist a proof that involves prime factors?
I suppose at least some proofs vaguely like that are possible because Gödel's incompleteness theorem is one, although I suspect that that same theorem puts some constraints on these "metaproofs."
Yep, I took several metamathematics classes at UCLA! Mostly on proof theory, but also analyzing metamathematical results (like Godel's theorems, Henkin construction, and so on).
Agreed! I'm realizing that my comment came across as negative, but I did appreciate seeing a second path to the same place. I also agree that the expression x^(1/x) feels like a more natural place to start.
You often see this I think, in "pretty" proofs compared with the more direct approach. A clever early step or some bit of startling insight.
It is a sort of automatic reflex when encountering stuff like $x^x$ in such contexts like here in mathematics to take logarithms. Working with logarithms is usually much simpler and easier to understand when having variables in both the exponent and the base, both for technical reasons (the logarithms will not be avoided anyway, as the other commenter said) and for intuition-gaining reasons. Multiplication of two functions is more intuitively understood (as one can work stuff such as signs, monotonicity etc easily) than exponentiation involving one function in the base and one in the exponent.
Plus this approach (IMO) visualised better than the proposed alternative.
I think it's arguable that OP's approach is cleaner for some value of clean, but the article's approach gave me happy flashbacks to doing number theory in undergrad that OP's approach didn't.
When communicating to the wider world, aesthetics do matter, and yeah, everything you said as well.
Since m and n are distinct, we may assume that m > n >= 2. From the equation and unique factorization, we know that n divides m, so write m = nd.
Then (nd)^n = n^(nd).
Hence d^n = n^{n(d-1)}, which yields d = n^{d-1} >= 2^{d-1}.
Claim: if k is an integer greater than or equal to 3, then k < 2^{k - 1}.
Proof: the base case is clear: 3 < 4. Suppose k > 3 and k - 1 < 2^{k - 2}.
Then k < 2^{k-2} + 1 <= 2^{k-1}, where the last inequality holds because 2^{k-1} - 2^{k-2} = 2^{k-2} >= 1. QED
So, d must be less than 3. Since m = nd and m and n are distinct, d is not 1, so d = 2.
Since d = n^{d-1} and d - 1 = 1, we have n = d, so m = 4.
How exactly do we know that n divides m? I think it's clear that they have the same prime factors, but it's not clear why they couldn't be like n = 12 and m = 18.
Say m contains the prime factor p^i and n contains the prime factor p^j, then m^n will contain p^in and n^m will contain p^jm. For them to be equal we need in = jm or i/j = m/n but for a given solution m/n is constant and the prime factor multiplicity ratio must be equal to that. And that applies to all prime factors, their multiplicities must all have the same constant ratio, therefore the smaller of the two numbers must have smaller multiplicity for all prime factors and therefore divide the larger number. There is probably a simpler way to say this, this feels too complicated.
I think it's even simpler after you figured out that m = kn = n^k => n is an integer > 1 and equal to the (k-1)-th root of k where k is an integer > 1.
The only integer (k-1)-th root of k is 2 for k = 2. Thus, n = 2, k = 2, m = 4.
No, it's a carefully constructed logical argument. Miles ahead of anything GPT is capable of producing, based on my experiments with it.
There's a clear signal that it's constructed by a person with advanced mathematical training: the laconic, "high points only" style. It makes small, but nontrivial and correct, logical leaps, expecting that you will work through the details needed to justify the leap yourself.
But if this were a college homework assignment you would be expected to explain briefly why n divides m, the comment did skip over it which can confuse people.
What drove you to this conclusion? We have an enlightening reply explaining how we know it isn't, but I would also like to know what heuristic you used (so that we know to be wary of it!)
I am not a mathematician, but in the discrete domain of integers:
1) you have two functions, essentially we look for where the two 3d space surfaces intersect, if z is taken as the result for each equation.
2) the integers "are disctinct", so (0,0) and (1,1) are out, plus (2,2) (3,3) etc. Basically a whole linear diagonal in the instersection of both 3d spaces is excluded (why though, to what useful end?)
3) Starting points for the ranges is therefore (0,1) and (1,0)
4) 1^y is always 1, and x^0 is always 1 so there is a constant starting value of 1 both on both axis
5) but x^y will always be larger than y^x, for y>x and x^y > y^x
(prove this by taking the first derivative to get rate of change. Do you use Laplace z domain for discrete, instead of s for continuous?)
6) and the converse to 5.
So once you have found one solution, you know to stop looking, the two surfaces keep diverging from each other.
Why resort to the continuous domain to solve a problem n the discrete, is this even a valid approach?
eg Is there a formal proof that says integers strictly follow that same rules as the continuous domain, just as a subset? I'm interested.
Does this come under group theory, a set with an applied operation?
As I said I am not a mathematician, I am an electrical engineer so probably one of the worst abusers of pure math in a formal sense, but the more I think about this the more questions this raises in my thinking.
> 5) but x^y will always be larger than y^x, for y>x and x^y > y^x (prove this by taking the first derivative to get rate of change. Do you use Laplace z domain for discrete, instead of s for continuous?)
This is false, though! Or, it's almost always true, but there are some exceptions. We of course have an exception at (2,4) and (4,2), where they're equal, and of course at (n,n), where they're obviously equal. And of course 0's and 1's will cause problems for you.
But also, most interestingly, there's an exception at (2,3) and (3,2)! 2<3, and yet, 3^2 > 2^3. Any proof has to account for this!
(There's a fair bit I could say about this exception, but perhaps I should just let you think about it instead. :) )
> Why resort to the continuous domain to solve a problem n the discrete
Because oftentimes this is easier. (In a number of cases it's much easier.) Also... you did this? Like to the extent that your step (5) is valid, you seem to have "proven" it by using the first derivative. That's a continuous tool! I'm not sure what you're talking about with the "Laplace z domain". Or are you using "first derivative" to mean "first difference", or something?
> is this even a valid approach?
Yes, why wouldn't it be? In fact this is actually one of the big reasons for introducing larger number systems, that they let you prove things about the original smaller number system. The rational numbers let you prove things about the integers; the complex numbers let you prove things about the reals; the real numbers and p-padic numbers let you prove things about the rationals; etc. (The integers let you prove things about the whole numbers!)
Proving statements about integers by means of complex numbers is a whole field in itself, i.e., analytic number theory. And in the case of Goodstein's theorem, one famously proves a statement about the whole numbers by passing to the ordinals...
Apropos: difference calculus is a fascinating topic. It has lots of results that are analogous to differential calculus.
Yes, p-adic numbers are also really interesting to look into.
It's also interesting to re-derive much of analysis (like limits and derivatives etc) in the context of the dual numbers (https://en.wikipedia.org/wiki/Dual_number):
> They are expressions of the form a + b * ε, where a and b are real numbers, and ε is a symbol taken to satisfy ε^2 = 0 with ε ≠ 0.
You can sort-of pretend that ε is an infinitesimal, but with a sound theoretical footing.
This is an oversimplification. Nonstandard analysis, the hyperreals, are one way of adding in infinitesimals to the reals, and definitely not the one I'd recommend for all use cases (although going from context, they may be appropriate here).
There are plenty of ways to make a number system that add in infinitesimals to the reals, such as yes the hyperreals, and also I'd count the dual numbers among them, but there's also e.g. the surreal numbers: https://en.wikipedia.org/wiki/Surreal_number
So, why am I kind of down on the hyperreals? Well, thing is, as I understand it, nonstandard analysis isn't really the study of the hyperreals; it's the use of the hyperreals to study the reals. I mentioned above that one big use of passing to a larger number system is that it reflects on the smaller number system; however, as best I can tell, the hyperreals are pretty much used purely in this way. They're used pretty much entirely as a tool for proving statements about the real numbers, rather than an object of study in their own right.
And there's a reason for that; people often talk about "the" hyperreals, but actually, they're not uniquely defined. There's not really the system of hyperreal numbers, so much as there are potentially different systems of hyperreal numbers, which is annoying, but not if you only want to use them as a tool to study the reals, because their (relevant) relation to the reals is all the same. It's a bit icky.
So yeah if you want to do analysis or calculus -- which might be the case, given that the earlier context was dual numbers, and that's what one would typically use dual numbers or -- then sure, use hyperreals. But if you just want to play around with a nifty number system that includes both reals and infinitesimals... eh, they're not great. You're likely to have more fun with the surreals.
(More generally, of course, it's worth remembering that there's no need to stick to well-known systems of numbers... you can invent your own! Like, if for some reason you need infinitesimals, but you don't want them to square to zero like in the dual numbers, but you also don't want all the stuff that's in the hyperreals or surreals, there's nothing wrong with using R[ε] (or R(ε), or other variants depending on exactly what you're doing) to get a sort of minimal reals-with-infinitesimals...)
[Edit: Is there no way to do bold anymore? Those R's in the above paragraph were supposed to be bold, to indicate the real numbers...]
> [if] you also don't want all the stuff that's in the hyperreals or surreals, there's nothing wrong with using R[ε] (or R(ε), or other variants depending on exactly what you're doing) to get a sort of minimal reals-with-infinitesimals...)
> [Edit: Is there no way to do bold anymore? Those R's in the above paragraph were supposed to be bold, to indicate the real numbers...]
I believe you can do bold in Unicode. (see: 𝐛𝐨𝐥𝐝) As far as I know HN has never supported bold as a markup style.
Standard number sets can be done the same way; I would represent the reals as ℝ. My standard method is to go to the wikipedia page for "blackboard bold" and copy the letter I want.
I'd like to know more about the sets you refer to, ℝ[ε] and ℝ(ε), but I don't recognize them. Do they have names I can search for?
> I believe you can do bold in Unicode. (see: 𝐛𝐨𝐥𝐝) As far as I know HN has never supported bold as a markup style.
Oh, I think you're right, I'd forgotten. Grr, HN's limited subset of Markdown is quite annoying sometimes. Well, I'm going to be lazy and just write "R".
> I'd like to know more about the sets you refer to, ℝ[ε] and ℝ(ε), but I don't recognize them. Do they have names I can search for?
No, they don't, that's part of my point -- that you don't need to use recognized systems with names, you can use the usual constructions (or unusual constructions...) to make your own in the way that mathematicians always do. I mean I guess they sort of have names in that the notation would be pretty understandable to most everyone, although pronouncing it is annoying in that they'd both most typically just be pronounced "R adjoin epsilon", although I guess you could say "ring-adjoin" or "field-adjoin" to disambiguate. But note that there's plenty of other variants one could make as well, not just these.
Basically go learn some abstract algebra, is what I would say. Or go read about ring adjunction (and polynomials) or field adjunction (and rational functions).
Btw, I wouldn't describe these as "sets", that's not really the appropriate word to use here. When we talk about systems of numbers, we are, well, talking about systems, or algebraic structures -- rings, fields, ordered rings or fields, topological rings or fields, etc. To say "set" implies that what is important is the elements of these things, the contents; but these elements have no meaning on their own, they're given meaning by the structure -- the permitted operations, relations, etc. (Addition, multiplication, negation, less than or equal to...) Formally, an algebraic structure is a tuple, with the base set being just the first element of that tuple, even if we typically abuse notation and use the same symbol to refer to the set and the structure on that set.
"The Z-transform is a mathematical tool which is used to convert the difference equations in discrete time domain into the algebraic equations in z-domain."
I was taught in engineering math that z-domain is the Laplacian discrete equivelent of the s domain, which is continuous and used by EE as well in analog.
z^-1 (or z(-1) means last sample. It is commonly used in digital signal processing, FFT etc.
s domain is used for (from memory) the operater e^-jw, which is for electrical engineers the transform for use with sine waves, such that impedance is 1/sC for capacitance and sL for inductance.
z domain has useful properties like "The differentiation in z-domain property of Z-transform states that the multiplication by n in time domain corresponds to the differentiation in zdomain."
I am sure I have made some technical mistakes in the above, but it is how I remember it and they don't impact my ability to apply it for my limited EE needs.
As to the question about the valid approach, intuitively I see this, but wondering if there was a formal proof of some kind, or is it taken as given?
Oh, I see! So it's the Fourier transform (in the abstract sense) from the integers to the circle group, except that then you extend to all of C (well, possibly minus the origin). Interesting! Yeah as someone previously unfamiliar with it, it would have been substantially clearer had you referred explcitly to applying the Z-transform, rather than switching domains, which is the sort of thing that will only make sense to someone who already knows about this.
So wait does that mean you weren't actually sure if this approach to that step would work? I assumed you were saying you had a proof, not just outlining an approach you thought would work. (I mean, obviously you didn't have a proof of the whole thing as the overall statement is false, but individual steps might have worked.)
But I have to note -- even if the proof works, then if you're applying the Z-transform, then you are once again not sticking to the realm of the discrete! Complex numbers are a continuum matter. So that approach still doesn't yield an integers-only proof!
> As to the question about the valid approach, intuitively I see this, but wondering if there was a formal proof of some kind, or is it taken as given?
I'm not really sure how to answer this -- what would a formal proof here even consist of? It's easy enough to do it in any instance, but the problem is, how would you even formally state the general principle?
Like you could do large classes of statements, certainly. So for instance, if what we're doing is purely algebraic, then you could say, if A and B are algebraic structures, and i:A->B is an injective homomorphism, and S is a set of algebraic equations all of which are always in the image of i, then all of them are always true in A; but of course there's way more types of statements one can make than algebraic equations.
So, uh, yeah, one can write down any number of statements like that, but I don't know how you'd formally abstract it into a general principle...?
Or, to say a bit more about this -- I guess the basic principle here is "it's OK to use auxiliaries?" Introducing a larger number system is no different from drawing an auxiliary line in geometry (for instance). There's no rule that the proof of a statement may only contain the entities appearing in the statement! You can introduce whatever auxiliaries you like.
A subset won’t have more solutions than its containing set; but it may have fewer.
Many problems are easier to solve in the reals (due to being complete), and you can then restrict that solution to your (sub)set of interest — in this case, the integers.
You see the same thing with Pythagorean triples being simpler to solve by doing the math over the complex numbers and then restricting your answers.
I'm trying to understand 5). If you're claiming that both (A) y > x, and (B) x^y > y^x hold, then x^y > y^x ("will always be larger") holds. You're satisfying your claim by assumption. Nothing new is deduced.
However, if only A) needs to be satisfied: y = 3, x = 2 is a counterexample, as x^y = 2^3 = 8 < 9 = 3^2 = y^x.
edit: looks like someone had the same thought as me as I was typing my reply!
I am not a mathematician either, I always maintained the university made a mistake granting make a math teacher degree :) It was also long ago enough a lot of you weren't even alive and I haven't done any such work in a quarter century. Nonetheless...
> but x^y will always be larger than y^x, for y>x and x^y > y^x (prove this by taking the first derivative to get rate of change. Do you use Laplace z domain for discrete, instead of s for continuous?)
I am struggling to follow what are you saying here
>
Why resort to the continuous domain to solve a problem n the discrete, is this even a valid approach?
I can't make heads or tails of this question. The proof says, correctly, for all (real) x != y, x < y solutions it is true that 0<x<e and e<y. He found this by investigating the derivative and establishing the monotonically increasing / decreasing nature of the function. Only after finding this out using does he go back to the original question: what integer x could be? Since he restricted x to be 0 < x < e , we only need to investigate the case of 1 and 2.
Yes, m=n is a solution, or set of solutions, but why arbitrarily exclude some solutions and not others?
There is no explanation as to why - eg maybe it could be if the equations represent a physical system and m=n implied the same physical space was occupied by two objects, for want of a better example.
But if the equations represented some kind of abstract concept, why are not all solutions valid and of interest?
It seems like me saying, find an equation that only yields all the primes that use any digit once - but why, to what end and value?
For today’s interesting problem, we are looking to find all positive integer pairs n and m which satisfy the equation above, when n and m are distinct. If we play around with small n and m we can quickly see that 2⁴ = 16 = 4² , so the pair 2 and 4 are certainly one solution. Actually, it is the only solution.
I noticed this interesting fact during my secondary school education. Additionally you can use this to prove that e is the point where a change in the exponent causes a larger result than a change in the base. E.g. (x+d)^y > x^(y+d) when y >= e for any real value d.
Alternative proof without any real numbers, logs or the like:
After relabeling we may assume n<m.
Step 1: We claim that there exists a natural number k such that m=kn.
Write the equation as
n^n n^(m-n) = m^n
n^(m-n) = (m/n)^n
The left-hand-side is always an integer. The right-hand-side is an integer if and only if such a factor k exists.
Step 2: Inserting this ansatz, we may take n-th roots and reduce the problem to n^k =k n.
Step 3: Monotonicity. Clearly we have equality for k=1.
We claim that the difference n^k-kn strictly increases with k for fixed n unless n=2, k=1. That claim in particular gives the result.
We prove this by induction in k. That is, suppose that n^k\geq kn holds for all k\leq K, then
n^(K+1) = n n^K \geq n K n \geq Kn +Kn \geq Kn +n = (K+1)n.
The second inequality is strict unless n=2 and the last one is strict unless K=1.
(m/n)^n is only an integer if m/n is an integer: true but this is quite a big piece of mathematics on its own.
The proof that no rational number squares to 2 is a famous result of number theory. That no rational non-integer, raised to any power, is any integer, surely has to be justified here.
True, but considering that the linked post uses e, log, limits and derivatives without any further comment, I am going to err on the side of "nah, enough detail".
I was interested in the proof of this ("if x is a rational number, and x raised to the power of an integer is an integer, then x is an integer"), so I asked the new Bing. It actually gave a correct proof (despite no relevant links found in Google)!
------BELOW IS BING'S OUTPUT
This statement is true. If x is a rational number and x raised to the power of an integer is an integer, then x is an integer. This statement is called the rational root theorem. The proof of this theorem is based on the fact that if x is not an integer, then it can be expressed as a fraction p/q where p and q are integers and q ≠ 0. Then, x raised to the power of an integer can be expressed as (p/q)^n which can be simplified to p^n/q^n. Since p^n and q^n are integers, p^n/q^n is a rational number. If x raised to the power of an integer is an integer, then q^n must divide p^n. Since p and q are relatively prime, this implies that q = 1. Therefore, x is an integer.
> It actually gave a correct proof (despite no relevant links found in Google)!
That is just wrong. It is the completely standard proof and can be found everywhere (the proof is over 2000 years old). Add to that, that it forgets to choose p and q as relatively prime, but then uses that at the end.... Not impressive.
Nah, k=1 is the trivial base case and is allowed. The "assumption" is no such thing, but just notational convenience. One can of course tidy that up (e.g. n\leq m and trivial case is obvious), but for a random HN post it seems not worth the effort.
Well, I guess you're right. I mean, we're looking for two disctinct numbers as the final solution, but that doesn't mean we can't use k=1 in the proof.
Showing the graph of the functions without showing that there are only two possible integers to work with (since we're looking for solutions in ℕ, not ℝ) on the left side of the maximum feels like missing the most important part of the drawing things out. You can even leave the y axis unlabeled, but label the x-axis (especially since we've already determined f(1)=0) and draw some dots on the curve to show that those are the only values that can be in our solution space.
"There's a maximum at e, so the only two possible values on the left of e that can be used in these pairs are 1, and 2. But really there's only one, because 1 is the power identity, so we can't use that. If there is a solution, it has to use 2."
Which means we're solving one function for one unknown, in the natural numbers. 2ⁿ=n² gives n=4. We have found the only pair of integers that satisfies our constraint function.
Why was "Distinct Positive" left out of the submission title? The submission smells like clickbait without it, yet it is part of the actual article title and the article appears to be submitted by the author.
m = n would be trivial and pointless. Also, I'm curious why he didn't specify precisely clarify "distinct" to mean "pairs of unequal values" (e.g., exclude {n,m} where n=m) and "pairs not equal to others" (e.g., {x,y} excludes {y,x}).
The condition is as precise as possible "pairs of distinct integers" is not equal to "distinct pairs of integers". Don't blame the writer when you fail to read what is written.
But if you make it a sentence, "my pair of integers is 1 and 1", it sounds a little bit off. Because "pair" for the same concept twice is often an error.
Picking a mathematical formulation like you did can avoid that kind of implication, but there wasn't a template for which formulation to use and the way you wrote that makes it look like order matters which isn't right either.
I like problems like these. I was posed one at the end of a 'Constructive Mathematics' course I took at Stanford. Eight and nine are right next to each other (+-1). One is a cube, and the other is a square.
Do a cube and a square land next to each other ever again?
Note it's not asking if they land on each other, that is trivial - 1000000 is 100^3 and 1000^2.
My solution was to form an equation, divide by x and solve the resulting quadratic equation. It produce new solutions: 0 and 1, 0 and -1! As well as 2 and 3. So I was encouraged to say no, those were the only solutions.
I was following until the very end. This felt very handwavy:
>Now, since e is somewhere between 2 and 3, we have to conclude that n must be 2. Further, from our curve, we can see that there can only be one other value of x for which (lnx)/x = (ln2)/2, and since we know that 2⁴ = 4², we know that the other integer must be 4.
Is this sort of thinking something that could be justifiably written in a real proof? Sure, conveniently fishing out 2 as n seems to narrow down the search space for m, but I'd imagine this wouldn't work for harder problems
Really fun article to read, and I'm not even that interested in math.
Somewhat unrelated, but I cant help but ask why authors are still using Medium as opposed to Substack or some other alternative. The hook of the post happened to interest me to the extent where I went looking for an un-auth-walled version, but I'm sure there have been countless cases where I have simply abandoned it because I wasn't invested enough to seek out the mirrors -- and I'm sure no author wants that.
Similar pathway for classic high school problem I really enjoyed about finding which one is bigger e^pi or pi^e. I was unreasonably pleased with that problem. Won't spoil.
I'm not formally trained Mathematician, but I think this is not an analytic proof - it lacks rigour. It is more a graphical approach that appeals to the reader's intuition. I could be wrong here though.
Second, I really really thought the last graph will include y = x line and was very surprised to not see it :)
No, the graph parts are explanation. The analysis of the derivative function maybe would be better if it were rewritten with the usual limes notations, nonetheless it's correct in the crucial conclusion of the ln x / x function is monotonically increasing (0, e) and monotonically decreasing (e, inf) and as such one of every pair of equal but distinct values must be from (0,e) the other from (e,inf). There are not many integers between 0 and e: 1 and 2. At 1, ln x / x is zero and in (e,inf) this function has a strict lower boundary of 0, it is never 0. Thus the only possible solution is 2 and it is.
In general, couldn't a function be monotonically increasing in some interval and still have many points such that y = x? A curve that jiggles around the line y = x for instance.
So while the original problem was x^y=y^x he have rearranged this into ln x / x = ln y / y and so now are investigating the ln x / x function. Yes, the lettering could be clearer because y at some points means one variable and later on it means the value of the function.
Yes, that's basically what I said and meant in the parent post. I was suggesting the article call it a "graphical" solution approach, not analytical.
And if I wasn't clear - it wasn't my intent to treat this any less "useful". If anything, it is more interesting (for me) to read about solutions that give an intuitive feel, than dry analysis.
I would have approached this problem this way: set a lower bound say 3. Then we wish to prove there doesn't exist any n,m>=3 that satisfies n^m=m^n. This is much more easily proved using just double induction. Then you can just search for a case where n=2.
This doesn't prove anything, unfortunately. It doesn't even make use of the fact that m,n are integers. If m,n are not restricted to integers, there exists an infinite number of solutions (e.g. m=8.043, n=1.46).
The mistake in your proof is that you considered the shape of the log function while implicitly holding m on the left-hand-side constant (i.e. you should have concluded "for a constant m", there is only 1 n satisfying this equality"). However, since m is variable, you have to consider an infinite number of log functions. If I missed something, let me know.
This was one of our bonus homework problems in Höhere Mathematik in engineering school in my misspent youth. I tried the factorization but got stuck. The offical solution was the log one shown in the article, somehow not as satisfying.
fake quote> Anyone that took a Calculus course knows how to analyze this function and get the minimum, maximum and increasing and decreasing intervals. The calculation is unsurprising and boring, so I will skip it and use a graph instead.
fake quote> Anyone that didn't take a Calculus course will not understand the technical details. The calculation is long enough to be distracting and boring, so I will skip it and use a graph instead.
Unless the main public of the blog are students from the first year of the university, I agree with the author that it's better to use a graph and left the calculations as an exercise.
[Perhaps the analytical calculation could have been a note at the bottom, but it's long and unsurprising enough to be boring to write it clearly and carefully. Just left it as an exercise :) .]
A hand-drawn graph can be a great tool for reasoning, because it only depicts the features we choose — everything else is imprecise, and it looks imprecise. For me, that's actually a better way of building a visual intuition for how a function works.
What you say can be true, but it sets a certain bar for clearly conveying the message artistically. I would say that bar was not met in this case. I found it so distracting I was compelled to go off to Desmos to draw it myself.
Differentiating the above function yields (1/x^2)(1-log(x))x^(1/x), which is positive when log(x) < 1 and negative when log(x) > 1. So the function has a maximum at e and decreases on either side of it. Therefore one of our integers must be less than e, and the other greater than it. For the smaller integer there are only two possibilities, 1 and 2. Using 1 doesn't give a solution since the equation x^(1/x) = 1 only has the solution 1. So the only remaining possibility for the smaller number is 2, which does yield the solution 2^4 = 4^2. Since x^(1/x) is strictly decreasing when x > e, there can't be any other solutions with the same value.