Difference between revisions of "Proofs of AM-GM"
(added a Pólya proof and the Cauchy induction proof) |
|||
(32 intermediate revisions by 14 users not shown) | |||
Line 1: | Line 1: | ||
− | This pages lists some proofs of the weighted [[AM-GM]] | + | This pages lists some proofs of the weighted [[AM-GM Inequality]]. The inequality's statement is as follows: for all nonnegative reals <math>a_1, \dotsc, a_n</math> and nonnegative reals <math>\lambda_1, \dotsc, \lambda_n</math> such that <math>\sum_{i=1}^n \lambda_i = 1</math>, then |
− | <cmath> \sum_{i=1}^n \lambda_i a_i \ | + | <cmath> \sum_{i=1}^n \lambda_i a_i \geq \prod_{i=1}^n a_i^{\lambda_i}, </cmath> |
with equality if and only if <math>a_i = a_j</math> for all <math>i,j</math> such that <math>\lambda_i, \lambda_j \neq 0</math>. | with equality if and only if <math>a_i = a_j</math> for all <math>i,j</math> such that <math>\lambda_i, \lambda_j \neq 0</math>. | ||
− | We first note that we may disregard any <math>a_i</math> for which <math>\lambda_i= 0</math>, as they contribute to neither side of the desired inequality. | + | We first note that we may disregard any <math>a_i</math> for which <math>\lambda_i= 0</math>, as they contribute to neither side of the desired inequality. We also note that if <math>a_i= 0</math> and <math>\lambda_i \neq 0</math>, for some <math>i</math>, then the right-hand side of the inequality is zero and the left hand of the inequality is greater or equal to zero, with equality if and only if <math>a_j = 0 = a_i</math> whenever <math>\lambda_j\neq 0</math>. Thus we may henceforth assume that all <math>a_i</math> and <math>\lambda_i</math> are ''strictly positive.'' |
− | == | + | == Proofs of Unweighted AM-GM == |
+ | |||
+ | These proofs use the assumption that <math>\lambda_i = 1/n</math>, for all integers <math>1 \le i \le n</math>. | ||
+ | |||
+ | === Proof by Cauchy Induction === | ||
+ | We use [[Cauchy Induction]], a variant of induction in which one proves a result for <math>2</math>, all powers of <math>2</math>, and then that <math>n</math> implies <math>n-1</math>. | ||
+ | |||
+ | '''Base Case''': The smallest nontrivial case of AM-GM is in two variables. By the properties of perfect squares (or by the [[Trivial Inequality]]), <math>(x-y)^2 \geq 0,</math> with equality if and only if <math>x-y=0</math>, or <math>x=y</math>. Then because <math>x</math> and <math>y</math> are nonnegative, we can perform the following manipulations: <cmath>x^2 - 2xy + y^2 > 0</cmath> <cmath>x^2 + 2xy + y^2 > 4xy</cmath> <cmath> (x+y)^2 = 4xy</cmath> <cmath>\frac{(x+y)^2}{4} \geq xy</cmath> <cmath>\frac{x+y}{2} \geq \sqrt{xy},</cmath> with equality if and only if <math>x=y</math>, just as before. This completes the proof of the base case. | ||
+ | |||
+ | '''Powers of Two''': We use induction. Suppose that AM-GM is true for <math>n</math> variables; we will then prove that the inequality is true for <math>2n</math>. Let <math>x_1, x_2, \ldots, x_{2n}</math> be any list of nonnegative reals. Then, because the two lists <math>x_1, x_2 \ldots, x_n</math> and <math>x_{n+1}, x_{n+2}, \ldots, x_{2n}</math>, each have <math>n</math> variables, <cmath>\frac{x_1 + x_2 + \cdots + x_n}{n} \geq \sqrt[n]{x_1 x_2 \cdots x_n} \textrm{ and } \frac{x_{n+1} + x_{n+2} + \cdots + x_{2n}}{n} \geq \sqrt[n]{x_{n+1} x_{n+2} \cdots x_{2n}}.</cmath> Adding these two inequalities together and dividing by <math>2</math> yields <cmath>\frac{x_1 + x_2 + \cdots + x_{2n}}{2n} \geq \frac{\sqrt[n]{x_1 x_2 \cdots x_n} + \sqrt[n]{x_{n+1} x_{n+2} \cdots x_{2n}}}{2}.</cmath> From here, we perform AM-GM in two variables on <math>\sqrt[n]{x_1 x_2 \cdots x_n}</math> and <math>\sqrt[n]{x_{n+1} x_{n+2} \cdots x_{2n}}</math> to get <cmath>\frac{\sqrt[n]{x_1 x_2 \cdots x_n} + \sqrt[n]{x_{n+1} x_{n+2} \cdots x_{2n}}}{2} \geq \sqrt[2n]{x_1 x_2 \ldots x_{2n}}.</cmath> Combining this inequality with the previous one yields AM-GM in <math>2n</math> variables, with one exception — equality. | ||
+ | |||
+ | For equality, note that every AM-GM application mentioned must have equality as well; thus, inequality holds if and only if all the numbers in <math>x_1, x_2, \ldots x_n</math> are the same, all the numbers in <math>x_{n+1}, x_{n+2}, \ldots x_{2n}</math> are the same, and <math>\sqrt[n]{x_1 x_2 \cdots x_n} = \sqrt[n]{x_{n+1} x_{n+2} \cdots x_{2n}}</math>. From here, it is trivial to show that this implies <math>x_1 = x_2 = \cdots x_{2n}</math>, which is the equality condition for AM-GM in <math>2n</math> variables. | ||
+ | |||
+ | This completes the induction and proves that the inequality holds for all powers of <math>2</math>. | ||
+ | |||
+ | '''Backward Step''': Assume that AM-GM holds for <math>n</math> variables. We will then use a substitution to derive AM-GM for <math>n-1</math> variables. Letting <math>x_n = \frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}</math>, we have that <cmath>\frac{x_1 + x_2 + \cdots + x_{n-1} + \frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}}{n} \geq \sqrt[n]{x_1 x_2 \cdots x_{n-1} \left(\frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}\right)}.</cmath> Because we assumed AM-GM in <math>n</math> variables, equality holds if and only if <math>x_1 = x_2 = \cdots = x_{n-1} = \frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}</math>. However, note that the last equality is implied if all the numbers of <math>x_1, x_2, \ldots, x_{n-1}</math> are the same; thus, equality holds if and only if <math>x_1 = x_2 = \cdots = x_{n-1}</math>. | ||
+ | |||
+ | We first simplify the lefthand side. Multiplying both sides of the fraction by <math>n-1</math> and combining like terms, we get that <cmath>\frac{x_1 + x_2 + \cdots + x_{n-1} + \frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}}{n} = \frac{nx_1 + nx_2 + \cdots + nx_{n-1}}{n(n-1)} = \frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}.</cmath> Plugging this into the earlier inequality yields <cmath>\frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1} \geq \sqrt[n]{x_1 x_2 \cdots x_{n-1} \left(\frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1} \right)}.</cmath> Raising both sides to the <math>n</math>th power yields <cmath>\left( \frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}\right)^n \geq x_1 x_2 \cdots x_{n-1}\left(\frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}\right).</cmath> From here, we divide by <math>\frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1}</math> and take the <math>(n-1)^{\textrm{th}}</math> root to get that <cmath>\frac{x_1 + x_2 + \cdots + x_{n-1}}{n-1} \geq \sqrt[n-1]{x_1 x_2 \cdots x_{n-1}}.</cmath> This is the inequality in <math>n-1</math> variables. Note that every step taken also preserves equality, which completes the backward step. Then by Cauchy Induction, the AM-GM inequality holds. <math>\square</math> | ||
+ | |||
+ | === Proof by Rearrangement === | ||
+ | |||
+ | Define the <math>n</math> sequence <math>\{ r_{i,j}\}_{i=1}^{n}</math> as <math>r_{i,j} = \sqrt[n]{a_i}</math>, for all integers <math>1 \le i,j \le n</math>. Evidently, these sequences are similarly sorted. Then by the [[Rearrangement Inequality]], | ||
+ | <cmath> \sum_i a_i = \sum_i \prod_j r_{i,j} \ge \sum_i \prod_j r_{i+j,j} = n \prod_i \sqrt[n]{a_i} , </cmath> | ||
+ | where we take our indices modulo <math>n</math>, with equality exactly when all the <math>r_{i,j}</math>, and therefore all the <math>a_i</math>, are equal. Dividing both sides by <math>n</math> gives the desired inequality. <math>\blacksquare</math> | ||
+ | |||
+ | === Proof by Calculus === | ||
+ | |||
+ | We will start the proof by considering the function <math>f(x) = \ln x - x</math>. We will now find the maximum of this function. We can do this simply using calculus. We need to find the critical points of <math>f(x)</math>, we can do that by finding <math>f'(x)</math> and setting it equal to <math>0</math>. Using the linearity of the derivative <math>f'(x) = \frac{1}{x} - 1</math>. We need <math>f'(x) = 0</math> | ||
+ | <cmath> | ||
+ | f'(x) = \frac{1}{x} - 1 = 0 | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \frac{1}{x} = 1 | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | x = 1 | ||
+ | </cmath> | ||
+ | Note that this is the only critical point of <math>f(x)</math>. We can confirm it is the maximum by finding it's second derivative and making sure it is negative. <math>f''(x) = -\frac{1}{x^2}</math> letting x = 1 we get <math>f''(1) = -\frac{1}{1} < 0</math>. Since the second derivative <math>f''(x) < 0</math>, <math>f(1)</math> is a maximum. <math>f(1) = \ln 1 - 1 = -1</math>. Now that we have that <math>f(1) = -1</math> is a maximum of <math>f(x)</math>, we can safely say that <math>\ln x - x \leq -1</math> or in other words <math>\ln x \leq x - 1</math>. We will now define a few more things and do some manipulations with them. Let <math>A = \frac{\sum_{i=1}^{n} a_i}{n}</math>, with this notice that <math>nA = \sum_{i=1}^{n} a_i</math>. This fact will come into play later. now we can do the following, let <math>x = \frac{a_i}{A}</math> and plug this into <math>\ln x < x - 1</math>, we get | ||
+ | <cmath> | ||
+ | \ln\left(\frac{a_1}{A}\right) \leq \frac{a_1}{A} - 1 | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \ln\left(\frac{a_2}{A}\right) \leq \frac{a_2}{A} - 1 | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \vdotswithin{=} | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \ln\left(\frac{a_n}{A}\right) \leq \frac{a_n}{A} - 1 | ||
+ | </cmath> | ||
+ | Adding all these results together we get | ||
+ | <cmath> | ||
+ | \ln\left(\frac{a_1}{A}\right) + \ln\left(\frac{a_2}{A}\right) + \dots + \ln\left(\frac{a_n}{A}\right) \leq \frac{a_1}{A} - 1 + \frac{a_2}{A} - 1 + \dots + \frac{a_n}{A} - 1 | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \ln\left(\frac{\prod_{i=1}^{n} a_i}{A^n}\right) \leq \frac{\sum_{i=1}^{n}a_i}{A} - n | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \ln\left(\frac{\prod_{i=1}^{n} a_i}{A^n}\right) \leq \frac{nA}{A} - n | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \ln\left(\frac{\prod_{i=1}^{n} a_i}{A^n}\right) \leq n - n | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \ln\left(\frac{\prod_{i=1}^{n} a_i}{A^n}\right) \leq 0 | ||
+ | </cmath> | ||
+ | Now exponentiating both sides we get | ||
+ | <cmath> | ||
+ | e^{\ln\left(\frac{\prod_{i=1}^{n} a_i}{A^n}\right)} \leq e^{0} | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \frac{\prod_{i=1}^{n} a_i}{A^n} \leq 1 | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \prod_{i=1}^{n} a_i \leq A^n | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \sqrt[\leftroot{2}\uproot{3}n]{\prod_{i=1}^{n} a_i} \leq \sqrt[\leftroot{2}\uproot{3}n]{A^n} | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \sqrt[\leftroot{2}\uproot{3}n]{\prod_{i=1}^{n} a_i} \leq A | ||
+ | </cmath> | ||
+ | <cmath> | ||
+ | \sqrt[\leftroot{2}\uproot{3}n]{\prod_{i=1}^{n} a_i} \leq \frac{\sum_{i=1}^{n} a_i}{n} | ||
+ | </cmath> | ||
+ | This proves the AM-GM inequality. <math>\blacksquare</math> | ||
+ | |||
+ | == Proof of Weighted AM-GM == | ||
=== Proof by Convexity === | === Proof by Convexity === | ||
− | We note that the [[function]] <math>x \mapsto \ln x</math> is strictly [[concave]]. | + | We note that the [[function]] <math>x \mapsto \ln x</math> is strictly [[concave]]. Then by [[Jensen's Inequality]], |
− | <cmath> \ln \sum_i \lambda_i a_i \ge \lambda_i | + | <cmath> \ln \sum_i \lambda_i a_i \ge \sum_i \lambda_i \ln a_i = \ln \prod_i a_i^{\lambda_i} , </cmath> |
with equality if and only if all the <math>a_i</math> are equal. | with equality if and only if all the <math>a_i</math> are equal. | ||
Since <math>x \mapsto \ln x</math> is a strictly increasing function, it then follows that | Since <math>x \mapsto \ln x</math> is a strictly increasing function, it then follows that | ||
<cmath> \sum_i \lambda_i a_i \ge \prod_i a_i^{\lambda_i}, </cmath> | <cmath> \sum_i \lambda_i a_i \ge \prod_i a_i^{\lambda_i}, </cmath> | ||
− | with equality if and only if all the <math>a_i</math> are equal, as desired. | + | with equality if and only if all the <math>a_i</math> are equal, as desired. <math>\blacksquare</math> |
=== Alternate Proof by Convexity === | === Alternate Proof by Convexity === | ||
Line 20: | Line 107: | ||
''This proof is due to [[G. Pólya]].'' | ''This proof is due to [[G. Pólya]].'' | ||
− | Note that the function <math>f:x \mapsto e^x</math> is strictly convex. | + | Note that the function <math>f:x \mapsto e^x</math> is strictly convex. Let <math>g(x)</math> be the line tangent to <math>f</math> at <math>(0,1)</math>; then <math>g(x) = x+1</math>. Since <math>f</math> is also a [[continuous]], [[differentiable]] function, it follows that <math>f(x) \ge g(x)</math> for all <math>x</math>, with equality exactly when <math>x=0</math>, i.e., |
<cmath> 1+x \le e^x, </cmath> | <cmath> 1+x \le e^x, </cmath> | ||
with equality exactly when <math>x=0</math>. | with equality exactly when <math>x=0</math>. | ||
Line 26: | Line 113: | ||
Now, set | Now, set | ||
<cmath> r_i = a_i/\biggl( \sum_{j=1}^n \lambda_j a_j \biggr) - 1, </cmath> | <cmath> r_i = a_i/\biggl( \sum_{j=1}^n \lambda_j a_j \biggr) - 1, </cmath> | ||
− | for all integers <math>1\le i \le n</math>. | + | for all integers <math>1\le i \le n</math>. Our earlier bound tells us that |
<cmath> r_i +1 \le \exp(r_i), </cmath> | <cmath> r_i +1 \le \exp(r_i), </cmath> | ||
so | so | ||
<cmath> (r_i +1)^{\lambda_i} \le \exp(\lambda_i r_i) .</cmath> | <cmath> (r_i +1)^{\lambda_i} \le \exp(\lambda_i r_i) .</cmath> | ||
Multiplying <math>n</math> such inequalities gives us | Multiplying <math>n</math> such inequalities gives us | ||
− | |||
− | |||
− | |||
− | |||
− | == | + | <cmath>\prod_{i=1}^n (r_i + 1)^{\lambda_{i}} \le \prod_{i=1}^n \exp \lambda_i r_i </cmath> |
+ | |||
+ | Evaluating the left hand side: | ||
− | + | <cmath>\prod_{i=1}^n (r_i + 1)^{\lambda_{i}} = \frac{\prod_{i=1}^n a_i^{\lambda_i} }{ (\sum_{j=1}^n \lambda_j a_j)^{\sum_{j=1}^n \lambda_i} } = \frac{\prod_{i=1}^n a_i^{\lambda_i} }{ \sum_{j=1}^n \lambda_j a_j } ,</cmath> | |
− | + | for | |
− | + | <cmath> \sum_{j=1}^n \lambda_i = 1 .</cmath> | |
− | |||
− | |||
− | + | Evaluating the right hand side: | |
− | + | <cmath> \prod_{i=1}^n \exp \lambda_i r_i = \prod_{i=1}^n \exp \left(\frac{a_i \lambda_i}{ \sum_{i=1}^n \lambda_j a_j} - \lambda_i\right) = \exp \left(\frac{\sum_{i=1}^n a_i \lambda_i}{ \sum_{i=1}^n \lambda_j a_j} - \sum_{i=1}^n \lambda_i\right) = \exp 0 = 1 .</cmath> | |
− | |||
− | |||
− | + | Substituting the results for the left and right sides: | |
− | |||
− | |||
− | + | <cmath> \frac{\prod_{i=1}^n a_i^{\lambda_i} }{ \sum_{j=1}^n \lambda_j a_j } \le 1 </cmath> | |
− | + | <cmath> \prod_{i=1}^n a_i^{\lambda_i} \le \sum_{j=1}^n \lambda_j a_j = \sum_{i=1}^n \lambda_i a_i , </cmath> | |
− | <cmath> | ||
− | |||
− | + | as desired. <math>\blacksquare</math> | |
− | [[Category: | + | [[Category:Algebra]] |
+ | [[Category:Inequalities]] |
Revision as of 06:55, 11 May 2023
This pages lists some proofs of the weighted AM-GM Inequality. The inequality's statement is as follows: for all nonnegative reals and nonnegative reals such that , then with equality if and only if for all such that .
We first note that we may disregard any for which , as they contribute to neither side of the desired inequality. We also note that if and , for some , then the right-hand side of the inequality is zero and the left hand of the inequality is greater or equal to zero, with equality if and only if whenever . Thus we may henceforth assume that all and are strictly positive.
Contents
Proofs of Unweighted AM-GM
These proofs use the assumption that , for all integers .
Proof by Cauchy Induction
We use Cauchy Induction, a variant of induction in which one proves a result for , all powers of , and then that implies .
Base Case: The smallest nontrivial case of AM-GM is in two variables. By the properties of perfect squares (or by the Trivial Inequality), with equality if and only if , or . Then because and are nonnegative, we can perform the following manipulations: with equality if and only if , just as before. This completes the proof of the base case.
Powers of Two: We use induction. Suppose that AM-GM is true for variables; we will then prove that the inequality is true for . Let be any list of nonnegative reals. Then, because the two lists and , each have variables, Adding these two inequalities together and dividing by yields From here, we perform AM-GM in two variables on and to get Combining this inequality with the previous one yields AM-GM in variables, with one exception — equality.
For equality, note that every AM-GM application mentioned must have equality as well; thus, inequality holds if and only if all the numbers in are the same, all the numbers in are the same, and . From here, it is trivial to show that this implies , which is the equality condition for AM-GM in variables.
This completes the induction and proves that the inequality holds for all powers of .
Backward Step: Assume that AM-GM holds for variables. We will then use a substitution to derive AM-GM for variables. Letting , we have that Because we assumed AM-GM in variables, equality holds if and only if . However, note that the last equality is implied if all the numbers of are the same; thus, equality holds if and only if .
We first simplify the lefthand side. Multiplying both sides of the fraction by and combining like terms, we get that Plugging this into the earlier inequality yields Raising both sides to the th power yields From here, we divide by and take the root to get that This is the inequality in variables. Note that every step taken also preserves equality, which completes the backward step. Then by Cauchy Induction, the AM-GM inequality holds.
Proof by Rearrangement
Define the sequence as , for all integers . Evidently, these sequences are similarly sorted. Then by the Rearrangement Inequality, where we take our indices modulo , with equality exactly when all the , and therefore all the , are equal. Dividing both sides by gives the desired inequality.
Proof by Calculus
We will start the proof by considering the function . We will now find the maximum of this function. We can do this simply using calculus. We need to find the critical points of , we can do that by finding and setting it equal to . Using the linearity of the derivative . We need Note that this is the only critical point of . We can confirm it is the maximum by finding it's second derivative and making sure it is negative. letting x = 1 we get . Since the second derivative , is a maximum. . Now that we have that is a maximum of , we can safely say that or in other words . We will now define a few more things and do some manipulations with them. Let , with this notice that . This fact will come into play later. now we can do the following, let and plug this into , we get Adding all these results together we get Now exponentiating both sides we get This proves the AM-GM inequality.
Proof of Weighted AM-GM
Proof by Convexity
We note that the function is strictly concave. Then by Jensen's Inequality, with equality if and only if all the are equal. Since is a strictly increasing function, it then follows that with equality if and only if all the are equal, as desired.
Alternate Proof by Convexity
This proof is due to G. Pólya.
Note that the function is strictly convex. Let be the line tangent to at ; then . Since is also a continuous, differentiable function, it follows that for all , with equality exactly when , i.e., with equality exactly when .
Now, set for all integers . Our earlier bound tells us that so Multiplying such inequalities gives us
Evaluating the left hand side:
for
Evaluating the right hand side:
Substituting the results for the left and right sides:
as desired.