GBB Services  
GBB Services : Resources : Mathematics : a^{2}b^{2}=(ab)(a+b) 
Note: (written November 6 2016) I'm glad you found this essay I wrote long ago. It's archived here on my website, together with my other essays from a decade ago. If you want to see my more recent content, my blog is the place to find it. Exploring the identity a^{2}b^{2} = (ab)(a+b)by Walter VanniniRectangles and Squares Money Hyperbolas Wave Equation Inner Products Multiplication Table Odd Numbers Cracking RSA Continued Fractions Infinite Product Feedback IntroductionThere are many possible approaches to understanding this identity. Recently Peter Hendrickson asked me who my target audience is, and I told him it was me, assuming that for some reason I was ignorant about the item under discussion. For this identity, I would be sure to tell myself that there are various algebraic rewritings of the identity. There are three broad classifications:
Throughout the discussion I would tell myself that the above identities hold in more general settings: the elements don't need to be real numbers, and the operations don't need to be multiplication and addition of reals. Also, with any algebraic result, I always want to know about geometric interpretations or consequences, so I'd be sure to mention something about that. Ditto for any physical interpretation. And, I'd mention something about generalizations and related identies: e.g. What about the difference of cubes, what about the product of three elements, what about the sum of squares, etc. Okay, that's the broad overview and plan of attack. Now let's dig in. Rectangles and SquaresThe square of a positive real s, as in "s raised to the power 2", i.e. s^{2}, is the area of a square with side length s. The product of two positive reals a and b is the area of a rectangle with sides of length a and b. Using these geometric interpretations, it's clear that the identity is saying something about areas of squares and rectangles. The above diagram gives some insight, via geometric dissection, about why the identity is true. At this point, I'd like to make a little digression into symmetry considerations. In the above geometric sequence, an arbitrary choice was made, and the initial symmetry of the picture was broken. We could just as easily have gone the other way: Strangely enough, this actually leads to a result. Considering both diagrams together, and forming a "superposition", we should expect that a^{2}b^{2} is some kind of weighted average of (ab)(a+b) and (a+b)(ab). Since the two items are the same, this doesn't produce anything new. But, what if they weren't the same? What if multiplication were not commutative (i.e. xy was not the same as yx). This would be the case if a and b were n by n square matrices, or if they were quaternions. In that case, the identity would not be true. Using quaternions for example: i^{2}j^{2} is 0, but (ij)(i+j) is not. Going back to algebra, this is because (ab)(a+b) = a^{2}+abbab^{2}, and ab and ba don't cancel each other out to give a final result of a^{2}b^{2}, unless ab=ba.
But, combining the two versions:
Assuming it makes sense to divide by two, we get a weighted average identity
I must admit at this stage that I don't know of any use for this, but I couldn't resist mentioning it. MoneyAfter that digression into abstract generality, it's time to go back to the concrete and specific.
Consider the special case of the identity where a is 1:
There is a nice financial interpretation of this. If an item has a 10% sales tax (corresponding to x=0.1), and has a 10% discount, then the final price is not the same as the initial price, but is instead reduced by 1% (since x^{2}=0.1^{2}=0.01 ie 1%). Also, since (1x)(1+x) and (1+x)(1x) are the same, it doesn't matter whether the discount is done first or last. Two Formulas for HyperbolasWhen I first learned about the hyperbola, back in high school, it was via the equation y = 1/x The instructor told us that the curve was called the hyperbola, and went on to talk about asymptotic behaviour, the two asymptotes (x=0 and y=0) and all the usual things. No problem! We then moved on to other topics. A few weeks later, we tackled second order equations in the plane: ie Ax^{2} + By^{2} + Cxy + Ex + Dy + F = 0. We then got to the equation x^{2}y^{2}=1, and we were told that it was a hyperbola, and the two asymptotes were x=y and x=y. It seems I was the only one who noticed and was disturbed by the fact that a "hyperbola" had been defined twice, and in two different ways that seemed unrelated. The instructor didn't talk about this. Our textbook, which was geared towards preparing for an exam, had nothing to say on the topic. I somehow convinced myself that the equations were defining the same kind of curve. I forget just how I did it, but I certainly didn't "understand". I probably manipulated some symbols, and magically everything worked out. It was years later that I finally realized that the two equations were easily reconciled via the a^{2}b^{2}=(ab)(a+b) identity. It's a little more obvious if y=1/x is rewritten as xy=1. Somehow, I don't think that would have been enough of a clue for me when I was in high school. Some things take time.
Just to say a little more about this, we have that the coordinate transform
The transform described here is a 45° rotation followed
by a scaling (with the scaling factor being √2).
If we stick to rotations, the transformation to consider is
Wave Equation
A solution U(x,t) to the
one dimensional wave equation in two variables x and t
is a solution to the partial differential equation:
As we saw earlier, the a^{2}b^{2} identity applies as long as ab=ba. In the case where a and b are the differential operators d/dx and d/dt, we have that ab=ba, ie (d/dx)(d/dt) = (d/dt)(d/dx) since d^{2}f/dxdt = d^{2}f/dtdx. So, the wave equation becomes (d/dx  d/dt) (d/dx + d/dt) U = 0
Writing
The solution to this is "trivial":
U = f(u) + g(v) = f((x+t)/2) + g((xt)/2), i.e.
The factorization given by the a^{2}b^{2} identity ends up telling us that any solution to the wave equation is the superposition of a wave moving to the left at speed 1, and a wave moving to the right at speed 1. Inner Products and Length Squared
For two vectors
u=(u_{1},u_{2},u_{3}) and
v=(v_{1},v_{2},v_{3}),
the inner product
u•v
is given by
For a vector
v=(v_{1},v_{2},v_{3}),
the length squared
v^{2}
is
It's pretty clear that they're related:
The a^{2}b^{2} identity applies in this situation, where multiplication is the dot product, and squaring is length squared: u^{2}v^{2} = (uv)•(u+v) This is mildly interesting, but the equivalent form of the identity, which relates a product to a difference of squares is really interesting: u•v = (u+v)/2^{2}(uv)/2^{2}
This tells us that the inner product function is derivable from
the length squared function:
And, of course, there's nothing special about R^{3} with the standard inner product. Everything generalizes to semidefinite products and to Hilbert spaces. Patterns in the Multiplication TableWhen I was in elementary school, we memorized our multiplication table all the way up to the 12 by 12 case. We were tested over and over, year after year. We got to know that multiplication table!
In high school, my good friend Jack Robin told me that he'd discovered a pattern in the multiplication table, and he started describing it to me. If you moved off the main diagonal, consisting of squares, via the other diagonal direction, the values went down by one. At first I didn't understand what he was talking about, but he showed it to me, and he was right. I'd never noticed that before.
The next question was: why?
As it turned out, we'd just been learning algebra, and I was able
to make the connection! It was just an application of the
a^{2}b^{2} identity:
Sums of Consecutive Odd Numbers
One of the special cases of the identity is:
After years of having my mind deformed by programming, it seems really natural to start at a positive integer n, and iterate down to 1. Doing that, we get:
Adding up the columns, we get
It's also true that if we hadn't gone all the way to one,
but only went down to m+1, we would get:
Cracking RSARSA encryption (named after Rivest Shamir and Adleman) wouldn't provide much protection if we knew how to factor large numbers quickly. In particular, given two large primes p and q, if we could factorize the product, N=pq quickly, RSA encryption would be broken. The a^{2}b^{2} identity gives us a way of factorizing quickly, if p and q aren't chosen properly. Before going into details, I'd like to digress a little and say a little bit more about RSA encryption. If you study the subject, something that you'll quickly learn is that what we really want to get is φ(N), where φ is Euler's phi function. Since φ(N)=(p1)(q1), factorizing N will certainly allow you to get φ(N). It's also true that knowing φ(N) will allow you to get the factorization. This is because knowing N=pq and φ(N)=(p1)(q1) will give you p+q, via N+1φ(N). Well, once you have pq and p+q, the identity gives us pq via pq = ((p+q)/2)^{2}  ((pq)/2)^{2}. And, once we have both p+q and pq we have p and q. Okay, that's the end of the digression. Although factorizing N=pq is hard in general, if p and q are chosen badly, then the factorization is easy. The obvious case, that takes little mathematical sophistication to realize, is that if p is small, say smaller that 10^{10}, then simply testing p=2, then p=3, then p=4, etc, will quickly give the game away. Gigahertz machines can check a mere 10^{10} cases pretty quickly. So, don't pick p and q to be to close to the extremes. i.e. if pq is a 200 digit number, don't pick p and q so that p has a mere 10 digits, and q has 190 digits. As it turns out, you don't want to be too close to the middle either, by which I mean don't pick p and q so that p has 100 digits, q has 100 digits, and p and q differ only in the least significant 55 digits. The reason is because of the identity! Suppose K^{2} is a perfect square greater that N. If K^{2}N is a perfect square J^{2}, then N=K^{2}J^{2} instantly gives us the factorization (KJ)(K+J). So, instead of iterating p starting from 2 and counting up, we can start from M=K^{2}N, where K^{2} is the smallest square greater than N, and increment by 2K+1, then 2K+3, then 2K+5 etc to get M=(K+1)^{2}N, M=(K+2)^{2}N, M=(K+3)^{2}N, and at each stage test to see if M is a square. In the 200 digit example above, it would take a mere 10^{10} steps to find p and q. By the way, currently there isn't any wellknown proof establishing that factorizing N=pq is difficult in general. Recursion and Continued FractionsThere's a recursive formula hiding in the identity, and it leads to continued fraction expressions. Here are the details:
The identity can be written as
This can be viewed as a way of writing xy in terms of x+y. But x+y can be written in terms of xy via simple addition, ie x+y=2y+(xy).
So, we have a formula for x+y in terms of x+y!
Here it is:
Well, once we have something like that, it's hard to resist
recursing infinitely to get:
Choosing x and y so that x^{2}y^{2} is something simple, like 1, we get that x must be √(1+y^{2}). Using this, the formula becomes: y+√(1+y^{2}) = 2y + 1/(2y+1/(2y+1/…)). In the well known continued fraction notation. This is saying y+√(1+y^{2}) = [2y,2y,2y,…] Some immediate consequences are:
for y=1/2, we get that φ, the golden ratio
(1+√5)/2 satisfies
φ = [1,1,1,…] leading to approximations via ratios of
Fibonacci numbers
Writing y=n, we get that
For example, taking n=1, we get that
Recursion and an Infinite ProductThere's another recursive formula hiding in the identity, and it leads to an infinite product, which in turn leads to an infinite series. Here are the details:
Consider the special case
Well, viewing x^{2} as a single variable
u, this is saying
Now apply the 1/(1x) formula to 1/(1u),
and we're well on our way to infinite recursion:
This process leads to the infinite product
Multiplying out, this gives the wellknown series 1/(1x) = 1 + x +x^{2} + x^{3} + x^{4} + … FeedbackIf you have corrections, additions, modifications, etc please let me know mailto:walterv@gbbservices.com
