Here is the answer, once and for all, on the answer to 2 + 2.......
I'm afraid I can't take credit for this.![]()
Let a & b each be equal to 1. Since a ^ b are equal,
b^2 = ab
Since a equals itself, it is obvious that
a^2 = a^2
Subtract equation 1 from equation 2. This yeilds
(a^2) - (b^2) = (a^2)-ab
We can factor both sides of the equation; (a^2)-ab equals a(a-b). Likewise, (a^2)-(b^2) equals (a + b)(a - b) (Nothing fishy is going on here. Ths statement is perfectly true. Plug in numbers and see for yourself!) Substituting into the equation 3 , we get
(a+b)(a-b) = a (a-b)
So far, so good. Now divide both sides of the equation by (a-b) and we get
a + b = a
b = 0
But we set b to 1 at the very beginning of this proof, so this means that
1 = 0
Now we can use this general idea and apply it to two. and can alter the number into any number. However the altered number can not be more than 1 away from the orginal number. therefore A+B= 5
A= 2 B=3
What does the ^ mean?