Let The roots of the equation x^2 + ax + b = 0 be alpha and beta.
Similarly, Let the roots of the equation x^2 + bx + a = 0 be alpha and gamma. [There is a common root.]
So, alpha + beta = -a rArr alpha = -(a + beta)
And alpha + gamma = -b rArr alpha = -(b + gamma)
So, -(a + beta) = -(b + gamma)
rArr a + beta = b + gamma
rArr beta - gamma = b -a
rArr (beta - gamma)^2 = b^2 - 2ab + a^2...............(i)
Again, alphabeta = b rArr beta = b/alpha
And, alphagamma = a rArr gamma = a/alpha
Then, beta - gamma = (b -a)/alpha and betagamma = (ab)/alpha^2
Putting this in eq (i)
((b-a)/alpha)^2 = (b-a)^2
rArr 1/alpha^2 = 1
rArr alpha = +-1
So, betagamma = +-ab and beta + gamma = (b + a)/alpha = +-(a + b)
So, The Required Equation, which has beta and gamma as roots is
x^2 +- (a + b)x +- ab = 0
But (a + b) = +- 1 [As, alpha = +-1, beta = b and gamma = a]
So, The equation becomes x^2 +-x +-ab = 0
Hence Proved.