Mathematics
posted by Sarah .
Show that 2i and 1i are both solutions to the equation x^2(1+i)x+(2+2i)=0 but their conjugates,2i and 1+i are not. Then explain why this does not violate the Conjugate Zeros Theorem.

The conjugate zeroes theorem says:
"If f(x) is a polynomial having only real coefficients and if a + bi is a zero of f(x), then a – bi is also a zero of f(x)."
Does the given equation satisfy all the conditions required for the theorem to apply?