Saturday

April 19, 2014

April 19, 2014

Posted by **kim** on Sunday, December 3, 2006 at 6:13pm.

Prove that the roots of

ax^2 + (a + b)x+b are real for all values of k

note the "x"s aren't multiplication signs.

a x^2 + bx + c has the discriminant of

D = b^2 - 4ac.

If D is nonnegative then the function has real roots.

In this case you have

D = (a + b)^2 - 4 a b = (a-b)^2

which is larger than or equal to zero because it is a square.

**Related Questions**

MATH,HELP - Can someone show me how to even do this problem. Find all positive ...

Discrete Math - Prove that if a and b are integers with a not equal to zero and ...

Precalculus - "Show that x^6 - 7x^3 - 8 = 0 has a quadratic form. Then find the ...

maths - if a, b,c are real numbers and not all equal, prove that the equation (c...

maths Please help!!!!! - if a, b,c are real numbers and not all equal, prove ...

Discrete Math - I have a review problem I am having problems with. This is the ...

math: polynomials - Please help me with these three problems! 1. solve for x: 3x...

Math - Which describes the number and type of roots of the equation x^2-625=0? A...

Algebra II - Which describes the number and type of roots of the equation x^2 -...

Math - For how many ordered triples of positive real numbers (a,b,c), do the ...