Math
posted by A.S. .
Suppose a scientist on Earth throws a baseball upward. The scientist lets go of the ball 2 meters above the ground with an initial velocity of 10 meters per second. How long does it take for the ball to hit the ground (H=0)? (Use the quadratic formula to solve this problem, show all work). Write a statement to interpret your results.

h = ho + Vi t (g/2)t^2 assume g = 9.8m/s^2
0 = 2 + 10 t 4.9 t^2
4.9 t^2  10 t  2 = 0
t = [10 +/sqrt(100+39.2)]/9.8
t = [ 10 +/11.8 ]/ 9.8
(negative result is before you threw it)
t = 2.22 seconds