A tennis player hits the ball horizontally with a speed of 30m/s at 1.5m above the ground.
How far away will it land if the ball was hit at a speed of 35m/s
h = Vo*t + 0.5g*t^2 = 1.5 m.
0 + 4.9t^2 = 1.5
t^2 = 0.306
Tf = 0.553 s. = Fall time.
D1 = V*Tf = 30m/s * 0.553s = 16.6 m.
D2 = 35m/s * 0.553s = 19.4 m.
posted by Henry