Please show work to how you find this answer.

A baseball is thrown with an initial velocity of 30 m/s at an angle of 45 degrees. Neglecting air resistance, how far away will the ball land?

I am not going to do it for you to copy and paste. Get real.

break up the intialvelocity vector in to vertical and horizonal components.

Using the vertical, solve for the time in air.

hfinal=hinitial+Viv*time-1/2 g time^2

solve for time

Then,having time,
horizontaldistance= initialhorizontalveloicyt*time