how do you calculate the time when average speed is given

time = distance divided by speed

the distance is 50 and the speed is 60 which equals 0.8333.... not sure is that answer makes sense

Yes, it makes sense. The time is 0.83 hour.

60 * 0.83 = 50 minutes

To calculate the time when the average speed is given, you need to know the total distance traveled. Once you have both the average speed and the total distance, you can use the formula:

Time = Distance / Average Speed

Here's a step-by-step explanation of how to calculate the time:

1. Determine the average speed: This is given in the problem statement. For example, let's say the average speed is 60 miles per hour.

2. Identify the total distance: This could also be given in the problem statement or you may need to calculate it using other given information. For instance, if the total distance is 180 miles, you have this value.

3. Apply the formula: Divide the total distance by the average speed as per the formula mentioned earlier.

Time = Distance / Average Speed

For example, using the values from the previous steps:
Time = 180 miles / 60 mph
Time = 3 hours

Therefore, in this example, it would take 3 hours to travel a distance of 180 miles at an average speed of 60 miles per hour.