Question
A jogger runs a distance of six miles at an average speed of four miles an hour, how many minutes did this take him?
Answers
d = r*t.
6 = 4 * t.
t = Time in hours.
Convert to min.
6 = 4 * t.
t = Time in hours.
Convert to min.
Related Questions
A jogger leaves home and runs at a constant speed of 4 miles per hour. Nine minutes later, a bicycli...
A car traveled 281 miles in 4 hours 41 minutes. What was the average speed of the car in miles per h...
A hiker can walk 2 miles in 43 minutes.
What is his average speed in miles per hour? Round your a...
Max completed a 10 mile run in 1 hour 20 minutes. Which dimensional analysis would you use to find...