In baseball, the pitcher's mound is 60.5 feet from home plate. If the fastest pitch ever recorded in Major League Baseball was 105.1 miles per hour, how many seconds did it take for the ball to travel from the pitcher's mound to home plate? Round the answer to the nearest tenth.

Respuesta :

Answer:

t=0.4 sec

Step-by-step explanation:

Remember that

1 mile=5,280 feet

1 hour=3,600 seconds

Let

s -----> the speed in ft/sec

d ----> the distance in ft

t -----> the time in sec

s=d/t

Solve for t

t=d/s

step 1

Convert  miles/hour to ft/sec

105.1 mi/h=105.1(5,280/3,600)=154.15 ft/sec

step 2

Find the time

t=d/s

we have

s=154.15 ft/sec

d=60.5 ft

substitute

t=60.5/154.15

t=0.4 sec

Answer:

t = 0.4

Step-by-step explanation:

(

1 hour

105.1 miles

)(

1 mile

5280 feet

)(

60 minutes

1 hour

)(

60 seconds

1 minute

)(60.5 feet) ≈ 0.4 seconds