Asked by NENE
Two spheres are launched horizontally from a 1.1m -high table. Sphere A is launched with an initial speed of 5.0M/S . Sphere B is launched with an initial speed of 3.0M/S
A. What is the time for the sphere A to hit the floor? and sphere B to hit the ground
B. What is the distance that sphere A travels from the edge of the table? And what is the distance for sphere B travels from edge of the table??
A. What is the time for the sphere A to hit the floor? and sphere B to hit the ground
B. What is the distance that sphere A travels from the edge of the table? And what is the distance for sphere B travels from edge of the table??
Answers
Answered by
drwls
Since they are both launched horizontally, there is zero initial vertical velocity component in each case and the times that they take to hit the ground are the same.
That time T can be obtained by solving
(g/2) T^2 = 1.1 m
The horizontal distance travaled when they hit the ground is Vo*T
Vo = 3.0 or 5.0 m/s
That time T can be obtained by solving
(g/2) T^2 = 1.1 m
The horizontal distance travaled when they hit the ground is Vo*T
Vo = 3.0 or 5.0 m/s
Answered by
NENE
the second question is very tricky. can you please give me the equation to use to find the distance??
Answered by
NENE
i got the answer. thanks
There are no AI answers yet. The ability to request AI answers is coming soon!
Submit Your Answer
We prioritize human answers over AI answers.
If you are human, and you can answer this question, please submit your answer.