A bullet is fired with a certain velocity at an angle above the horizontal at a location where g= 10.0 m/s^2. The initial x any components of its velocity are 86.6 m/s and 50 m/s respectively. How long does it take before the bullet hits the ground?

2 answers

v₀=sqrt{v₀² (x) +v₀²(y)}=
=sqrt{86.6²+50²} =
=100 m/s.

tan α= v₀(y)/v₀ (x)=50/86.6
α= tan ⁻¹(50/86.6) =30°

t= 2vₒ•sinα/g
5.0 s