Asked by kieran
The drag on an airplane traveling at velocity v is D=av^2+(b/v^2) where a and b are positive constants. At what speed does the airplane experience the least drag.
I have so far that D'=-2b/v^3 do i set this equal to zero then i get 0=-2 don't think that is the right way where do i go from here
I have so far that D'=-2b/v^3 do i set this equal to zero then i get 0=-2 don't think that is the right way where do i go from here
Answers
Answered by
Steve
I get
D' = 2av - 2b/v^3
then set D'=0 to get
2av - 2b/v^3 = 0
v = ∜(b/a)
D' = 2av - 2b/v^3
then set D'=0 to get
2av - 2b/v^3 = 0
v = ∜(b/a)
Answered by
kieran
how do you get 2av should that be zero since a is a constant or did i miss something in class
Answered by
Steve
a is a constant, just like b.
d/dv(av^2) = (a)(2v) = 2av
it doesn't make the whole term 0, it just multiplies. Funny you kept the b, but not the a.
d/dv(av^2) = (a)(2v) = 2av
it doesn't make the whole term 0, it just multiplies. Funny you kept the b, but not the a.
There are no AI answers yet. The ability to request AI answers is coming soon!
Submit Your Answer
We prioritize human answers over AI answers.
If you are human, and you can answer this question, please submit your answer.