Physics, asked by Anonymous, 5 months ago

A satellite is observed from two points A and B at a distance 1.3x107 m apart on earth. If angle subtended at the satellite is 1°30’, find the distance of the satellite from earth
I want the correct answer

Answers

Answered by mohit810275133
2

Explanation:

ANSWER

ϕ=1°54 =114

=(114×60)"=114×60×4.85×10 −6

rad=3.32×10 −2 rad

Basis,b=AB=1.276×10 7 m

So,the distance of the moon from the earth,

S= ϕB = 3.32×10 −21.276×10 7 =3.84×10 8

Similar questions