radio signals travel at a rate of 3x10^8 meters per second. how many seconds will it take for a radio signal to travel from a satellite to the surface of the Earth if the satellite is orbiting at a height of 3.6x10^7

1.) 8.3 seconds
2.) 10.8 x 10^15 seconds
3.) 1.2 x 10^-1 seconds
4.) 1.08 x 10^16 seconds

thank you!! c:

Respuesta :

time = distance/speed

[tex] T = \frac{3.6 \times 10^7}{3 \times 10^8}= \boxed{1.2 \times 10^{-1} sec} [/tex]

Answer:

Option 3

The time taken for a radio signal to travel from a satellite to the surface of the Earth is [tex]1.2\times10^{-1}[/tex]

Step-by-step explanation:

Given : Radio signals travel at a rate of [tex]3\times10^8[/tex] meters per second. The satellite is orbiting at a height of [tex]3.6\times10^7[/tex].

To find : How many seconds will it take for a radio signal to travel from a satellite to the surface of the Earth?

Solution :

The speed of radio signal is  [tex]3\times10^8[/tex] meters per second.

The distance or height of satellite orbiting is [tex]3.6\times10^7[/tex].

[tex]\text{Time}=\frac{\text{Distance}}{\text{Speed}}[/tex]

Substitute the value in the formula,

[tex]\text{Time}=\frac{3.6\times10^7}{3\times10^8}[/tex]

[tex]\text{Time}=1.2\times10^{-1}[/tex]

Therefore, Option 3 is correct.

The time taken for a radio signal to travel from a satellite to the surface of the Earth is [tex]1.2\times10^{-1}[/tex]