The Gateway to Computer Science Excellence
First time here? Checkout the FAQ!
+1 vote

Suppose the length of a 10Base5 cable is $2500$ m. If the speed of propagation in a thick coaxial cable is $200,000,000$ m/s, how long does it take for a bit to travel from the beginning to the end of the network? Assume that there is a $10$ μs delay in the equipment due to repeaters and there is a repeater for every $500$ meters?

  1. $26.5$ micro seconds
  2. $52.5$ microseconds
  3. $47.5$ micro seconds
  4. $37.5$ micro seconds
asked in Computer Networks by Boss (17.2k points)
edited by | 125 views

2 Answers

+2 votes

Propagation delay ($T_{p}$)=$\frac{500m}{2*10^8 m/sec}$= $2.5{\mu }$sec

Total time taken= $2.5*5+10*4$= $52.5{\mu }$sec

Hence Option B) is correct


answered by Boss (15k points)

 why we are not taking transmission time on each  repeter......for first repeater it will zero 



@Arjun sir @GATEBOOK

It is negligible.
0 votes

The answer is 52.5 microseconds 

40  microseconds  due to the router and 12.5 due to the propagation delay


answered by Active (1.9k points)

Related questions

Quick search syntax
tags tag:apple
author user:martin
title title:apple
content content:apple
exclude -tag:apple
force match +apple
views views:100
score score:10
answers answers:2
is accepted isaccepted:true
is closed isclosed:true
50,122 questions
53,241 answers
70,480 users