Can Fibre Optic Ethernet Cables Be Longer Than The Standard ?

Short Answer is “It depends, but usually yes. And you probably should” Long answer follows:

Laser Power

The length of a fibre optic cable run is primarily determined by whether a strong enough laser signal is received by the Gigabit transceiver, commonly known as an SFP1. Although there are other considerations such as modal dispersion and signal integrity for signal propagation,  lets choose to keep this simple for now

When the IEEE creates the standards, part of the technical considerations is the strength of “laser launch” or transmitter, and the sensitivity of “laser receiver”. Both of these items are measured in terms of dBM2. For 1000BaseLX/LH using a single mode laser:


1000Base LX/LH Gigabit Ethernet
Laser Transmitter Characteristics: Minimum Launch Power -9.5 dBm Maximum Launch Power -3 dBm
Receiver Characteristics: Minimum Receiver Sensitivity -20 dBm Maximum Input Power -3 dBm


Note that for 1000BaseSX the figures are significantly different. this is why the shorter cable length comes into play, less power budget to work with, but in a given situation we may well be able to exceed the standard and get away with it.


10000BaseSX Gigabit Ethernet
Laser Transmitter Characteristics: Minimum Launch Power -10 dBm Maximum Launch Power -4 dBm
Receiver Characteristics: Minimum Receiver Sensitivity -17 dBm Maximum Input Power 0 dBm


Note that different vendors display their min/max figures differently e.g. Cisco documents the max/min values like this for 1000BASE-SX:
Transmit Power Range -3 to -9.5 dBm
Receive Power Range 0 to -17 dBm


The variation could be due to many factors such manufacturing process, quality control, quality of materials and so on. Manufacturers contribute to the standards, usually, so that they can ensure they are able to make products that fit the standards.

Power Sums (with Math)

Now from this we can deduce that the worst case scenario would be when the launch power from the single mode launch laser is transmitting at the lowest possible power of -9.5dBm and the receiver sensitivity is at least sensitive of -20dBM which gives us a worst case of 11.5dBm of “headroom” for guaranteed function. If the power level at the receiver was less than this, then there are no guarantees that your receiver could decide the serial signal correctly.

That is, once the received signal drops below the power level you might be screwed.

In any given circumstance, launch power is affected by the quality of the fibre connector and the laser installed, not much we can do there. Receiver Sensitivity is determined by the amount laser power received at the other end, which is degraded by patch leads, fibre length, fusion splice or mechanical splice, types of connectors etc.

Loss on Fibre Optic Path

In a perfect case, the laser output of the SFP would be directly connected to the receiver. The use of a fibre optic cable causes loss of the light wave. The laser signal is attenuated by the cable itself, each connector from the SFP to patch lead, the patch lead to fibre backbone. The fibre backbone might consist of several splices or joins in the glass core as well as shown in this diagram:

Fibre optic txvr 1

In short, every break in the fibre core causes some type of loss. Losses for fibre are quite small in most circumstances.

Does it work ?

The interesting thing is that even on cabling where there is more than 11.5dB of loss, then your connection may well work because you might have good launch characteristics i.e. up to -3dBm of launch giving up to 6.5 dbM of headroom. Potentially, the receiver could work on an even lower input power. Thus the recommended length of 550m for 1000BaseLX on 62.5um multimode could be longer depending on how good the end to end performance of the cabling system is.

However, the loss in the cable can vary over time since loss can fluctuate with temperature and other environmental issues, and this may cause intermittent errors. The gamble with overrunning the standard lengths is that we risk intermittent and indeterminate faults, often with no apparent cause.

So yes, it depends

The best way to run long cables is to use a fibre optic cabling test that measures the end to end loss, including path leads. But you can vary this to your own requirements.

In most cases, the cabling contractors have oversold the requirements for fibre cabling and forced most customers to use fusion spliced backbone cabling which has less than 0.1dBm loss per connection when mechanical splicing would be cheaper and faster to complete. Even on a big campus with 2 kilometres run of single mode, it’s unlikely that you would incur enough loss to be a concern (although model dispersion might then be a concern – but that’s another story)

The EtherealMind View

But it’s not guaranteed so you need to consider whether even this very low level of risk. In my view, you can easily double the length of fibre cabling runs with good quality transceivers provided that you test the installed cabling.

For educational campus, I would certainly recommend this practice since it would save a huge amounts of money in otherwise useless switch purchases to act as signal regenerators.

Remember, the IEEE is a zero risk, backward looking, nothing can change too much and don’t rock the boat organisation. Conservative doesn’t even begin to describe their approach to physical transmission performance.

Go ahead, go long.

  1. SFP Small Form factor Pluggable module.
  2. dBm (sometimes dBmW) is an abbreviation for the power ratio in decibels (dB) of the measured power referenced to one milliwatt (mW). It is used in radio, microwave and fiber optic networks as a convenient measure of absolute power because of its capability to express both very large and very small values in a short form. Compare dBW, which is referenced to one watt (1000 mW). Wikipiedia
  • PG

    I know of a number of 62.5 multimode cables running LH SFPs at >800m with no errors for yrs. Not even using the mode conditioning patch leads. The fibre
    We didn’t try this until the fibre had been tested, which was how we learned how long the runs were.

    • PG

      It’s also worth mentioning that Cisco will support multimode runs with LH SFPs up to 2km if certain conditions are met.

    • Matt

      I’ve seen similar, runs over 1km with LH GBICs on MMF that ran fine for years. I’d probably still go for mode conditioned cables to be on the safe side.

  • legox

    if we are out of the campus (metro or wide ..) a common practice is to add the EOL (end of life).
    This parameter is a cumulative and statistic misure of the future fiber loss (cut/repair, laser aging loss etc..).
    For eg. for 15 year end of life 3 db is a common value.

  • Wes Felter

    Related question: What is the deal with 10GBASE-SR vs. LRM range? I have some multimode fiber between two buildings and it’s not clear which type of transceiver would be better. 

  • Pingback: Packet Pushers – Show 83 – We Are An Independent Show — My Etherealmind

  • Pingback: Show 83 – We Are An Independent Show

  • Guest

    >transmitting at the lowest possible power of -9.5dBm
    >receiver sensitivity is at least sensitive of -20dBM which
    >gives us
    a worst case of 11.5dBm of “headroom”
    I don’t understand how did you get value in dBm units when you subtract dBms from dBms. In my opinion, subtracting dBms from dBms is equal to taking the logarhythm of division of relevant milliWatts, which results in dimensionless value. So, I think, -20dBm-(-9.5dBm)=-10.5 _dB_, not _dBm_.