Optical Fiber Cabling
Fiber Testing FAQ
Optical Power Measurements
Optical power is measured in linear units of milliwatts (mW), microwatts (uW - really the greek letter "mu"W),
nanowatts (nW) and decibels (dB).
What is the difference between "dBm" and "dB"?
dB is a ratio of two powers, for example the loss in a fiber optic cable. When power is measured in linear units
(mW, uW or nW), dB is calculated on a log scale using this formula:
power (dB) = 10 log (power1/power2)
If we are measuring absolute power levels, the measurement is generally referenced to 1 milliwatt (mW), is
expressed as "dBm" and the calculation becomes:
power (dBm) = 10 log (power/1 mW)
Thus 1 mW = 0 dBm, 1 uW = -30 dBm, 1 nW = -60 dBm and two equal powers compared are 0dB (eg. power
being the same, there is no loss.)
LED: -10 to -25 dBm into 62.5/125 fiber
Telcom/LAN laser: 0 to -13 dBm into singlemode fiber, to +20 dBm with DWDM and fiber amplifier systems
CATV Laser : +16 to 0 dBm into singlemode fiber
It depends on the network and type of source. When measured at the end of the network cable, the source
output power will usually be in these ranges:
LAN/LED: -20 to -35 dBm into 62.5/125 fiber
Telcom/LAN laser: -20 to -45 dBm into single-mode fiber
CATV Laser : 0 to -10 dBm into single-mode fiber
The loss budget is a calculation of how much attenuation a link should have. You compare that loss to the
dynamic range of the networking equipment to see if the range and link loss are compatible.
All optical power meters which are calibrated to NIST (the US standards body) or any national standards lab will
measure optical power to an uncertainty of about +/- 0.2 dB or 5%. Therefore, since every power meter has an
uncertainty of +/- 0.2 dB, any two meters can differ by 0.4 dB in the worst case (one at +0.2 dB and one at -0.2
dB) even if both are within their specification!
- Are more complex or higher priced FO power meters more accurate?
- The high priced meters offer better dynamic range and more features, but not better absolute measurement
- uncertainty.
Why is the measurement uncertainty so high? That is because there are three to four calibration transfers from
the NIST absolute optical power standard before the meter is delivered to the customer. The NIST standard has
an uncertainty of about 1% itself and every transfer adds errors of about 1%.
NIST only offers calibrations at 850, 1300 and 1550 nm, so those meters that have calibrations at other
wavelengths have to extrapolate to those values, increasing the measurement uncertainty at those
wavelengths.
If my source is at a slightly different wavelength from the standard calibration wavelength, doesn't that add to
measurement error?
Perhaps, but the wavelength of most sources is not known by the person making the measurement. If everyone
uses meters calibrated at only a few specific wavelengths, everyone can be testing to the same standard and
will get more closely correlated measurements on sources of unknown wavelengths.
Use an OTDR on cable plants, OCWR on patch cords. ORL testing with ORL tester (or what Telcordia/Bellcore
calls an OCWR or optical continuous wave reflectomenter) is only applicable to short patch cords. If you try using
them on an installed cable plant, the integrated backscatter from the length of the fiber will overwhelm the
connector back reflection. Twenty km of fiber gives about the same backscatter as a connector with 20 dB ORL
and you cannot tell where the backscatter comes from! It's better to use an OTDR to find ORL problems on
installed cable plant.