RF Cafe Software
About RF Cafe
1996 - 2022
BSEE - KB3UON
RF Cafe began life in 1996 as "RF Tools" in an AOL screen name web space totaling 2 MB. Its primary purpose was to provide me with ready access to commonly needed formulas and reference material while performing my work as an RF system and circuit design engineer. The Internet was still largely an unknown entity at the time and not much was available in the form of WYSIWYG ...
All trademarks, copyrights, patents, and other rights of ownership to images and text used on the RF Cafe website are hereby acknowledged.
My Hobby Website:
Try Using SEARCH
to Find What You Need.
There are 1,000s of Pages Indexed on RF Cafe !
A lot of people ask, so here's the answer to the eternal question, "How did 50 Ω get to be the standard RF transmission line impedance?" Here are a few stories.
There are probably lots of stories about how 50 ohms came to be. The one I am most familiar goes like this. In the early days of microwaves - around World War II, impedances were chosen depending on the application. For maximum power handling, somewhere between 30 and 44 Ω was used. On the other hand, lowest attenuation for an air filled line was around 93 Ω. In those days, there were no flexible cables, at least for higher frequencies, only rigid tubes with air dielectric. Semi-rigid cable came about in the early 50s, while real microwave flex cable was approximately 10 years later.
Somewhere along the way it was decided to standardize on a given impedance so that economy and convenience could be brought into the equation. In the US, 50 Ω was chosen as a compromise. There was a group known as JAN, which stood for Joint Army and Navy who took on these matters. They later became DESC, for Defense Electronic Supply Center, where the MIL specs evolved. Europe chose 60 Ω. In reality, in the U.S., since most of the "tubes" were actually existing materials consisting of standard rods and water pipes, 51.5 Ω was quite common. It was amazing to see and use adapter/converters to go from 50 to 51.5 Ω. Eventually, 50 won out, and special tubing was created (or maybe the plumbers allowed their pipes to change dimension slightly).
Further along, the Europeans were forced to change because of the influence of companies such as Hewlett-Packard which dominated the world scene. 75 Ω is the telecommunications standard, because in a dielectric filled line, somewhere around 77 Ω gives the lowest loss. (Cable TV) 93 Ω is still used for short runs such as the connection between computers and their monitors because of low capacitance per foot which would reduce the loading on circuits and allow longer cable runs.
Volume 9 of the MIT Rad Lab Series has some greater details of this for those interested. It has been reprinted by Artech House and is available.
Gary Breed wrote in his High Frequency Electronics publication that one explanation offered via an old househam's tale is that, "The most common story is that 50-ohm high power coaxial lines were first made using standard sizes of copper pipe, such as 3/4 inch for the inner conductor and 2 inch for the outer conductor."
Check this out - someone referenced this page on Wikipedia