Ian said:
OK.
At 450MHz and temperature T1 centigrade, your 5000 foot of coax will
have an attenuation of A1 dB.
[The value of A1 obviously depends on the characteristics of that
particular type of coax, and on its length. For a different frequency /
temperature / length / type of coax, the value of A1 will be different.]
At temperature T2, the change of attenuation will be approximately
(T2-T1) x 0.02 x A1 dB.
At 11GHz and temperature T1, your 5 foot of coax will have an
attenuation of A2 dB.
[Again, for a different frequency / temperature / length / type of coax,
different value of A1.]
At temperature T2, the change of attenuation will be approximately
(T2-T1) x 0.02 x A2 dB.
Essentially, you don't have to know anything about the frequency /
length / type of coax. Although these do determine the amount of initial
attenuation, all you need to know is the actual value of the initial
attenuation, and the change of temperature. The higher the initial
attenuation (for whatever reason), the greater will be the change of
attenuation.
You still need to know the length. I used to have to do the
calculations for CATV system design on a four banger calculator, before
PCs were common. Levels at the highest & lowest temperature was one
limiting factor of the amplifier spacings. Some cables that were almost
identical at first glance in the catalogs could be 100 feet difference
in maximum length. At that time, pressurized, fused disk was the best
.750 cable available, but it was so fragile we wouldn't use it. Then you
had to wade through all the various foam types, the DC resistance per
100 feet, maximum certified frequency and a dozen other numbers.
A 5000 foot cable will have twice the change of a 2500 foot cable
forthe same temperature change.
You still haven't stated if your .02 dB
change/degree is per fook, per 100 feet, or per mile.