CI Calibration: Difference between revisions

From RadiWiki
Jump to navigation Jump to search
No edit summary
 
(17 intermediate revisions by 3 users not shown)
Line 4: Line 4:
Let say you want to calibrate 10 mA (20 dbmA), with a tolerance of 0,5 dB
Let say you want to calibrate 10 mA (20 dbmA), with a tolerance of 0,5 dB
This would mean that all generated levels between 10 mA and 10,59254 mA are accepted as correct values.
This would mean that all generated levels between 10 mA and 10,59254 mA are accepted as correct values.
The first thing you may wonder, why no fill in 0dB? This is possible, but will generate an large amount of warning because the desired levels can not be reached.
The first thing you may wonder, why not just fill in 0dB? This is possible, but will generate a large amount of warnings, because the desired levels can not be reached.
0 dB tolerance is only possible in an ideal world, and unfortunately our measuring equipment is not ideal.
0 dB tolerance is only possible in an ideal world, and unfortunately our measuring equipment is not ideal.
So what is the correct setting? this is depending on your needs, it is always a trade off between measuring speed and accuracy.
So, what is the correct setting? This is depending on your needs, it is always a trade off between measuring speed and accuracy.
When the desired level has been reached, this level is stored inside the calibration file.
When the desired level has been reached, this level is stored inside the calibration file.
Meaning, if 10 watts is generating 11 mA, then 11 mA is stored together with the 10 watts.
Meaning, if 10 watts is generating 11 mA, then 11 mA is stored together with the 10 watts.
But if I do a substitution test? will it reproduce the 11 mA again if I set it to 10 mA?
But if I do a substitution test, will it reproduce the 11 mA again if I set it to 10 mA?
The answer is no, RadiMation will calculate power needed to achieve 10 mA and level to this.
The answer is no. RadiMation will calculate the power needed to achieve 10 mA and level to this value.
So in theory you can calibrate on 1 mA and perform testing at 100 mA, in practice you calibrate on 2 or 3 levels.
In theory you can calibrate on 1 mA and perform testing at 100 mA, in practice you calibrate on 2 or 3 levels. For example 10 and 100 mA.
for example 10 and 100 mA.


= Insertion loss of the 150 to 50 Ω adapter =
= Insertion loss of the 150 {{Ohm}} to 50 {{Ohm}} adapter =


[[Image:150 ohm to 50 ohm adapter.jpg]]
[[Image:150 ohm to 50 ohm adapter.jpg]]


The loss between the voltage on the 150 Ω port to the 50 Ω port is 9,5 dB.
The loss between the voltage on the 150 {{Ohm}} port to the 50 {{Ohm}} port is approximately 9,5 dB.


How do you calculate this?
Assuming that the voltage on the 150 {{Ohm}} port is 1 V, (which corresponds to 0 dBV), the voltage over the 50 {{Ohm}} port is a third of the voltage of the 150 {{Ohm}} port.


The easiest way is by example.
<math> U_{50 \Omega} = 1 V*{\frac{50}{100 + 50}} = \frac{1}{3}\ V = 20*\log10({\frac{1}{3}})\ dBV = -9,542425094\ dBV \approx -9,5\ dBV</math>


We assume that the voltage on the 150 Ω port is 1 V, we chose this because this is 0 dBV and makes it easy to calculate.
So, everything that is measured with a 50 {{Ohm}} power sensor, is approximately 9,5 dB lower then the actual value.


The voltage over the 50 Ω port is 1 / 3 of the voltage of the 150 Ω port.
= Calculating the power meter reading back to Vrms in the Calibration =
in formula:


<math> U_{50 \Omega} = 1 V*{\frac{50}{100 + 50}} = \frac{1}{3}\ V</math>
[[Image:Setup cdn bci calibration.gif]]


And 1/3 V is about -9,5 dBV


in formula:
With the above picture of the calibration set up we will calculate the Vrms on the CDN.


<math>U_{50 \Omega} = 20*log10(1/3) \approx -9,542425094 \ dBV </math>
In this example the power meter reading is -2.6 dBm.


<math> -2.6 \ dBm = 10^{\frac{-2.6}{10}} \ mW \approx 0.55\ mW = 0,00055\ W</math>
From power to voltage:
<math>P=\frac{U^2}{R} \to U_{rms} =\sqrt{P*R} \to U_{rms} = \sqrt{0,00055*50} \to U_{rms}\approx 0.18 \ V</math>
<math> dBU_{rms} = 20 * log^{10}(0.18) \approx -14.8 \ dBVrms</math>
With the known attenuation of 30 dB of the attenuator and the 15,6 dB of the 150 to 50 ohm network and e.m.f. value. The actual reading is +30.8 dBVrms.
'''Note:'''
The factor 6 (15,6 dB) arises from the e.m.f. value specified for the test level.
The matched load level is half the e.m.f. level and the further 3:1 voltage division is caused by the 150 Ω to 50 Ω adapter terminated by the 50 Ω measuring equipment.


[[Category:RadiMation]]
[[Category:RadiMation]]
[[Category:Calibration]]
[[Category:Calibration]]

Latest revision as of 11:09, 15 August 2013

Tolerance[edit]

The tolerance in the CI calibration window gives you more control over the calibration. Let say you want to calibrate 10 mA (20 dbmA), with a tolerance of 0,5 dB This would mean that all generated levels between 10 mA and 10,59254 mA are accepted as correct values. The first thing you may wonder, why not just fill in 0dB? This is possible, but will generate a large amount of warnings, because the desired levels can not be reached. 0 dB tolerance is only possible in an ideal world, and unfortunately our measuring equipment is not ideal. So, what is the correct setting? This is depending on your needs, it is always a trade off between measuring speed and accuracy. When the desired level has been reached, this level is stored inside the calibration file. Meaning, if 10 watts is generating 11 mA, then 11 mA is stored together with the 10 watts. But if I do a substitution test, will it reproduce the 11 mA again if I set it to 10 mA? The answer is no. RadiMation will calculate the power needed to achieve 10 mA and level to this value. In theory you can calibrate on 1 mA and perform testing at 100 mA, in practice you calibrate on 2 or 3 levels. For example 10 and 100 mA.

Insertion loss of the 150 Ω to 50 Ω adapter[edit]

150 ohm to 50 ohm adapter.jpg

The loss between the voltage on the 150 Ω port to the 50 Ω port is approximately 9,5 dB.

Assuming that the voltage on the 150 Ω port is 1 V, (which corresponds to 0 dBV), the voltage over the 50 Ω port is a third of the voltage of the 150 Ω port.

So, everything that is measured with a 50 Ω power sensor, is approximately 9,5 dB lower then the actual value.

Calculating the power meter reading back to Vrms in the Calibration[edit]

Setup cdn bci calibration.gif


With the above picture of the calibration set up we will calculate the Vrms on the CDN.

In this example the power meter reading is -2.6 dBm.

From power to voltage:

With the known attenuation of 30 dB of the attenuator and the 15,6 dB of the 150 to 50 ohm network and e.m.f. value. The actual reading is +30.8 dBVrms.

Note:
The factor 6 (15,6 dB) arises from the e.m.f. value specified for the test level. 
The matched load level is half the e.m.f. level and the further 3:1 voltage division is caused by the 150 Ω to 50 Ω adapter terminated by the 50 Ω measuring equipment.