significant figures in averaging samples












1












$begingroup$


I can't seem to find anything about this but I thought that for every 10 samples (of the same thing) that you averaged together you gained 1 significant figure. You'd maybe need 100 samples to gain 2 figures.



I'm talking about the context where maybe a computer or pulse generator is firing a pulse laser in a loop and you're looking at a change in absorbance or fluorescence with a photodiode or photomultiplier tube. You see a trace on a digital storage scope and you average hundreds or thousands of them for improved accuracy.



I used to work for chemists, but my job was electronics to their specifications. I built the amplifier for the photodiode, interfaced the storage scope to computers. I know improving accuracy was the purpose of averaging, don't remember how the significant figures worked out. It was also 30+ years ago.










share|cite|improve this question









$endgroup$












  • $begingroup$
    Hi & welcome to MSE. I'm not an expert in statistics, so maybe somebody else can give you a more definitive response. I don't think there's any specific number of values to average to get an extra significant digit of accuracy. It depends a lot on how much variance there is in the data. If it's all very close to each other, then even several may work. However, if for example, there is a $1000$ times variation between the minimum and maximum values, with you also getting everything in between, it'll require many more values for you to have reasonable confidence in an extra digit of accuracy.
    $endgroup$
    – John Omielan
    Jan 13 at 1:30










  • $begingroup$
    No, nothing like 1000 times variance, this was more like a signal to noise ratio thing. Ideally the samples would have been the same but there were various error sources that only averaging could really hope to diminish.
    $endgroup$
    – Alan Corey
    Jan 13 at 1:48










  • $begingroup$
    I was just stating $1000$ to give emphasis to how important the variance can be. As for it being the "signal to noise ratio" thing (as well as any other error sources), the answer to your question will depend largely on the range & variance of the errors you may reasonably expect to encounter. This is something that obviously depends on your particular situation, such as the type of machine you're using, so there's something you should try to determine, or at least get a rough idea of them. Somebody here can likely give you a reasonable formula of some sort to use once you have these values.
    $endgroup$
    – John Omielan
    Jan 13 at 1:53












  • $begingroup$
    Welcome to Mathematics Stack Exchange community! The quick tour will help you get the most benefit from your time here.
    $endgroup$
    – dantopa
    Jan 13 at 2:28










  • $begingroup$
    Actually what got me started here is that I monitor outdoor temperatures with an Oregon Scientific thermometer but also record the values transmitted over radio on a computer. In 2.5 years I've gotten 1.5 million data points. If I say the mean temperature for a given month was x, how should I limit the precision in my printf? By default I see 6 places. It's about 1 data point per minute.
    $endgroup$
    – Alan Corey
    Jan 13 at 13:49
















1












$begingroup$


I can't seem to find anything about this but I thought that for every 10 samples (of the same thing) that you averaged together you gained 1 significant figure. You'd maybe need 100 samples to gain 2 figures.



I'm talking about the context where maybe a computer or pulse generator is firing a pulse laser in a loop and you're looking at a change in absorbance or fluorescence with a photodiode or photomultiplier tube. You see a trace on a digital storage scope and you average hundreds or thousands of them for improved accuracy.



I used to work for chemists, but my job was electronics to their specifications. I built the amplifier for the photodiode, interfaced the storage scope to computers. I know improving accuracy was the purpose of averaging, don't remember how the significant figures worked out. It was also 30+ years ago.










share|cite|improve this question









$endgroup$












  • $begingroup$
    Hi & welcome to MSE. I'm not an expert in statistics, so maybe somebody else can give you a more definitive response. I don't think there's any specific number of values to average to get an extra significant digit of accuracy. It depends a lot on how much variance there is in the data. If it's all very close to each other, then even several may work. However, if for example, there is a $1000$ times variation between the minimum and maximum values, with you also getting everything in between, it'll require many more values for you to have reasonable confidence in an extra digit of accuracy.
    $endgroup$
    – John Omielan
    Jan 13 at 1:30










  • $begingroup$
    No, nothing like 1000 times variance, this was more like a signal to noise ratio thing. Ideally the samples would have been the same but there were various error sources that only averaging could really hope to diminish.
    $endgroup$
    – Alan Corey
    Jan 13 at 1:48










  • $begingroup$
    I was just stating $1000$ to give emphasis to how important the variance can be. As for it being the "signal to noise ratio" thing (as well as any other error sources), the answer to your question will depend largely on the range & variance of the errors you may reasonably expect to encounter. This is something that obviously depends on your particular situation, such as the type of machine you're using, so there's something you should try to determine, or at least get a rough idea of them. Somebody here can likely give you a reasonable formula of some sort to use once you have these values.
    $endgroup$
    – John Omielan
    Jan 13 at 1:53












  • $begingroup$
    Welcome to Mathematics Stack Exchange community! The quick tour will help you get the most benefit from your time here.
    $endgroup$
    – dantopa
    Jan 13 at 2:28










  • $begingroup$
    Actually what got me started here is that I monitor outdoor temperatures with an Oregon Scientific thermometer but also record the values transmitted over radio on a computer. In 2.5 years I've gotten 1.5 million data points. If I say the mean temperature for a given month was x, how should I limit the precision in my printf? By default I see 6 places. It's about 1 data point per minute.
    $endgroup$
    – Alan Corey
    Jan 13 at 13:49














1












1








1





$begingroup$


I can't seem to find anything about this but I thought that for every 10 samples (of the same thing) that you averaged together you gained 1 significant figure. You'd maybe need 100 samples to gain 2 figures.



I'm talking about the context where maybe a computer or pulse generator is firing a pulse laser in a loop and you're looking at a change in absorbance or fluorescence with a photodiode or photomultiplier tube. You see a trace on a digital storage scope and you average hundreds or thousands of them for improved accuracy.



I used to work for chemists, but my job was electronics to their specifications. I built the amplifier for the photodiode, interfaced the storage scope to computers. I know improving accuracy was the purpose of averaging, don't remember how the significant figures worked out. It was also 30+ years ago.










share|cite|improve this question









$endgroup$




I can't seem to find anything about this but I thought that for every 10 samples (of the same thing) that you averaged together you gained 1 significant figure. You'd maybe need 100 samples to gain 2 figures.



I'm talking about the context where maybe a computer or pulse generator is firing a pulse laser in a loop and you're looking at a change in absorbance or fluorescence with a photodiode or photomultiplier tube. You see a trace on a digital storage scope and you average hundreds or thousands of them for improved accuracy.



I used to work for chemists, but my job was electronics to their specifications. I built the amplifier for the photodiode, interfaced the storage scope to computers. I know improving accuracy was the purpose of averaging, don't remember how the significant figures worked out. It was also 30+ years ago.







average significant-figures






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Jan 13 at 1:21









Alan CoreyAlan Corey

61




61












  • $begingroup$
    Hi & welcome to MSE. I'm not an expert in statistics, so maybe somebody else can give you a more definitive response. I don't think there's any specific number of values to average to get an extra significant digit of accuracy. It depends a lot on how much variance there is in the data. If it's all very close to each other, then even several may work. However, if for example, there is a $1000$ times variation between the minimum and maximum values, with you also getting everything in between, it'll require many more values for you to have reasonable confidence in an extra digit of accuracy.
    $endgroup$
    – John Omielan
    Jan 13 at 1:30










  • $begingroup$
    No, nothing like 1000 times variance, this was more like a signal to noise ratio thing. Ideally the samples would have been the same but there were various error sources that only averaging could really hope to diminish.
    $endgroup$
    – Alan Corey
    Jan 13 at 1:48










  • $begingroup$
    I was just stating $1000$ to give emphasis to how important the variance can be. As for it being the "signal to noise ratio" thing (as well as any other error sources), the answer to your question will depend largely on the range & variance of the errors you may reasonably expect to encounter. This is something that obviously depends on your particular situation, such as the type of machine you're using, so there's something you should try to determine, or at least get a rough idea of them. Somebody here can likely give you a reasonable formula of some sort to use once you have these values.
    $endgroup$
    – John Omielan
    Jan 13 at 1:53












  • $begingroup$
    Welcome to Mathematics Stack Exchange community! The quick tour will help you get the most benefit from your time here.
    $endgroup$
    – dantopa
    Jan 13 at 2:28










  • $begingroup$
    Actually what got me started here is that I monitor outdoor temperatures with an Oregon Scientific thermometer but also record the values transmitted over radio on a computer. In 2.5 years I've gotten 1.5 million data points. If I say the mean temperature for a given month was x, how should I limit the precision in my printf? By default I see 6 places. It's about 1 data point per minute.
    $endgroup$
    – Alan Corey
    Jan 13 at 13:49


















  • $begingroup$
    Hi & welcome to MSE. I'm not an expert in statistics, so maybe somebody else can give you a more definitive response. I don't think there's any specific number of values to average to get an extra significant digit of accuracy. It depends a lot on how much variance there is in the data. If it's all very close to each other, then even several may work. However, if for example, there is a $1000$ times variation between the minimum and maximum values, with you also getting everything in between, it'll require many more values for you to have reasonable confidence in an extra digit of accuracy.
    $endgroup$
    – John Omielan
    Jan 13 at 1:30










  • $begingroup$
    No, nothing like 1000 times variance, this was more like a signal to noise ratio thing. Ideally the samples would have been the same but there were various error sources that only averaging could really hope to diminish.
    $endgroup$
    – Alan Corey
    Jan 13 at 1:48










  • $begingroup$
    I was just stating $1000$ to give emphasis to how important the variance can be. As for it being the "signal to noise ratio" thing (as well as any other error sources), the answer to your question will depend largely on the range & variance of the errors you may reasonably expect to encounter. This is something that obviously depends on your particular situation, such as the type of machine you're using, so there's something you should try to determine, or at least get a rough idea of them. Somebody here can likely give you a reasonable formula of some sort to use once you have these values.
    $endgroup$
    – John Omielan
    Jan 13 at 1:53












  • $begingroup$
    Welcome to Mathematics Stack Exchange community! The quick tour will help you get the most benefit from your time here.
    $endgroup$
    – dantopa
    Jan 13 at 2:28










  • $begingroup$
    Actually what got me started here is that I monitor outdoor temperatures with an Oregon Scientific thermometer but also record the values transmitted over radio on a computer. In 2.5 years I've gotten 1.5 million data points. If I say the mean temperature for a given month was x, how should I limit the precision in my printf? By default I see 6 places. It's about 1 data point per minute.
    $endgroup$
    – Alan Corey
    Jan 13 at 13:49
















$begingroup$
Hi & welcome to MSE. I'm not an expert in statistics, so maybe somebody else can give you a more definitive response. I don't think there's any specific number of values to average to get an extra significant digit of accuracy. It depends a lot on how much variance there is in the data. If it's all very close to each other, then even several may work. However, if for example, there is a $1000$ times variation between the minimum and maximum values, with you also getting everything in between, it'll require many more values for you to have reasonable confidence in an extra digit of accuracy.
$endgroup$
– John Omielan
Jan 13 at 1:30




$begingroup$
Hi & welcome to MSE. I'm not an expert in statistics, so maybe somebody else can give you a more definitive response. I don't think there's any specific number of values to average to get an extra significant digit of accuracy. It depends a lot on how much variance there is in the data. If it's all very close to each other, then even several may work. However, if for example, there is a $1000$ times variation between the minimum and maximum values, with you also getting everything in between, it'll require many more values for you to have reasonable confidence in an extra digit of accuracy.
$endgroup$
– John Omielan
Jan 13 at 1:30












$begingroup$
No, nothing like 1000 times variance, this was more like a signal to noise ratio thing. Ideally the samples would have been the same but there were various error sources that only averaging could really hope to diminish.
$endgroup$
– Alan Corey
Jan 13 at 1:48




$begingroup$
No, nothing like 1000 times variance, this was more like a signal to noise ratio thing. Ideally the samples would have been the same but there were various error sources that only averaging could really hope to diminish.
$endgroup$
– Alan Corey
Jan 13 at 1:48












$begingroup$
I was just stating $1000$ to give emphasis to how important the variance can be. As for it being the "signal to noise ratio" thing (as well as any other error sources), the answer to your question will depend largely on the range & variance of the errors you may reasonably expect to encounter. This is something that obviously depends on your particular situation, such as the type of machine you're using, so there's something you should try to determine, or at least get a rough idea of them. Somebody here can likely give you a reasonable formula of some sort to use once you have these values.
$endgroup$
– John Omielan
Jan 13 at 1:53






$begingroup$
I was just stating $1000$ to give emphasis to how important the variance can be. As for it being the "signal to noise ratio" thing (as well as any other error sources), the answer to your question will depend largely on the range & variance of the errors you may reasonably expect to encounter. This is something that obviously depends on your particular situation, such as the type of machine you're using, so there's something you should try to determine, or at least get a rough idea of them. Somebody here can likely give you a reasonable formula of some sort to use once you have these values.
$endgroup$
– John Omielan
Jan 13 at 1:53














$begingroup$
Welcome to Mathematics Stack Exchange community! The quick tour will help you get the most benefit from your time here.
$endgroup$
– dantopa
Jan 13 at 2:28




$begingroup$
Welcome to Mathematics Stack Exchange community! The quick tour will help you get the most benefit from your time here.
$endgroup$
– dantopa
Jan 13 at 2:28












$begingroup$
Actually what got me started here is that I monitor outdoor temperatures with an Oregon Scientific thermometer but also record the values transmitted over radio on a computer. In 2.5 years I've gotten 1.5 million data points. If I say the mean temperature for a given month was x, how should I limit the precision in my printf? By default I see 6 places. It's about 1 data point per minute.
$endgroup$
– Alan Corey
Jan 13 at 13:49




$begingroup$
Actually what got me started here is that I monitor outdoor temperatures with an Oregon Scientific thermometer but also record the values transmitted over radio on a computer. In 2.5 years I've gotten 1.5 million data points. If I say the mean temperature for a given month was x, how should I limit the precision in my printf? By default I see 6 places. It's about 1 data point per minute.
$endgroup$
– Alan Corey
Jan 13 at 13:49










1 Answer
1






active

oldest

votes


















0












$begingroup$

I think the statistical fact you need to know in this case is that if $X_1, X_2, X_3, dots , X_n$ are independent, identically distributed random variables, each with standard deviation $sigma$, then the standard deviation of their mean, $overline{X}$, is $sigma / sqrt{n}$. So if you have some estimate of the error in each of your $n$ temperature readings, then just divide that error estimate by $sqrt{n}$ to find the error in the average.



In terms of significant digits, this means that in order to get one more significant digit in your results, you need 100 times as many samples as in your original data. If you want two more significant digits, you need to increase your number of samples by a factor of 10,000.



All this assumes that your readings are in fact independent and unbiased. If that assumption is false, all bets are off.



Reference: An Introduction to Error Analysis: the Study of Uncertainties in Physical Measurements, Second Edition by John R. Taylor.






share|cite|improve this answer











$endgroup$













  • $begingroup$
    OK, there's a mishmash of concepts here too. I get at most 3 significant figures (I've seen 21.6 C) but the A/.D converter in the sensor works in bits of resolution. Natively it does degrees C but 21.6 is probably 216 counts. I doubt it's an 8-bit A/D but it could be 10.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:21










  • $begingroup$
    Looking at last year's data I see 509966 samples for 12 months so each month is ~42497. I guess I'll set the printf to %.3f and keep 3 digits after the decimal. I'm partly looking at how difficult it is to see global warming. You sort of have to know it's there, it doesn't jump out at you. But I've only got 2.5 years of data too.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:29










  • $begingroup$
    For anybody that might want to do this, get an Oregon Scientific sensor and an RTL2832 dongle. Download rtl_433 at github.com/merbanan/rtl_433. It runs on a Raspberry Pi, maybe a Zero, so you can do this for well under $50.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:34











Your Answer





StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3071608%2fsignificant-figures-in-averaging-samples%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























1 Answer
1






active

oldest

votes








1 Answer
1






active

oldest

votes









active

oldest

votes






active

oldest

votes









0












$begingroup$

I think the statistical fact you need to know in this case is that if $X_1, X_2, X_3, dots , X_n$ are independent, identically distributed random variables, each with standard deviation $sigma$, then the standard deviation of their mean, $overline{X}$, is $sigma / sqrt{n}$. So if you have some estimate of the error in each of your $n$ temperature readings, then just divide that error estimate by $sqrt{n}$ to find the error in the average.



In terms of significant digits, this means that in order to get one more significant digit in your results, you need 100 times as many samples as in your original data. If you want two more significant digits, you need to increase your number of samples by a factor of 10,000.



All this assumes that your readings are in fact independent and unbiased. If that assumption is false, all bets are off.



Reference: An Introduction to Error Analysis: the Study of Uncertainties in Physical Measurements, Second Edition by John R. Taylor.






share|cite|improve this answer











$endgroup$













  • $begingroup$
    OK, there's a mishmash of concepts here too. I get at most 3 significant figures (I've seen 21.6 C) but the A/.D converter in the sensor works in bits of resolution. Natively it does degrees C but 21.6 is probably 216 counts. I doubt it's an 8-bit A/D but it could be 10.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:21










  • $begingroup$
    Looking at last year's data I see 509966 samples for 12 months so each month is ~42497. I guess I'll set the printf to %.3f and keep 3 digits after the decimal. I'm partly looking at how difficult it is to see global warming. You sort of have to know it's there, it doesn't jump out at you. But I've only got 2.5 years of data too.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:29










  • $begingroup$
    For anybody that might want to do this, get an Oregon Scientific sensor and an RTL2832 dongle. Download rtl_433 at github.com/merbanan/rtl_433. It runs on a Raspberry Pi, maybe a Zero, so you can do this for well under $50.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:34
















0












$begingroup$

I think the statistical fact you need to know in this case is that if $X_1, X_2, X_3, dots , X_n$ are independent, identically distributed random variables, each with standard deviation $sigma$, then the standard deviation of their mean, $overline{X}$, is $sigma / sqrt{n}$. So if you have some estimate of the error in each of your $n$ temperature readings, then just divide that error estimate by $sqrt{n}$ to find the error in the average.



In terms of significant digits, this means that in order to get one more significant digit in your results, you need 100 times as many samples as in your original data. If you want two more significant digits, you need to increase your number of samples by a factor of 10,000.



All this assumes that your readings are in fact independent and unbiased. If that assumption is false, all bets are off.



Reference: An Introduction to Error Analysis: the Study of Uncertainties in Physical Measurements, Second Edition by John R. Taylor.






share|cite|improve this answer











$endgroup$













  • $begingroup$
    OK, there's a mishmash of concepts here too. I get at most 3 significant figures (I've seen 21.6 C) but the A/.D converter in the sensor works in bits of resolution. Natively it does degrees C but 21.6 is probably 216 counts. I doubt it's an 8-bit A/D but it could be 10.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:21










  • $begingroup$
    Looking at last year's data I see 509966 samples for 12 months so each month is ~42497. I guess I'll set the printf to %.3f and keep 3 digits after the decimal. I'm partly looking at how difficult it is to see global warming. You sort of have to know it's there, it doesn't jump out at you. But I've only got 2.5 years of data too.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:29










  • $begingroup$
    For anybody that might want to do this, get an Oregon Scientific sensor and an RTL2832 dongle. Download rtl_433 at github.com/merbanan/rtl_433. It runs on a Raspberry Pi, maybe a Zero, so you can do this for well under $50.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:34














0












0








0





$begingroup$

I think the statistical fact you need to know in this case is that if $X_1, X_2, X_3, dots , X_n$ are independent, identically distributed random variables, each with standard deviation $sigma$, then the standard deviation of their mean, $overline{X}$, is $sigma / sqrt{n}$. So if you have some estimate of the error in each of your $n$ temperature readings, then just divide that error estimate by $sqrt{n}$ to find the error in the average.



In terms of significant digits, this means that in order to get one more significant digit in your results, you need 100 times as many samples as in your original data. If you want two more significant digits, you need to increase your number of samples by a factor of 10,000.



All this assumes that your readings are in fact independent and unbiased. If that assumption is false, all bets are off.



Reference: An Introduction to Error Analysis: the Study of Uncertainties in Physical Measurements, Second Edition by John R. Taylor.






share|cite|improve this answer











$endgroup$



I think the statistical fact you need to know in this case is that if $X_1, X_2, X_3, dots , X_n$ are independent, identically distributed random variables, each with standard deviation $sigma$, then the standard deviation of their mean, $overline{X}$, is $sigma / sqrt{n}$. So if you have some estimate of the error in each of your $n$ temperature readings, then just divide that error estimate by $sqrt{n}$ to find the error in the average.



In terms of significant digits, this means that in order to get one more significant digit in your results, you need 100 times as many samples as in your original data. If you want two more significant digits, you need to increase your number of samples by a factor of 10,000.



All this assumes that your readings are in fact independent and unbiased. If that assumption is false, all bets are off.



Reference: An Introduction to Error Analysis: the Study of Uncertainties in Physical Measurements, Second Edition by John R. Taylor.







share|cite|improve this answer














share|cite|improve this answer



share|cite|improve this answer








edited Jan 13 at 16:19

























answered Jan 13 at 15:29









awkwardawkward

6,12011022




6,12011022












  • $begingroup$
    OK, there's a mishmash of concepts here too. I get at most 3 significant figures (I've seen 21.6 C) but the A/.D converter in the sensor works in bits of resolution. Natively it does degrees C but 21.6 is probably 216 counts. I doubt it's an 8-bit A/D but it could be 10.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:21










  • $begingroup$
    Looking at last year's data I see 509966 samples for 12 months so each month is ~42497. I guess I'll set the printf to %.3f and keep 3 digits after the decimal. I'm partly looking at how difficult it is to see global warming. You sort of have to know it's there, it doesn't jump out at you. But I've only got 2.5 years of data too.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:29










  • $begingroup$
    For anybody that might want to do this, get an Oregon Scientific sensor and an RTL2832 dongle. Download rtl_433 at github.com/merbanan/rtl_433. It runs on a Raspberry Pi, maybe a Zero, so you can do this for well under $50.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:34


















  • $begingroup$
    OK, there's a mishmash of concepts here too. I get at most 3 significant figures (I've seen 21.6 C) but the A/.D converter in the sensor works in bits of resolution. Natively it does degrees C but 21.6 is probably 216 counts. I doubt it's an 8-bit A/D but it could be 10.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:21










  • $begingroup$
    Looking at last year's data I see 509966 samples for 12 months so each month is ~42497. I guess I'll set the printf to %.3f and keep 3 digits after the decimal. I'm partly looking at how difficult it is to see global warming. You sort of have to know it's there, it doesn't jump out at you. But I've only got 2.5 years of data too.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:29










  • $begingroup$
    For anybody that might want to do this, get an Oregon Scientific sensor and an RTL2832 dongle. Download rtl_433 at github.com/merbanan/rtl_433. It runs on a Raspberry Pi, maybe a Zero, so you can do this for well under $50.
    $endgroup$
    – Alan Corey
    Jan 13 at 17:34
















$begingroup$
OK, there's a mishmash of concepts here too. I get at most 3 significant figures (I've seen 21.6 C) but the A/.D converter in the sensor works in bits of resolution. Natively it does degrees C but 21.6 is probably 216 counts. I doubt it's an 8-bit A/D but it could be 10.
$endgroup$
– Alan Corey
Jan 13 at 17:21




$begingroup$
OK, there's a mishmash of concepts here too. I get at most 3 significant figures (I've seen 21.6 C) but the A/.D converter in the sensor works in bits of resolution. Natively it does degrees C but 21.6 is probably 216 counts. I doubt it's an 8-bit A/D but it could be 10.
$endgroup$
– Alan Corey
Jan 13 at 17:21












$begingroup$
Looking at last year's data I see 509966 samples for 12 months so each month is ~42497. I guess I'll set the printf to %.3f and keep 3 digits after the decimal. I'm partly looking at how difficult it is to see global warming. You sort of have to know it's there, it doesn't jump out at you. But I've only got 2.5 years of data too.
$endgroup$
– Alan Corey
Jan 13 at 17:29




$begingroup$
Looking at last year's data I see 509966 samples for 12 months so each month is ~42497. I guess I'll set the printf to %.3f and keep 3 digits after the decimal. I'm partly looking at how difficult it is to see global warming. You sort of have to know it's there, it doesn't jump out at you. But I've only got 2.5 years of data too.
$endgroup$
– Alan Corey
Jan 13 at 17:29












$begingroup$
For anybody that might want to do this, get an Oregon Scientific sensor and an RTL2832 dongle. Download rtl_433 at github.com/merbanan/rtl_433. It runs on a Raspberry Pi, maybe a Zero, so you can do this for well under $50.
$endgroup$
– Alan Corey
Jan 13 at 17:34




$begingroup$
For anybody that might want to do this, get an Oregon Scientific sensor and an RTL2832 dongle. Download rtl_433 at github.com/merbanan/rtl_433. It runs on a Raspberry Pi, maybe a Zero, so you can do this for well under $50.
$endgroup$
– Alan Corey
Jan 13 at 17:34


















draft saved

draft discarded




















































Thanks for contributing an answer to Mathematics Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3071608%2fsignificant-figures-in-averaging-samples%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

Can a sorcerer learn a 5th-level spell early by creating spell slots using the Font of Magic feature?

Does disintegrating a polymorphed enemy still kill it after the 2018 errata?

A Topological Invariant for $pi_3(U(n))$