By using this site, you agree to our Privacy Policy and our Terms of Use. Close
DonFerrari said:
Alby_da_Wolf said:
DonFerrari said:
Alby_da_Wolf said:
To people suggesting a number like 370022 for a given platform in a given week be rounded to 370000: it wouldn't be correct. If calculations bring to a result of 370022 ± error margin, rounding it to 370000 ± error margin wouldn't be the same. Look at the first message of this thread, it's explained quite clearly: that 370022 value is the midpoint of the probability curve of the extrapolation of sales data collected that week. BTW most probably a rounding to the closest unit already happened, as sales are integer numbers, start values are integers and final values must be too, but intermediate values very often won't and calculations will be made keeping all the available decimals in every intermediate result, and rounding will happen only at the end to limit the growth of rounding error, that would just be added to the error already present in the extrapolation.
I'm puzzled, I thought that some basic rules for rounding and about measurement errors were taught also in high-school physics courses all around the world, not just at university.

And aparrently you haven't learn them since you want to use more signficance numbers than the tolerance permit.

Just no. When you receive sales data, you collect integer numbers and you aren't introducing yourself an additional measurement error like if you were measuring a length, say, with a 1mm graduated ruler. Those integer numbers can be precise, if the store can give them to you for a given week, or they'll be approximations plus or minus an error margin. You'll take them into account, and you'll have to take into account also the precision available for your internal calculations and the approximation error accumulation, plus obviously your estimate of the error in the formula you devised for the extrapolation of your data but even then, rounding the final result of the central value to an integer different from the closest one (or the closest greater or lower one) would be an error in every case.

Not sure if you know metrology, but you can also decentralize your "mean"... instead of saying 10.0+-0.05 you could say 9.95+0.1. In statistic don't know it would be acceptable... but I never saw a statistic report on election being 22.5+-.5% of intention (and they use integer numbers even tough they state results in percentages), and then they give margin of error as 5 pp (percentual points)... by your logic they should state it as 22.5341% (if that was the exact approximation value on their method on lets say 1430 people polled).

I know you can do it and yes, if you decentralize and you do like you say, that is taking into account the additional error you introduce doing so, then it's correct, but this way you are unnecessarily increasing the error margin, and unnecessarily worsening the error accumulation in the lifetime total. And yes, it also makes sense for immediateness to round to an integer percent value when you communicate a percent result, or to just one decimal, like ioi does in the bar chart in VGC's front page, but you don't want to do the same for data that you'll reuse for further calculations, like adding weekly absolute numbers to get monthly, yearly, LTD and lifetime totals. Then, when using those numbers to draw a curve, rounding those integers to the resolution available will again be correct, but it's correct to do it for FINAL data that you won't reuse. If you'll want to draw another graph, you'll start again from the unrounded values, you won't get them scanning the graph you previously made rounding the values. If you round too much and too early data that must be reused, the error on further results could skyrocket. BTW, rounding to the closest hundred or thousand can be a significant additional percent error for small weekly numbers.
Then there is a method issue: ioi devised his formulas for extrapolation and he needs to constantly refine, update and possibly improve them or at least not let them worsen as the market changes, to achieve it he cannot arbitrarily round numbers, he could do it for some samples if he found that a given source is systematically above or under the average and its error isn't random, but biased up or down, but ouside of these cases, he'll want to keep all his numbers as centered as possible, then he'll only round data that he just presents to the public without reusing them, like the percent values and rounded totals of the front page bar charts. Another thing: real sales numbers are integers, so after applying extrapolation formulas, that will mostly give results with decimals, rounding to the closest integer is right and necessary. But after doing it, an integer like 370022 has the same "dignity" as 370000, rounding it to the closest hundred or thousand presenting the results is just cosmetic, you'll do it regularly if you, for example, present the data as 370k, 0.37M or 0.4M and so on, but again, you won't add these quick representations of weekly totals to calculate yearly totals. The 0.37M example can give you a hint: if you write, for example, 10.1M for a lifetime total, your rounding error is at most ±0.5%, but if you decide to use just ONE decimal and you write 0.1M  for a weekly total that before cosmetic rounding was a given number somewhere between 50000 and 150000, your rounding error can be up to 50%: even when deciding to do cosmetic roundings, you'll have to consider the range of actual values before deciding the acceptable rounding.


EDIT: I completed this post after taking some time and a long pause and in the meantime ioi wrote another comment on this issue, read it too, as he's enormously more skilled than me in the statistics and probability field (I, despite not liking them very much, except for robotics, found that I'm better at automatic controls, and, for example, I have many time had the temptation to represent ioi's adjustments as a linear system with a feedback loop  ).



Stwike him, Centuwion. Stwike him vewy wuffly! (Pontius Pilate, "Life of Brian")
A fart without stink is like a sky without stars.
TGS, Third Grade Shooter: brand new genre invented by Kevin Butler exclusively for Natal WiiToo Kinect. PEW! PEW-PEW-PEW!