Using Circular Error Probable to charcterize rifle performance

I don't think anyone is suggesting changing how scores are calculated. Scores represent how you did on that particular day -- as they should.

The whole point of using CEP or any other way of doing a more sophisiticated analysis is to do exactly what bsl135 has said:

It's relevant because precision measurement is valuable! If you really want to know how much effect variable 'x' has on your precision, you have some choices for how to investigate that variable.

C-T-C doesn't provide the same level of accuracy in doing that investigation as CEP. If someone just wants to burn through ammunition and barrels and has the money, then good for them, but if you can get more accurate information with fewer shots I'd think that would be a good thing.

As far as "Competitive Benchrest, where every shot counts" goes, it is exactly the very high level of performance and the corresponding slim difference between top competitors which is why more sensitive methods of determining equipment performance is needed.

Restating another excellent point made by bsl135, if none of this is true then why bother at all with chronographs?
 
OK...

As much as anything it's the "C-T-C" thing that's bothering me......exactly HOW is this flawed?

The center of a bullet hole is an infinitely small POINT, there AIN'T NO better place to measure from/to. And it's absolutely dead-nuts consistent regardless of caliber, humidity, target composition, lunar phase etc...... Why would anyone have a problem with center to center? And a bunch of these posts DO? Bryan does say "measure c-t-c AND apply CEP" but what the hey??? Then all CEP becomes is another way of fudging the numbers.

It's like the age-old "SD vs ES" argument. The math geeks will always argue that SD results in "more accurate results" but ONE dropped shot due to bad ES will most certainly ruin your day in competition.

It's the whole "more accurate way to measure" thing that gets me going..... CEP and SD simply produce averages which correspond "more nearly to the mean"..... SO FREAKIN' WHAT?


So this accountant is shooting 1K......

The first group is 56" left of the bull, 26" grp.

He dials over and shoots a 30 incher 52" RIGHT of the 10ring........

His group-center vertical is centered up.

"Well, 2" left and dead on for vertical, LOOKS LIKE I WIN!"

:p

al
 
I want to know this....when I shoot a group of .200 or less....how in the heck do you know EXACTLY where each shot landed.

Another thing....things like the size of the single bullet hole, the type of cut in the paper and whether it's verticle or horizontal predominate in the group is way more important than what some damn stastical info is going to give me.

The last registered match I shot...the first morning, 100yd LV my groups went like this "approximatly" .326, .260, .190, .160, .095 The first would have been a mid one if a wind gust/condidtion change hadn't bit me. These were shot at a range I had never been to before, in weather I had never shot in before, with a barrel that had NEVER been shot at 100yds, with bullets I had only used once before in a match and with three different powder charges and two different bullet seating depths. How did the groups improve without statistical data....I looked at the bullet holes in the first target to improve the second to see if seating depth was right, I looked at the verticle in the group, my brain rationalized the wind conditions and I looked at the fired cases. NOW...I want to know exactly what in this program takes all that into consideration because if you don't....you data is only going to tell you one thing....you groups suck...

Sorry but I will stick with the real world and leave the fantasy to the elf's

Hovis
 
(I'm the other McMillan on the paper)

HovisKM: At the risk of starting a flame war, I most respectfully submit that you have not read the entire paper. Please go do so; your criticism will be infinitely more valuable.

That said, the answer to your question lies in shooting at a different (nearby) target for each single shot. You avoid blowing the center of the target out, and are still able to measure each position precisely.

In a more general reply to the thread, I wish to make it clear that we are not advocating changing scoring methods, we do not say that c-t-c is useless, and we are certainly not minimizing any other aspect of the skill involved in the handling of a rifle. Nowhere do we suggest that this is a method to "convince yourself that your equipment is better than it is."

Bryan hit it on the head earlier - this technique is most useful for comparing things. It can tell you with a lower chance of error than c-t-c whether you know enough about the performance of two loads to say that one is better than the other. Sure, you could shoot 5x5 shot groups, and average them, but the statistics can tell you that one is noticeably better than the other in fewer shots (or that your 5x5 shot groups are not statistically different, so you need more shots to know for sure).

We do not say that it is a method for everyone, and we most certainly are not suggesting taking it to a competition. We merely present it as a potentially useful tool some may wish to take advantage of while developing and tuning their systems.
 
Last edited by a moderator:
I want to know this....when I shoot a group of .200 or less....how in the heck do you know EXACTLY where each shot landed.

This problem exists regardless of whether you use CEP or not.


Another thing....things like the size of the single bullet hole, the type of cut in the paper and whether it's verticle or horizontal predominate in the group is way more important than what some damn stastical info is going to give me.

If the group is offset or has a dominate vertical or horizontal spread that is large enough, you will see that by looking at it. The statistical information lets you measure how much, and if the offset or spread is too small to see, it will still let you measure it. I don't think I'd be comfortable with a gunsmith who set headspacing just by how it looked and who made no measurement whatsoever.


The last registered match I shot...the first morning, 100yd LV my groups went like this "approximatly" .326, .260, .190, .160, .095 The first would have been a mid one if a wind gust/condidtion change hadn't bit me. These were shot at a range I had never been to before, in weather I had never shot in before, with a barrel that had NEVER been shot at 100yds, with bullets I had only used once before in a match and with three different powder charges and two different bullet seating depths. How did the groups improve without statistical data....I looked at the bullet holes in the first target to improve the second to see if seating depth was right, I looked at the verticle in the group, my brain rationalized the wind conditions and I looked at the fired cases. NOW...I want to know exactly what in this program takes all that into consideration because if you don't....you data is only going to tell you one thing....you groups suck...

First, .326, .260, .190, .160, .095 is a set of statistical data. Whether it is "damn stastical info" or just the regular kind I can't say. But, statistical data it is -- a rose by any other name still smells as sweet.

Second, no one has claimed that statistical data is necessary for groups to improve. It's not necessary to use a car necessary to get from North Dakota to Texas, but it's faster than walking. But, it is interesting that the above quote includes a set of statistical data which is used to backup the statement that groups improved without using statistical data. To do this without statistical data would have required posting images of all five groups so that people could somehow argue about them without making any measurements, since that would result in a set of statistical data.

Finally, a set of statistical data, including the set in the above quote, will tell you more than just if your groups "suck". It will tell you how much they "suck", and how much each "sucks" as compared to the others. If the right statistics are used, it will also tell you if and how one "sucks" differently from the others.

Every time a scorer measures a group and calculates a score he has produced a statistic, and a group of scores is a set of statistical data. All that has been presented in the post which started this thread was a way to calculate a set of statistical data which could be used for things like comparing two loads which are too similar to easily distinguish using the common c-t-c statistic.


Sorry but I will stick with the real world and leave the fantasy to the elf's

No one has even suggested that anyone use any kind of statistics they don't want to. The original post was simply to say; Here's some information that might be useful.

No one is forced to read this thread. If someone CHOOSES to read this thread, and then reads the paper referred to in the post then they might have a legitimate criticism of what has been done. They might even just want to say they don't think any of this is worthwhile.

Fine, but why can't it just be left at that and then leave the "elf's" to their "fantasy". "Elf's" like Dr. Franklin Mann were writing about this kind of thing 100 years ago. He even has gunsmithing tools named after him and it hasn't hurt the sport any. So what's the point in getting so worked up about it? Just ignore the thread, move on, and let people who might be interested and might want to ask a question do it without having to be insulted.
 
Charles and Paul,
I've had a chance to spend a bit more time on your paper and have a few questions. First (to clear up what was probably a misconception on my part), the circle that covers 50% of the shots with CEP appears to have its center constrained to be at xbar, ybar, correct? I think I mistakenly interpreted your footnote 1 to mean that the circle could be moved around arbitrarily. Second, ellipticity in Table 3 seems to be different (reciprocal?) than its definition in the sentence below. Further, equations 3.5 and 3.6 seem to support the definition rather than the Table value.
Thanks for all your effort to share this. Looks like what I have been been looking for, but have been too busy (or lazy:eek:) to find/derive for myself.

Cheers,
Keith
 
The key phrase in much of the discussion herein has been "comparison". Whether it be through the use of circular error, or center to center, they are still both statistics used for comparison of either more than one gun, or one gun shot at different times, or under different loads.

Either statistic accomplishes the same thing, i.e. the comparison of multiple events on an apples to apples basis. The effective difference between "CEP" or "C-T-C" is not one of better or worse, or even more accurate. The difference is simply which one is adopted for comparative analysis in a given discipline.

If benchrest shooters had been using CEP, then C-T-C would be an alternative that could be debated.

CEP will use more data in its description of a 5 shot group than C-T-C, which uses only two points. And CEP will in fact show how probable it is that a certain number of those five shots will lie in a certain proximity to one another.

C-T-C will only tell us how tightly those five shots were grouped, as that group is described by the two extremes in location. It is this measurement that we seem to have adopted as our measure of accuracy, rather then CEP, which, while providing more data, describes something else.
 
Corrections to Table 3

Keith -

Thank you for reading the paper carefully. You are right, Table 3 is in error given the definition of ellipticity in the paper. I will get that corrected and get another version of the paper up.

You are correct, the CEP we have shown in this paper ignores the offset of xbar, ybar and it is calculated about xbar, ybar. In the literature, this is referred to as CEP_MPI - CEP about the Mean Point of Impact. There is also well developed thought in the literature to include bias in the CEP calculation. I'll plan to extend our work to include bias in a future paper, however, the main point of the first paper was to work out ways to use CEP to characterize rifles as well as a fairly easy measurement technique.
 
Lower variability for CEP than c-t-c

For those interested (clearly not everyone), the point Bryan makes about lower variability for CEP than c-t-c is shown in the right portion of Figure 1 in the paper. It appears to be true regardless of the number of shots fired.
 
Eratta for CEP paper

This morning I posted a copy of our CEP paper to www.statshooting.com. This corrects the error Keith noted between the definition of ellipticity and Table 3. The updated version of the paper is dated Dec. 6, 2008 and is version 1.01. Many thanks to Keith for finding this error.
 
Ok....I guess I'll let the cat out of the bag as to one thing I was meaning but no one has posted as to how this information is most flawed.

Now assuming this information is to be used by competitive shooters. This is where the problem lays.

Rifle barrels vary between each barrel even of the same maker and require different loads....that's a known fact. Most competitive shooters shoot in varying conditions in vary geographical locations. Now by the time this statistical data becomes of use because you have to include data of different altitudes, humidity, tempertures, and locations. The barrel will be long wore out before you get enough information to make it useful for that barrel and the places a person shoots.

That is why I included group sizes in the previous post. Sure it created statistical data but that data is only good for that day at that moment and anything outside of that...it's just a reference with large tolerances. So between matches...while cleaning and adjusting loads...are we suppose to carry a computer around to figure this and then decide what to do?? It's just not pratical. In the competitive world...data that is not up to the half hour to hour current is pretty useless other than being general, it's just not precise.

Just don't see the usefullness in it...that's all and my opinion. Flame away.....I have thick skin....developed by years of experience....not a computer.

Hovis
 
I was going to comment on how the point made above about variability and how it also relates to how group size (as measured with c-t-c, aka extreme spread, ES) is seen to increase with the number of shots, but maybe I'll get back to that later.

I think Hovis raises some very important questions in his last post about problems of doing and using statistical analysis of target groups. There are definitely limitations and some "costs" and these have to be balanced for the circumstances and the goals of the individual. I agree that for some purposes the costs outweigh the benefits and it's not worth it. Also, the results of any statistical analysis are only valid for certain circumstances. So I'm not trying to say here that the points raised are wrong, but just something about where the limits of applying this kind of analysis are as a way of trying to answer some of the questions raised.

I apologize for this being a rather long post, but the questions raised are good and as often with good questions, the answers can be long -- maybe someone else could provide shorter and better written answers.

When raising CEP or any other form of statistical analysis which involves something other than what is commonly done I tend to point out the advantages which may make it seem like I'm saying c-t-c is "wrong" or useless -- that's not what I mean to say. It is a legitimate tool for certain purposes and not for others. Just like a hammer is good for driving nails, but for driving machine screws, a screwdriver is a better tool.

One of the biggest misunderstandings made in using statistical results is in distinguishing between predictions and descriptions. Descriptions are easy. The group sizes (or scores) for a target in a competition are descriptive. They are meant to say how the shooter did right then and there. It doesn't matter if he had a lucky day, an unlucky day, or his average (good or bad) day.

Predictions are different. In that case you are saying what will happen in the future. So for example people often quote a group size as a measure of how accurate a rifle is. The implication is that that group size is somehow representative of the accuracy you can expect from the rifle IN THE FUTURE. Is it a 1 MOA rifle, 0.8 MOA, etc.? As Hovis correctly points out, varying conditions will ultimately affect the accuracy achieved in any given circumstances so predictions are really limited to certain circumstances.

You have to know how barrel condition (wear, cleanliness, temperature, etc.) can be expected to affect the predictions, as well as different loads, weather, etc. In principle measurements could all be made of how these variables would affect accuracy of the equipment or the equipment and shooter, but in practice it might take forever. So while possible in principle, it is impossible in practice. But, if you limit the conditions, the predictions become useful. So if someone says rifle X is a 1 MOA rifle and rifle Y is a 3 MOA rifle, you take it to mean that under conditions reasonably similar to how they were tested, you can expect X to be considerably more accurate than Y. Knowing exactly what conditions to accept as "reasonably similar" depends on your knowledge and experience of equipment and with shooting. It would be foolish, and I haven't seen, and hope no one would suggest, that statistical results be thought of as something to take the place of knowledge and experience. Without the knowledge and experience to say what is "reasonably similar" in the first place, the statistical results aren't much use.

So "reasonably similar" puts a pretty strong limitation on how you can use a statistical analysis. But still people find it useful to say things like this is a 1 MOA rifle. Some may not care if it is 0.8 MOA or 1 MOA or 1.2 MOA, as long as it's in that range. In that case it is probably true that a more sophisticated way of doing a statistical analysis is a waste of time. On the other hand, there are constantly discussions on these forums about some new way of modifying a rifle, or some new attachment which will squeeze just a little more accuracy out of it. For someone interested in that, this kind of analysis can be very useful and I'll get back to that below.

Consider that 1 MOA rifle again, and let's say it was characterized by shooting 5 5-shot groups. Why five? Because people know that even under effectively identical conditions each group will be different in the placement of each individual shot and the variation from group to group needs to be "averaged" out to get a "representative" result. (Why not four or six? This is a qustion for another post.) Suppose that the next time the rifle is fired it is under "reasonably similar" conditions. Will the next shot fall exactly 1 MOA from the point of aim? Will the next five all fall within 1 MOA of the POI? Maybe, maybe not. There will be some variation from shot to shot no matter how close the conditions are to when the rifle was characterized. This natural variation that everyone knows is there can be described by statistics. The 1 MOA number is one statistic. It is commonly calculated by c-t-c (extreme spread). The problem is that when making a PREDICTION there is a tolerance (also called an uncertainty) in the this 1 MOA number, and here is where a big part of the complication comes in. When you go from talking about using statistical results for more than "that day at that moment" (that is, a description) to talking about later (that is, a prediction) that uncertainty becomes critical in knowing how much trust you can have in the prediction.

Suppose the whole characterization of the rifle was done over again under reasonably identical conditions. (So we assume the barrel was broken in and that the wear in the barrel from 25 shots isn't significant.) This time we might get 1.2 MOA. A third time 0.9 MOA. If we kept testing, we might find that 68% of the tests gave us a result of 1 +/- 0.2 MOA and 95% gave us 1 +/- 0.4 MOA. And maybe after all of this testing we would start to see the effect of wear on the barrel and so our results which were true for the original rifle are no longer of much use.

But, the whole mathematics of statistics and probability has been developed and TESTED over a few hundred years in solving this kind of problem. So by doing a more complete statistical analysis on the original 5 x 5 group data, you could get the 1 +/- 0.2 MOA result without having to shoot so many shots that you actually begin to wear the barrel.

Certainly for a lot of people the +/- 0.2 MOA information wouldn't be worth the extra trouble of doing the statistical calculations. But for some it might, and it might explain some of the "flyers" coming from a "1 MOA" rifle.



Now to the question of making a modification or adding some kind of "device" to a rifle to make it just a little more accurate. I'm not saying ANYTHING about the merits of any particular modification or device. These forums are full of heated arguments over such things. But, what a more complete statistical analysis does allow is a way of measuring whether any particular device provides an improvement, and if so, how much.

The common c-t-c (extreme spread) statistic for group size is very nice in that it can be done simply and quickly with minimal calculation and it's very easy to visualize. Suppose that we use a 5 x 5 set of shots and get a result of 1 MOA. Now we haven't calculated the uncertainty in that number but it's still there. Next we add our device, follow the same procedure and get 0.9 MOA. Again, we haven't calculated an uncertainty, but it's still there. So did the device really produce an improvement? Maybe, or maybe it was just the random variation we know exists from one set of shots to the next. This maybe is one reason some of the arguments about this kind of thing can go on forever and get so heated.

So here is where a more sophisticated analysis (such as CEP) than c-t-c is the right tool for the job. With the right set of statistics, but using the same data, the uncertainty in the predicted results can be calculated. So in the above case, you might find that without the device you could expect the rifle to average 1.1 +/- 0.2 MOA groups, and with it 1.0 +/- 0.2 MOA. The difference is so small that it might just be random chance. This tells you that to say with more certainty you have to shoot a few more groups to "average out" more of the uncertainty. Then you might get 1.1 +/- 0.1 and 0.9 +/- 0.1 and you could say with very good certainty that the device really made a difference. On the other hand, if the device didn't really do anything then you might get 1.1 +/- 0.1 MOA and 1.0 +/- 0.1 MOA and again it looks like the improvement is either less than about 0.1 MOA or just the result of random variation in shot groups.

As far as what "very good certainty" means, it's possible to calculate what that means in terms of a specific number using statistics and probability.

The down-side compared to the c-t-c statistic is that the calculations are long and tedious enough that doing them without a computer is really painful. Also, you lose the simplicity of being able to easily visulize what how the statistic is calculated. However, with a little experience, you can develop a "feel" for the more complicated statistics. Also, the simplicity of the c-t-c statistic also tends to give people a false sense of confidence in what it actually can predict.

The particular numbers used here are just to illustrate the idea, they aren't meant to say anything about any particular real device. But the point is that people spend a lot of time and money on trying to squeeze out a little more accuracy and the arguments over what works and doesn't never end. By using some more complete statistics, such as CEP, then it's possible to make more precise measurements so that it could be seen what works and what doesn't with out being mislead by the random variations from one target to the next.

I agree it wouldn't make sense to carry a computer around to try to figure out what to do during a match. The place it seems it would be useful would be in trying to improve AVERAGE performance of equipment and shooter so that you could expect to do better on average. That might even be more than I should say. Probably the best thing to say would be that using more complete statistical analysis can help characterize performance more accurately and precisely. Whether that is useful or not depends on what the individual is trying to achieve.
 
Awsome!

I don't think anyone could have said it better (with more or less words:))
Thanks for taking the time to give such an in depth explanation.
-Bryan
 
Most shooters' time and resources are limited, and given what we know about the relative importance of such things as trigger time to develop judgment in reading wind flags, and gain experience in keeping in tune, some things have to be weeded out of our “to do” lists based on our best guesses as to the likelihood of their increasing performance, rational prioritization if you will. There is not enough time to do everything that might be beneficial, so we have to make choices. IMO, for the vast majority of us, this is one of those weeds. I have yet to see a report that says, this is what I did, and this is why I think it worked better.
 
Back
Top