|
Re: Evaluation question |
Hi Roger: Let's say there are 3 outcomes to a prediction: HIT (in all 3 parameters), partial hit (one or more parameters), and a complete miss (no parameters where matched). You sound like you want to address both the partial hit and complete miss cases by some method which allows assigning a "degree of closeness" to the partial hit or miss (for training or evaluation purposes). My suggestion is to expand the range of the 3 stated parameters in turn (location,time window, magnitude) and then calculate the probability for that particular earthquake. For example, someone predicts a 3.3-4.3 quake in a given region and time window. Instead a mag. 5 happens. You could then calculate the probability of the expanded prediction to include the mag. 5. Now a rule with a decision threshold could be used. If the probability of the expanded prediction parameters earthquake was less than some number between 0.0 and 1.0 (say 33% or 25 % or some threshold percentage) then it would be retained as a partial-hit but with the probability of the "captured earthquake" so noted in a scoring column format. For example column headings could be: 1) Orig. Prediction - Time, Location, Mag (T,L,M) If no earthquake can be found nearby in the predicted region of time, location, and magnitude with a probability less than say 33% ( 1 in 3 or 4) then the prediction is ruled a complete miss. For each predictor an accumulated score will develop that can show the accumulated odds that have been showing for Hits, Misses, and Partial hits. The probabilities (all multiplied together over time) of the expanded predictions with the captured earthquakes gives an indication of how close the predictor is (or is not) over time. A sample might be as shown for predictor X ovewr time: 1 - Made 1000 predictions Accumulated odds = 1 / (0.25)**300 Note: A predictor could always adjust the parameters to do 1 in 4 odds predictions or 1 in 20 or whatever they choose. 2 - Using the probability threshold of 33 % to decide HIT or MISS, accumulated 70% Incorrect predictions (note a MISS is defined as an incorrect prediction and not an undetected earhquake that was missed altogether as far as making a prediction goes...). Remaining will be a range of earthquakes with varying probabilities depending on their parameters with regards to the original prediction. Let's say there were 5 earthquakes in expanded predictions with the following probabilities of happening (I can't spell occurrence?): 0.26 (26 %) Clearly the .08 earthquake will make the predictor look good whereas the 0.26 probability case will not. A decision could be made at this point to accumulate (as for the hits cases) either the best or worst (or both) probability case as part of the predictors record of success/failure. The benefit of this approach (assuming I explained it well enough) is that it becomes "geography" independent sort of by noit using "hardwired" ranges per magnitude (and other approaches) since seismicity varies around the globe and the hardwiring will create an evaluation bias IMO. Ok...so to try (again) and create a summary performance for predictor X it would be as follows: 1 - Made 100 predictions This method would quickly sort out whose is onto to something and who is not. Of course the question of handling aftershocks, etc. is another topic of discussion (though one could argue that the probabilities method above could account for this effect (and give low weighting to it automatically). Comments are welcome. Just my 2cents worth...
Follow Ups: ● A clarification or two - 2cents 18:53:31 - 3/12/2002 (13622) (0) |
|