This post is about a reductions-related problem that I find mysterious. There are two kinds of reductions analysis currently under consideration.

- Error limiting reductions. Here, the goal is to bound the error rate of the created classifier in terms of the error rate of the binary classifiers that you reduce to. A very simple example of this is that error correcting output codes where it is possible to prove that for certain codes, the multiclass error rate is at most 4 * the binary classifier error rate.
- Regret minimizing reductions. Here, the goal is to bound the
*regret*of the created classifier in terms of the*regret*of the binary classifiers reduced to. The regret is the error rate minus the minimum error rate. When the learning problem is noisy the minimum error rate may not be*0*. An analagous result for reget is that for a probabilistic error correcting output code, multiclass regret is at most 4 * (binary regret)^{0.5}.

The use of “regret” is more desirable than the use of error rates, because (for example) the ECOC error rate bound implies nothing when there is enough noise so that the binary classifiers always have error rate 0.25. However the square root dependence introduced when analyzing regret is *not* desirable. A basic question is: Can we have the best of both worlds? Can we find some algorithm doing multiclass classification with binary classifiers such that average regret *r* for the binary classifiers implies average regret bounded by *4r* for the multiclass classifier?

If the answer is “yes”, that reduction algorithm may be empirically superior to the one we use now.

If the answer is “no”, that is a sharp and unexpected distinction between error rate analysis and regret analysis.