question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

RFC classifiers trained by minimizing the Brier loss

See original GitHub issue

At the moment, our probablistic classifiers (e.g. logistic regression and gradient boosted trees) optimize the log loss, typically after taking a sigmoid or softmax inverse link function (typically part of the Cython loss).

However the log-loss is not the only proper-scoring rule to fit estimators of the expected conditional class probabilities, in particular the Brier loss is also a proper scoring rule and has the practical advantage of being upper bounded which should limit the impact of mislabeled examples in the training set.

Would it make sense to publicly such estimators in scikit-learn?

Note that for linear models, fitting the Brier loss is not equivalent to our RidgeClassifier because the latter does not take the softmax of its raw predictions before computing the square loss, hence it does not offer a well defined estimation of the conditional class probabilities.

Minimizing the Brier loss of a softmax linear model is no longer a convex optimization problem but I doubt that this would prevent a newton solver with a robust line search to converge in practice.

Issue Analytics

  • State:closed
  • Created a year ago
  • Comments:10 (10 by maintainers)

github_iconTop GitHub Comments

1reaction
lorentzenchrcommented, Jun 3, 2022

Pretty much any GLM literature (score equation for general links and canonical links is the key):

  • McCullagh & Nelder, Generalized Linear Models (maybe still the best book about GLMs!) Chapters 2.2.3, 4.3.1
  • https://dx.doi.org/10.2139/ssrn.3822407 Chapter 5.1.5 for the balance property of canonical links which implies good calibration.

For probabilistic classification in particular (no GLM context)

The last point about efficiency (for fitting) is a property of Maximum Likelihood Estimation Theory (achieving the Cramer-Rao lower bound asymptotically).

1reaction
GaelVaroquauxcommented, Jun 2, 2022

That’s RidgeClassifier, isn’t it? Ooops, didn’t read well.

I’d say: is there a good literature documenting practice (and theory) showing benefit of such classifiers. If not, I wouldn’t prioritize this: we have a lot of things on our plate already.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Normal tissue complication probability (NTCP) modelling of ...
Penalised logistic regression and random forest classification ... (lower log loss) and overall performance (lower Brier score) than the PLR ...
Read more >
arXiv:2111.02529v1 [cs.LG] 3 Nov 2021
Abstract. Minimizing expected loss measured by a proper scoring rule, such as Brier score or log-loss (cross-entropy), is a common objective.
Read more >
sklearn.metrics.brier_score_loss
Compute the Brier score loss. The smaller the Brier score loss, the better, hence the naming with “loss”. The Brier score measures the...
Read more >
Classifier accuracy with filtering | Download Scientific Diagram
Also, the H-measure and Brier Score of the MP-LSTM model outperformed that in Feng et al. [31] and Ala'raj et al. [7] ,...
Read more >
Engineering nonlinear epileptic biomarkers using deep ...
The Gini Importance of each feature was calculated from the pretrained RFC, enabling the most significant features (MSFs) for each task to be ......
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found