question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

[FR] Autologging functionality for scikit-learn integration with XGBoost (and LightGBM)

See original GitHub issue

MLflow Roadmap Item

This is an MLflow Roadmap item that has been prioritized by the MLflow maintainers. We’re seeking help with the implementation of roadmap items tagged with the help wanted label. Items tagged with the needs design label are awaiting a design sketch from an MLflow maintainer: @WeichenXu123.

For requirements clarifications and implementation questions, or to request a PR review, please tag @WeichenXu123 in your communications related to this issue.

Proposal Summary

Hello,

When trying to implement autolog() function for XGBoostClassifier and XGBoostRegressor, we are not able to see the metrics and parameters or models which are getting logged. Can you please address this issue.

Thanks.

PFA added code for XGBoostRegressor which is to be converted to .py file to replicate the issue boston.txt

Issue Analytics

  • State:closed
  • Created 2 years ago
  • Comments:25 (16 by maintainers)

github_iconTop GitHub Comments

1reaction
harupycommented, Oct 11, 2021

Logging feature impotance within sklearn autologging sounds good to me.

1reaction
dbczumarcommented, Oct 8, 2021

Autologging for XGBoost sklearn estimators should be initiated through mlflow.xgboost.autolog. We can add an internal (prefixed with _) autolog API in mlflow.sklearn with an xgboost_estimators=True/False param. If xgboost_estimators==true, we only patch XGBoost sklearn estimators and associated them with the XGBoost autologging integration via safe_patch("xgboost", ...). Otherwise, we patch the sklearn estimators (and not the xgboost estimators) via safe_patch("sklearn", ...), which is the current behavior. Let me know if this sounds good!

Yes, looks good to me! I will start to work on its implementation unless there are other suggestions. Thanks!

Thanks a bunch, @jwyyy ! Sounds good!

Read more comments on GitHub >

github_iconTop Results From Across the Web

Gradient Boosting with Scikit-Learn, XGBoost, LightGBM, and ...
This is a type of ensemble machine learning model referred to as boosting. Models are fit using any arbitrary differentiable loss function and ......
Read more >
Real-time Serving for XGBoost, Scikit-Learn RandomForest ...
Available as a Docker container and integrates easily with Kubernetes ... If you would like to try deploying your own XGBoost, LightGBM, ...
Read more >
Cross-version Testing in MLflow - The Databricks Blog
It integrates with many popular ML libraries such as scikit-learn, XGBoost, TensorFlow, and PyTorch to support a broad range of use cases.
Read more >
XGBoost vs LightGBM: How Are They Different - Neptune.ai
Both the algorithms treat missing values by assigning them to the side that reduces loss the most in each split. Feature Importance methods....
Read more >
Convert a pipeline with a XGBoost model — onnxcustom
sklearn -onnx can convert the whole pipeline as long as it knows the converter associated to a XGBClassifier. Let's see how to do...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found