[Question] How to get the number of target algorithm runs without fitting data?
See original GitHub issueHi there,
To get the number of target algorithm runs, auto-sklearn needs to fit data, and finally call sprint_statistics()
. Is there any other way to get the number of target algorithm runs without fitting data? Since in my case, it just got 10 runs under time_left_for_this_wask=3600. To evaluate more runs. I know I could increase the time limit. In addition to do in this way, does anyone know how auto-sklearn decides the number of target algorithm runs?
Thanks a lot.
Issue Analytics
- State:
- Created a year ago
- Comments:5 (2 by maintainers)
Top Results From Across the Web
Machine Learning: Target Feature Label Imbalance Problems ...
The goal of this post is to teach python programmers why they must have balanced data for model training and how to balance...
Read more >Overfitting and Underfitting With Machine Learning Algorithms
An underfit machine learning model is not a suitable model and will be obvious as it will have poor performance on the training...
Read more >Logistic Regression Questions - Analytics Vidhya
What do you mean by the Logistic Regression? It's a classification algorithm that is used where the target variable is of categorical nature....
Read more >Choosing the Correct Type of Regression Analysis
Fitted line plots: If you have one independent variable and the dependent variable, ... Not all count data follow the Poisson distribution because...
Read more >1.11. Ensemble methods — scikit-learn 1.2.0 documentation
As other classifiers, forest classifiers have to be fitted with two arrays: a sparse or ... tasks (where n_features is the number of...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Hi @jmren168,
We use SMAC3 as a Bayesian optimization (BO) framework. You can refer to the reference given there. Specifically, where we instansiate it is here.
I won’t go into details as BO is a huge field but the core principles:
x
can be seen as a point in the very high-dimensional spaceX
.x
, we would love to know its true performancep
ahead of time but we only know it after evaluating it.x -> u, s
, a mean and std deviations, so it builds a model of the entire spaceX
.x
evaluated, the BO model is updated and knows a little more about the whole landscapeX
. Essentially it’s a statistical model that learns to model the spaceX
.X
we don’t know much about will be tried once in a while but also to exploit regions ofX
where we do know a lot and shows good performance.Hi @eddiebergman,
After reading the paper you mentioned, there are 4 types of SMAC3, which one is really implemented in autosklearn? Some of them? All of them? Thanks.
SMAC4BB: SMAC for Low-dimensional and Continuous Black-Box Functions SMAC4HPO: SMAC for CASH and Structured Hyperparameter Optimization SMAC4MF: SMAC for Expensive Tasks and Automated Deep Learning SMAC4AC: SMAC for Algorithm Configuration