[Question] Best practices regarding Sobol sampling
See original GitHub issueIn the gpei_hartmann_service.ipynb
tutorial, before exploring using GPEI, a Sobol strategy has been used. Just curious if there is any reference on the best practices regarding the Sobol strategy: mainly what’s the relationship between the search space and the number of Sobol trials.
Issue Analytics
- State:
- Created a year ago
- Reactions:2
- Comments:5 (4 by maintainers)
Top Results From Across the Web
Understanding Sobol sampling - Zemax Knowledgebase
Sobol sampling takes a different approach. Rather than generating random numbers, it generates a uniform distribution in probability space.
Read more >To Sobol or not to Sobol? The effects of sampling schemes in ...
Here, we consider three sampling strategies: random sampling, Latin hypercube sampling (LHS), and Sobol sequences. Figure 1A shows a visual ...
Read more >Sobol Sequence Explained by Antoine Savine - Medium
Sobol Sequence Explained by Antoine Savine. It is considered best practice in financial Monte-Carlo simulations to apply quasi-random numbers ...
Read more >(PDF) Sampling based on Sobol' sequences for Monte Carlo ...
Quasi-random (QR) sequences such as the Sobol′ sequences are designed to generate a sample that is uniformly distributed over the unit ...
Read more >Efficient sampling algorithm for large-scale optimization under ...
Quasi-Monte Carlo methods based on low-discrepancy sequences (Niederreiter, 1992) like ... problem scrambled Sobol sampling was introduced by Chi et al.
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
In general, you should get away with relatively few Sobol points. This is really just an initialization and mostly in place to prime the model to be able to fit to some data. In general, if there are too few points to get a good model fit, the model uncertainty should be high across the entire search space and cause any reasonable acquisition function to behave more or less like random exploration. So if you have a small iteration budget it would be reasonable to use a small number of Sobol points. Using a few more points will probably reduce the chances of the model running into numerical issues and make the model fitting more stable (this is more an intuition / hypothesis than proven fact though, I’ve never formally investigated this).
Interestingly, for
GPEI
there were nearly identical results using a10/90
Sobol/GPEI vs. a46/54
Sobol/GPEI with 23 hyperparameters, see Figure 2 of https://doi.org/10.1016/j.commatsci.2022.111505.