Incorporating additional data to train gp_model
See original GitHub issueHi,
I have a few questions regarding the incorporation of additional data to train the existing gp_model. As GP is a nonparametric algorithm which can learn as the data grows. Is there a way to incorporate additional data in gpytorch to update the hyperparameters without recalculating again using the whole data. I found two methods in GPyTorch which seems to be related to that :
get_fantasy_data()
- problem with this is that the method is not available for multitask GPs
set_train_data()
My query is regarding the second method; in documentation, it is given that the function doesn’t refit the hyperparameter. Does that mean it doesn’t update the hyperparameters at all?
Thanks in advance!
Issue Analytics
- State:
- Created 4 years ago
- Comments:7 (2 by maintainers)
Top Results From Across the Web
How to Update Neural Network Models With More Data
Retraining Update Strategies. A benefit of neural network models is that their weights can be updated at any time with continued training. When ......
Read more >GP Regression with Grid Structured Training Data
In this notebook, we demonstrate how to perform GP regression when your training data lies on a regularly spaced grid. For this example,...
Read more >Implement A Gaussian Process From Scratch | by Shuai Guo
The following code trains a GP model and makes predictions on a grid of test data. It is common to scale the GP...
Read more >Is there a way to incorporate new data into an already trained ...
It arises because as optimizers get more efficient, the neural network ... If in the future you feed it a bunch of data...
Read more >incorporating prior knowledge into gaussian process models
Another potentially useful attribute of GP model is the possibility to incorporate prior knowl- ... Based on a set of N training data...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
The hyperparameters don’t have to be updated when you add new data. Updating them will give your model a better fit, but comes at the cost of additional computation.
GPyTorch doesn’t fit the hyperparameters of GP models automatically - you have to write the optimization loop yourself (as seen in our examples). You should:
set_train_data
(and include the new data + all your old data)However, you may find for your particular application that the hyperparameters don’t vary too much with new data - and it might make more sense to sparingly update the hyperparameters.
That’s not easily possible with the same likelihood (the default one is homoskedastic). You could instead use a
FixedNoiseGaussianLikelihood
where you instantiate the first part of the vector of noises to the inferred noise level of your model and set the other ones to a very small noise level.