Allow selected models in the model repository not to be loaded
See original GitHub issueIs your feature request related to a problem? Please describe.
Yes. Some models in my model repository runs just on CPU and some others just GPU. If launching the server without GPU support, it fails because backends&models with GPU support fails to load. Would like to be able to exclude some models being loaded in case not launching triton with GPU support.
Describe the solution you’d like
- Maybe naming the model folders with a prefix of “". So all model folder names in a repo starting with "” character will be omitted to be loaded.
- It will skip loading failed models and without crashing, continue with successful models.
Describe alternatives you’ve considered Currently, if I want to test CPU only models, I move GPU models to temporary folder.
Issue Analytics
- State:
- Created 2 years ago
- Comments:7 (4 by maintainers)
Top Results From Across the Web
Getting Started with Repositories
This document shows how to handle repositories through the web interface as well as through the terminal. There are no requirements if working...
Read more >MLflow Model Registry on Databricks
MLflow Model Registry is a centralized model repository and a UI and set of APIs that enable you to manage the full lifecycle...
Read more >Model Repository: DataRobot docs
These blueprints illustrate the algorithms (the preprocessing steps, selected estimators, and in some models, postprocessing as well) used to build a model, not...
Read more >Register and Deploy Models with Model Registry
With the SageMaker model registry you can catalog models for production, ... Model packages used in the AWS Marketplace are not versionable entities...
Read more >Develop models in Code Repositories
If configured, a model is automatically submitted whenever a transaction is successfully committed to that model. However, as it is not always desired...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
@mustafakasap You can try out
--exit-on-error=false
when launching server. See here for more details: https://github.com/triton-inference-server/server/blob/main/src/servers/main.cc#L310``