Replies: 3 comments
-
@GuanLuo @tanmayv25 Do you know if this is somehow doable? Could we just have the model frameworks load the binaries directly at runtime? |
Beta Was this translation helpful? Give feedback.
-
Yes. It is already supported. See the documentation on this feature here: https://github.com/triton-inference-server/server/blob/main/docs/protocol/extension_model_repository.md#load |
Beta Was this translation helpful? Give feedback.
-
If using the python clients then look at the config and file options: https://github.com/triton-inference-server/client/blob/main/src/python/library/tritonclient/grpc/__init__.py#L656 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
It would be beneficial if we there is a way to inject/passthrough model binary via API rather than current single way (via model file path). For example, in on premise use case which require model encryption/protection, user can writing their own mechanism to protect model and inject model into triton and unload without using filesystem as a medium
Beta Was this translation helpful? Give feedback.
All reactions