feat: register models for llama stack #2867
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Signed-off-by: Philippe Martin [email protected]
What does this PR do?
During llama stack container creation:
Screenshot / video of UI
(in the demo below, the chat completion with the model
MaziyarPanahi/Mistral-7B-Instruct-v0.3.Q4
fails because the output ofllama-stack-client models list
truncates the names of models - the name to use should have beenMaziyarPanahi/Mistral-7B-Instruct-v0.3.Q4_K_M
)You can see in the demo that all models are registered in the llama stack, and inference servers are started only for the ones used during a chat completion.
llama-stack-register-all-models.mp4
What issues does this PR fix or reference?
Fixes #2840
How to test this PR?