-
Notifications
You must be signed in to change notification settings - Fork 132
Open
Description
Is your feature request related to a problem? Please describe.
From a tooling standpoint, we need the ability to discover all running LLM endpoints, so we can pick one and use it as an AI assistant in an IDE (using the continue.dev extension in VS Code/IntelliJ for instance)
Describe the solution you'd like
The model endpoints should be listed with at least their label,type, and api url e.g.
- label: LLama 3
- provider: ollama (or instructlab...)
- apiUrl: https://my.cluster:12345/foo/
Describe alternatives you've considered
AFAIK, there's no other way to discover running inference engines at the moment.
cc @amfred
Metadata
Metadata
Assignees
Labels
No labels