-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Query server for model availability of OpenAI-compatible servers #609
Comments
To be sure, the use case here is to populate the dropdown if you're using the OpenAI class with a personal API token? |
That or any OpenAI model server, where models can have completely different names from that. but yea, the autopopulate part is correct |
i tried using continue with a small team, and it was horrendous to set up, especially since we had to switch models a few times. If I can help let me know |
@simon376 Thanks for the honesty here. Model setup is something we're planning to spend most of next week on, so if you don't mind I'd actually love to hear the entirety of your experience. Is GitHub issues the best way to communicate, Discord, or even a quick call? Whatever is most comfortable. I'll have many questions and can hopefully share some early ideas with you Our goal is like you said, to get this to a point where you wouldn't need to worry about the config file at all |
I will message you on discord when I find the time. my choice of words was a bit harsh tbh 😅 maybe I was a bit ill-prepared |
No worries! "horrendous" was properly descriptive in my opinion : ) Many of the noted improvements are now available in the pre-release. Still yet to autodetect available models, but I still think this is a good idea |
I am using the pre-release version v.09.80 now. At some point I actually got it running! Wow! Now, I can't get it. I am writing a guide for my users, and tried doing it step-by-step, but couldn't get it to probe. Some things I noticed when trying it:
If I add a model with the name AUTODETECT and no title, like this:
I end up with a model named "custom" on the model list, which points to... I don't know whom. Its response don't match my models. |
This issue was closed because it wasn't updated for 10 days after being marked stale. If it's still important, please reopen + comment and we'll gladly take another look! |
Validations
Problem
OpenAI-compatible inference servers have the
/v1/models
endpoint which lists their models and capabilities. e.g.:This returns a JSON list of the models and some of their capabilities. Compatible servers implement the same.
This would make the configuration of such servers much easier. Instead of writing the config.py manually, one can just query the server for the models...
Solution
No response
The text was updated successfully, but these errors were encountered: