Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Recommended approach for local models? i.e. Swappable model support. #37

Open
illuminatedjavascript opened this issue Mar 1, 2024 · 1 comment

Comments

@illuminatedjavascript
Copy link

Looks like litellm does a (too) good job of encapsulating the calls to openai, making calls to local openai-api-based models require a proxy to intercept and re-route.

Is this the recommended approach for the time being? Any plans to drop the litellm dependency, use one that's a little more open, or write your own layer?

It would be nice to use this with swappable models especially since AC seems to generalize across general instruct models and not require function-calling models.

@pslusarz
Copy link

pslusarz commented Oct 2, 2024

I added AWS Bedrock support on this MR. Not sure how good results would be on the local models, since they are kind of weak at development, but the code seems to support them, ie you can try something like model="huggingface/deepseek-ai/deepseek-coder-33b-instruct" in configuration.toml.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants
@pslusarz @illuminatedjavascript and others