llama-server as CLI
#11219
Replies: 1 comment 2 replies
-
I think something that can make calls without touching HTTP is useful, you can get that extra performance. But, it would still be nice to have a HTTP-based OpenAI one, since OpenAI is more popular and standard. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
@ericcurtin I'm creating a dedicated discussion in order not to make the thread become off-topic
FYI my recent refactor #10691 is aimed to (somewhat) doing this. The idea is that in the future, we can expose llama-server as an internal library. A downstream program (like llama-client as you mentioned) can make calls internally to llama-server without ever touching HTTP stack.
Not sure if this is what you're looking for, feel free to discuss more.
Beta Was this translation helpful? Give feedback.
All reactions