Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug] LLMParams and LLMParamsDoc Pydantic Model Error #1078

Open
6 tasks
jjmaturino opened this issue Dec 26, 2024 · 2 comments
Open
6 tasks

[Bug] LLMParams and LLMParamsDoc Pydantic Model Error #1078

jjmaturino opened this issue Dec 26, 2024 · 2 comments
Assignees
Labels
bug Something isn't working

Comments

@jjmaturino
Copy link

jjmaturino commented Dec 26, 2024

Priority

Undecided

OS type

Ubuntu

Hardware type

Xeon-GNR

Installation method

  • Pull docker images from hub.docker.com
  • Build docker images from source

Deploy method

  • Docker compose
  • Docker
  • Kubernetes
  • Helm

Running nodes

Single Node

What's the version?

N/A

Description

TL;DR: Pydantic Model accepts streaming parameter and not stream parameter as defined by TGI api spec.

There is a discrepancy between the TGI interface standard and the pydantic models defined.

The current difference is that the repo models define the stream parameter as streaming.

This causes the json that is accepted, (marshaled and unmarshalled) to expect streaming as the json key rather than the TGI standard of stream


Discovered during this PR where I was noticing during testing that when I posted to the chat endpoint via curl with the stream json key,

The program would not successfully unmarshal the json object.

After talking with @xiguiw , and looking at the TGI documentation, I believe that this is an error in the codebase.

This might be a breaking change if fixed.

Reproduce steps

streaming: bool = True

streaming: bool = True


https://huggingface.github.io/text-generation-inference/

Raw log

No response

Attachments

No response

@xiguiw
Copy link
Collaborator

xiguiw commented Jan 2, 2025

@jjmaturino

Thank for catching this!
@XinyaoWa will help to fix it.
It's working in process.

@joshuayao
Copy link
Collaborator

Hi @jjmaturino, the bug was fixed. Could you please help verify it with the latest code? Thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

4 participants