Replies: 1 comment 3 replies
-
I don't think anyone is working on Yarn long context for Qwen models. I'm not very familiar with the concept and personally find the 32k context enough for my needs. But there is no reason not to support it if it works, so feel free to add whatever changes are necessary. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I notice a lot of (I believe) fairly new changes related to Qwen models in convert_hf_to_gguf.py -- originally I thought it was probably in answer to bartowski's question. But the work appears incomplete.
I dont notice any plan or any public detail to align the work towards, which makes it hard to contribute. I was actually going to try to knock out bartowski's request after completing some work on a project I have ( llama-gguf-optimize), which I've done, but now that I am looking again to begin, I see these changes that I wasn't expecting.
This was added with Support for MiniCPM about 14 hours ago .. so, not terribly long ago and maybe there's something I am missing -- I could use some help in finding it. Is there any plan to add support for the yarn long context, as described in my comment on Bartowski's earlier post?
cc: @JFLFY2255
Beta Was this translation helpful? Give feedback.
All reactions