-
Notifications
You must be signed in to change notification settings - Fork 224
Add ColQwen3 Support #366
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Add ColQwen3 Support #366
Conversation
|
Some tests don't run. (2) The tests. should we bump the transformers version and we're good ? |
|
@hxssgaa thanks for bringing these over. Having a blast using them! |
We have fixed ruff lint error in our commits, the other lint errors are not from us. For the transformers version it must be at least |
|
I try to run tomoro-colqwen3-embed-8b with this PR, and I encounter following errors: How to fix it? or Is it my wrong configuration? |
Hi, note that Colpali format isn't directly compatible with current Tomoro-colqwen3 hf models as they are converted from this Colpali format, you can refer to the Tomoro HF repo for how to run the models for now. We intent to also share the conversion script later, but it shouldn't belong to this repo. |
|
So.one thing I don't understand yet is that this PR is not compatible with the Tomoro checkpoints you shared ? Isn't it better to make this compatible ? |
|
We based our huggingface repo off the colqwen2 implementation on transformers here: https://github.com/huggingface/transformers/tree/main/src/transformers/models/colqwen2, which uses a different naming convention for params compared to this repo. Should we unify the names or keep 2 separate versions? |
|
I trained a 2B smaller colqwen3 model on the Qwen3-VL-2B-Instruct model, welcome to follow and use~ |
ManuelFay
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is it possible to rebase on main so that linting gets corrected ?
The pther interrogation I have is whether the pyprojecxt should get updatred. From my understanding, qwen3 is only in newer versions of transformers so I am guessing we should bump the minimal transformers package right ?
The rest looks great !
) * looks like colqwen 2.5 omni support was accidentally removed in illuin-tech#339 EDIT: that was based upon just looking at the main __init__.py. looking at the other files, perhaps it was intentionally removed... * found & fixed resize_token_embeddings() breakage
* lint * lint examples
Already rebased to main, and yes, I have updated the minima transformer version to be |
Add ColQwen3 support
The fine-tuned colqwen3 models are below including with benchmark results: