-
Couldn't load subscription status.
- Fork 2.8k
[MOE] Fix issues with batch, use postponed_const when fusing expert weights #32546
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
[MOE] Fix issues with batch, use postponed_const when fusing expert weights #32546
Conversation
…into mateuszm/moe_fix_bs_postponed
| if: ${{ inputs.model_scope == 'precommit' }} | ||
| run: | | ||
| export PYTHONPATH=${MODEL_HUB_TESTS_INSTALL_DIR}:$PYTHONPATH | ||
| python3 -m pip install optimum-intel==1.25.2 transformers==4.53.3 --upgrade |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That is a bad practice to introduce test dependencies in the pipeline, why can't we use the common test requirements?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Upgrade to transformers and optimum-intel is required to use versions that support qwen3_moe models.
Ideally it would be best to upgrade tests/requirements_pytorch to align versions, however this might be bit risky without thorough testing since this requirements are used in large amount of model tests, including nightly large models.
I've initially tried running other precommit suites test_pa_transformation test_transformations and it seem they were successful with upgraded versions
Details:
Tickets: