-
Notifications
You must be signed in to change notification settings - Fork 6.2k
Update tokenizer.model_max_length if necessary (Fix #6415) #6632
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Hi Xiaosu, thanks for your concern about the handling of tokenizer in llama factory. We thought that this property does not affect both the training and inference, so we would like not to modify it in our framework. |
Thank you for your comments! You are right there will be no defects to do training and/or inference, except one place: The VLLM inference will read the If we create an vllm engine by Therefore, I think this pr makes less confusion for further use cases. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Thank you for your quick response ❤️ |
Former-commit-id: 58d029f
Former-commit-id: 58d029f
What does this PR do?
Fixes #6415
It modifies logic of
loader.py -> load_tokenizer(...)
to updatetokenizer.model_max_length
ifmodel_args.model_max_length
changes.Before submitting