-
Notifications
You must be signed in to change notification settings - Fork 4.8k
Added support of google/flan models #2321
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
|
||
| def match(self, model_path: str): | ||
| return "flan" in model_path.lower() | ||
|
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What is the instruction template for flan? Do we need to set it here?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @merrymercy, do you mean prompt template?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yes
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@wangzhen263 Any updates? IIRC, flan-t5 has a default prompt template.
bf7aa7e to
a81a04c
Compare
Added AutoModelForSeq2SeqLM to init google/flan models in 8bit
|
Closed due to inactivity. Feel free to reopen after you address these problems.
|
Why are these changes needed?
Related issue number (if applicable)
Checks
format.shto lint the changes in this PR.