Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

非安培架构显卡微调 #687

Open
98luobo opened this issue Oct 30, 2024 · 0 comments
Open

非安培架构显卡微调 #687

98luobo opened this issue Oct 30, 2024 · 0 comments

Comments

@98luobo
Copy link

98luobo commented Oct 30, 2024

大佬们,实验室显卡不支持flash_attention,我现在想微调internVL2-4b,是不是把pretrained的4b模型下的config.json中的"use_flash_attn"设置为false就行了?

还有个问题,我这个微调任务是面向grounding的,如果图片大于1000的话,是不是图片大小也得归一化一下

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant