We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
现在训练时样本长短不一,如果batch size设的过大会导致OOM,设的过小GPU算力有没有被利用起来,我想把数据根据大小排序或分组,然后希望训练时能根据根据当前batch的长度来动态调节batch size,达到算力的最优利用,气温能给我一些实现的建议吗?
The text was updated successfully, but these errors were encountered:
可以设计这样的算法:
Sorry, something went wrong.
@Jintao-Huang 谢谢你的回复,我的意思是,在代码的大概什么位置,可以动态的设置每个批次的batch size或者说可以调节每个批次取到的样本数,比如说哪个文件或哪个类
动态batchsize已经在transformers最新版中支持
Jintao-Huang
No branches or pull requests
现在训练时样本长短不一,如果batch size设的过大会导致OOM,设的过小GPU算力有没有被利用起来,我想把数据根据大小排序或分组,然后希望训练时能根据根据当前batch的长度来动态调节batch size,达到算力的最优利用,气温能给我一些实现的建议吗?
The text was updated successfully, but these errors were encountered: