Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

curriculum training #5

Closed
Jiakui opened this issue Jan 14, 2019 · 3 comments
Closed

curriculum training #5

Jiakui opened this issue Jan 14, 2019 · 3 comments

Comments

@Jiakui
Copy link

Jiakui commented Jan 14, 2019

In the paper, it is said that the model should be trained with three stages. However, in the code, I didn't see how you implement the three stages. Therefore, if we run the code, could we reproduce the accuracy mentioned in the paper?

Thanks!

@Canjie-Luo
Copy link
Owner

Yes, the MORAN v1 in our paper was trained using curriculum learning. But the review took a long time. And we developed and released MORAN v2 for more stable convergence and higher accuracy. The released version was trained end-to-end with only one stage. But it outperformed the MORAN v1. The results were reported in the README.md. Run the code and manually decrease the learning rate~You can reproduce the v2 results. By the way, if the MORAN v2 is trained using curriculum learning, it may yield better performance.

@BlakeXiaochu
Copy link

BlakeXiaochu commented Jan 19, 2019

Hi~ We can reproduce the v2 results by running the code and manually decreasing the learning rate. What's the strategy of decreasing learning rate? For example, divide the learning rate by 10 every 100000 iter? Thx~

@Canjie-Luo
Copy link
Owner

Please divide the lr by 10 after 3 epoch. Totally 6 epoch maybe enough for training. Enjoy yourself~

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants