Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Hello , Can LMFlow support Qwen1.5-1.8B model Fine-tuning? #796

Closed
13416157913 opened this issue Apr 28, 2024 · 3 comments
Closed

Hello , Can LMFlow support Qwen1.5-1.8B model Fine-tuning? #796

13416157913 opened this issue Apr 28, 2024 · 3 comments

Comments

@13416157913
Copy link

Hello , Can LMFlow support Qwen1.5-1.8B model Fine-tuning?

@research4pan
Copy link
Contributor

research4pan commented Apr 28, 2024

Thanks for your interest in LMFlow! We are integrating that feature right now, hopefully supporting it in 12-48 hours. Please stay tuned for our latest update 😄

@wheresmyhair
Copy link
Collaborator

wheresmyhair commented Apr 28, 2024

Hello , Can LMFlow support Qwen1.5-1.8B model Fine-tuning?

Hi, we've tested on Qwen1.5-1.8B and the script works fine.
qwen2

Please make sure you include --lora_target_modules q_proj, v_proj (only for qwen models) in the finetune shell script:
ft

Also, we strongly recommend you to:

  1. Use a conversation dataset to finetune the model. You could either:
    • To test the workflow, try to download a conversation dataset from out data server via:
       cd data && ./download.sh alpaca && cd -
      and specify the dataset to data/alpaca/train_conversation, or
    • Prepare your own conversation dataset (see here)
  2. Specify the conversation template to qwen2 for better performance.

@13416157913
Copy link
Author

Hello , Can LMFlow support Qwen1.5-1.8B model Fine-tuning?

Hi, we've tested on Qwen1.5-1.8B and the script works fine. qwen2

Please make sure you include --lora_target_modules q_proj, v_proj (only for qwen models) in the finetune shell script: ft

Also, we strongly recommend you to:

  1. Use a conversation dataset to finetune the model. You could either:

    • To test the workflow, try to download a conversation dataset from out data server via:

       cd data && ./download.sh alpaca && cd -

      and specify the dataset to data/alpaca/train_conversation, or

    • Prepare your own conversation dataset (see here)

  2. Specify the conversation template to qwen2 for better performance.

Thank you very much.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants