-
Notifications
You must be signed in to change notification settings - Fork 239
Issues: young-geng/EasyLM
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Dataset for the paper: The False Promise of Imitating Proprietary LLMs (https://arxiv.org/pdf/2305.15717)
#112
opened May 14, 2024 by
manandey
updated May 14, 2024
ERROR: Accessing retired flag 'jax_enable_async_collective_offload'
#109
opened Mar 15, 2024 by
LeoXinhaoLee
updated Mar 15, 2024
why 'LLaMATokenizer' object has no attribute 'sp_model'?
#101
opened Oct 24, 2023 by
zepen
updated Mar 8, 2024
Serving errors: deprecated dependencies and structure error
#103
opened Nov 7, 2023 by
sjw8793
updated Nov 8, 2023
Conflicting dependencies for jax[cuda11-pip]==0.4.14
#100
opened Oct 23, 2023 by
gbacon
updated Oct 23, 2023
What is the full batch size if mesh_dim is set to 1,1,-1, on TPUv3-8?
#94
opened Sep 23, 2023 by
TPFRL
updated Oct 5, 2023
For 30B LLama model, can server be supported by configuring mesh_dims on tpu v3-8 (128g)? I tried 8,1 and 4,1 but they don't seem to work.
#35
opened Apr 19, 2023 by
joytianya
updated Sep 5, 2023
How to convert the weights on HF into the format of EasyLM?
#7
opened Apr 4, 2023 by
joytianya
updated Aug 7, 2023
TPU Installation broken because of change in Orbax
#84
opened Aug 1, 2023 by
akhilkedia
updated Aug 2, 2023
Feature request: Use Orbax for checkpointing.
#80
opened Jul 25, 2023 by
OhadRubin
updated Jul 31, 2023
Use EasyLM to pre-train llama-7B using Nvidia GPU
#78
opened Jul 24, 2023 by
zhpacer
updated Jul 24, 2023
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.