-
Notifications
You must be signed in to change notification settings - Fork 493
Pull requests: bigscience-workshop/petals
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Support loading weights from Safetensors on server
#473
by borzunov
was merged Aug 22, 2023
Loading…
Temporarily require peft<0.5.0, transformers<4.32.0
#470
by justheuristic
was merged Aug 22, 2023
Loading…
Make client compatible with transformers' GenerationMixin
#464
by borzunov
was merged Aug 20, 2023
Loading…
Move SequenceManagerConfig -> ClientConfig, petals.dht_utils -> petals.utils.dht
#463
by borzunov
was merged Aug 14, 2023
Loading…
Prioritize short inference, unmerge pools for long inference
#458
by borzunov
was merged Aug 11, 2023
Loading…
Use torch.cuda.synchronize for compute throughput
#456
by justheuristic
was merged Aug 9, 2023
Loading…
benchmarks: Aggregate speed among workers, set default dtype torch32
#454
by borzunov
was merged Aug 9, 2023
Loading…
Test Llama, rebalancing, throughput eval, and all CLI scripts
#452
by borzunov
was merged Aug 8, 2023
Loading…
Force using --new_swarm instead of empty --initial_peers
#451
by borzunov
was merged Aug 8, 2023
Loading…
Prefer longer servers for fine-tuning, exclude unreachable
#448
by borzunov
was merged Aug 7, 2023
Loading…
[Refactor] extract block forward, backward and inference into a separate file
#435
by justheuristic
was merged Aug 7, 2023
Loading…
3 tasks done
Rewrite MemoryCache alloc_timeout logic
#434
by justheuristic
was merged Aug 28, 2023
Loading…
4 tasks done
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.