Stars
Concise, consistent, and legible badges in SVG and raster format
greengerong / EchoMimic
Forked from BadToBest/EchoMimicLifelike Audio-Driven Portrait Animations through Editable Landmark Conditioning
DLRover: An Automatic Distributed Deep Learning System
Minimalistic large language model 3D-parallelism training
Ongoing research training transformer models at scale
An Efficient "Factory" to Build Multiple LoRA Adapters
TuGraph Analytics is a distributed graph compute engine.
TuGraph is a high performance graph database.
GLake: optimizing GPU memory management and IO transmission.
Xwin-LM: Powerful, Stable, and Reproducible LLM Alignment
High Accuracy and efficiency multi-task fine-tuning framework for Code LLMs. This work has been accepted by KDD 2024.
PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.
GitHub Action that uploads coverage to Codecov ☂️
A blazingly fast multi-language serialization framework powered by JIT and zero-copy.
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
Making large AI models cheaper, faster and more accessible
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
DeepRec is a high-performance recommendation deep learning framework based on TensorFlow. It is hosted in incubation in LF AI & Data Foundation.
Kubernetes Operator for AI and Bigdata Elastic Training
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
《Machine Learning Systems: Design and Implementation》- Chinese Version
MindSpore is a new open source deep learning training/inference framework that could be used for mobile, edge and cloud scenarios.
Guide for building custom op for TensorFlow
搞定C++:punch:。C++ Primer 中文版第5版学习仓库,包括笔记和课后练习答案。
FlexFlow Serve: Low-Latency, High-Performance LLM Serving