Stars
A Lightweight Face Recognition and Facial Attribute Analysis (Age, Gender, Emotion and Race) Library for Python
The rewritten engine, originally for tensorflow. Now all other backends have been ported here.
Review/Check GGUF files and estimate the memory usage and maximum tokens per second.
MMD Tools is a blender addon for importing/exporting Models and Motions of MikuMikuDance.
Makes judging, compiling, and downloading problems for competitive programming easy.
💯 Curated coding interview preparation materials for busy software engineers
Grokking the Coding Interview: Patterns for Coding Questions Alternative
Fast inference engine for Transformer models
Simple frontend for LLMs built in react-native.
Learn how to design large-scale systems. Prep for the system design interview. Includes Anki flashcards.
On-device AI across mobile, embedded and edge for PyTorch
real time face swap and one-click video deepfake with only a single image
[NeurIPS 2024] Official implementation of "Faster Diffusion: Rethinking the Role of UNet Encoder in Diffusion Models"
Run PyTorch LLMs locally on servers, desktop and mobile
Cross-platform, customizable ML solutions for live and streaming media.
GoodbyeDPI — Deep Packet Inspection circumvention utility (for Windows)
A Compressed Stable Diffusion for Efficient Text-to-Image Generation [ECCV'24]
Everything you need to know to get the job.
T-GATE: Temporally Gating Attention to Accelerate Diffusion Model for Free!
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
Implementing a ChatGPT-like LLM in PyTorch from scratch, step by step
High-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model
Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
Free and Open Source, Distributed, RESTful Search Engine
a header-only, constexpr alternative to gperf for C++14 users
Auto detecting, masking and inpainting with detection model.
Optimum version of a UI for Stable Diffusion, running on ONNX models for faster inference, working on most common GPU vendors: NVIDIA,AMD GPU...as long as they got support into onnxruntime