-
NTUST
- Taiwan
- https://hackmd.io/@GlycerinLOL
- @GlycerinLOL
Highlights
- Pro
Block or Report
Block or report GlycerinLOL
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
Get up and running with Llama 3, Mistral, Gemma 2, and other large language models.
Easy token price estimates for 400+ LLMs. TokenOps.
The Patterns of Scalable, Reliable, and Performant Large-Scale Systems
List of papers on hallucination detection in LLMs.
Galgame翻译器,支持剪贴板、OCR、HOOK等。Visual Novel translate tool , support clipboard / OCR/ HOOK
33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU
Implementation of paper "Towards a Unified View of Parameter-Efficient Transfer Learning" (ICLR 2022)
A modular RL library to fine-tune language models to human preferences
A quick guide (especially) for trending instruction finetuning datasets
ReFT: Representation Finetuning for Language Models
Compilers Principles, Techniques, & Tools (purple dragon book) second edition exercise answers. 编译原理(紫龙书)第2版习题答案。
使用 NextJS + Notion API 实现的,支持多种部署方案的静态博客,无需服务器、零门槛搭建网站,为Notion和所有创作者设计。 (A static blog built with NextJS and Notion API, supporting multiple deployment options. No server required, zero threshold t…
A WebUI for Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Train transformer language models with reinforcement learning.
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
A curated list of Human Preference Datasets for LLM fine-tuning, RLHF, and eval.
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
Awesome-LLM: a curated list of Large Language Model
Paper List for In-context Learning 🌷
Source code of extensions for Tachiyomi/Mihon and variants.
Free and open source manga reader for Android
一款专注于Ai翻译的工具,可以用来一键自动翻译RPG SLG游戏,Epub TXT小说,Srt Lrc字幕等等。
Robust Speech Recognition via Large-Scale Weak Supervision
Efficient Attention for Long Sequence Processing
整理 pytorch 单机多 GPU 训练方法与原理