Stars
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
[ECCV2024 Oral🔥] Official Implementation of "GiT: Towards Generalist Vision Transformer through Universal Language Interface"
CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts
llama3 implementation one matrix multiplication at a time
This list of writing prompts covers a range of topics and tasks, including brainstorming research ideas, improving language and style, conducting literature reviews, and developing research plans.
RAG AutoML Tool - Find optimal RAG pipeline for your own data.
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
🦜🔗 Build context-aware reasoning applications
High accuracy RAG for answering questions from scientific documents with citations
ACL'2020: Biomedical Entity Representations with Synonym Marginalization
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
PubMedQA: A Dataset for Biomedical Research Question Answering
Dromedary: towards helpful, ethical and reliable LLMs.
True Few-Shot BioIE: Benchmarking GPT-3 In-Context and Small PLM Fine-Tuning
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
中文医疗信息处理基准CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark
Example models using DeepSpeed
🐫 CAMEL: Finding the Scaling Law of Agents. A multi-agent framework. https://www.camel-ai.org
Code and documentation to train Stanford's Alpaca models, and generate the data.