Stars
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
A high-throughput and memory-efficient inference and serving engine for LLMs
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Python packaging and dependency management made easy
Universal LLM Deployment Engine with ML Compilation
SGLang is a fast serving framework for large language models and vision language models.
Ongoing research training transformer models at scale
An open source implementation of CLIP.
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
Utilities intended for use with Llama models.
Minimalistic large language model 3D-parallelism training
Awesome Reasoning LLM Tutorial/Survey/Guide
Minimalistic 4D-parallelism distributed training framework for education purpose
This repository contains the official implementation of the research papers, "MobileCLIP" CVPR 2024 and "MobileCLIP2" TMLR August 2025
A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.
A reimplementation of Andrej Karpathy's repository for an RL self-learning AI agent that learns to play Pong through trial and error, using PyTorch

