-
Peking University
- Beijing, China
- https://www.microsoft.com/en-us/research/people/wangliang/
Highlights
Stars
Tongyi Deep Research, the Leading Open-source Deep Research Agent
[NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewards
Lightweight coding agent that runs in your terminal
SGLang is a fast serving framework for large language models and vision language models.
SkyRL: A Modular Full-stack RL Library for LLMs
Democratizing Reinforcement Learning for LLMs
RAGEN leverages reinforcement learning to train LLM reasoning agents in interactive, stochastic environments.
verl: Volcano Engine Reinforcement Learning for LLMs
🤗 smolagents: a barebones library for agents that think in code.
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & vLLM & Ray & Dynamic Sampling & Async Agentic RL)
Official inference framework for 1-bit LLMs
A bibliography and survey of the papers surrounding o1
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Large World Model -- Modeling Text and Video with Millions Context
LOFT: A 1 Million+ Token Long-Context Benchmark
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
⚡FlashRAG: A Python Toolkit for Efficient RAG Research (WWW2025 Resource)
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Generative Representational Instruction Tuning
Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI
LongEmbed: Extending Embedding Models for Long Context Retrieval (EMNLP 2024)


