Stars
Tools for merging pretrained large language models.
《代码随想录》LeetCode 刷题攻略:200道经典题目刷题顺序,共60w字的详细图解,视频难点剖析,50余张思维导图,支持C++,Java,Python,Go,JavaScript等多语言版本,从此算法学习不再迷茫!🔥🔥 来看看,你会发现相见恨晚!🚀
[ICLR'24 spotlight] An open platform for training, serving, and evaluating large language model for tool learning.
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
This repo aims to record resource of role-playing abilities in LLMs, including dataset, paper, application, etc.
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
A bibliography and survey of the papers surrounding o1
[ICLR2025 Spotlight] MagicPIG: LSH Sampling for Efficient LLM Generation
A framework for detecting, highlighting and correcting grammatical errors on natural language text. Created by Prithiviraj Damodaran. Open to pull requests and other forms of collaboration.
A quick guide (especially) for trending instruction finetuning datasets
Tools to download and cleanup Common Crawl data
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.
Awesome things about LLM-powered agents. Papers / Repos / Blogs / ...
RAGChecker: A Fine-grained Framework For Diagnosing RAG
List of papers on hallucination detection in LLMs.
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)
[EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs
Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
ACL 2024 | LooGLE: Long Context Evaluation for Long-Context Language Models
Implementation of paper Data Engineering for Scaling Language Models to 128K Context
Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarks

