Lists (1)
Sort Name ascending (A-Z)
Starred repositories
Data validation using Python type hints
A topic-centric list of HQ open datasets.
Blind&Invisible Watermark ,图片盲水印,提取水印无须原图!
Modeling, training, eval, and inference code for OLMo
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
Ring attention implementation with flash attention
Large World Model -- Modeling Text and Video with Millions Context
Windows Calculator: A simple yet powerful calculator that ships with Windows
A minimal GPU design in Verilog to learn how GPUs work from the ground up
Recipes for shrinking, optimizing, customizing cutting edge vision models. 💜
Automatically Discovering Fast Parallelization Strategies for Distributed Deep Neural Network Training
Qlib is an AI-oriented Quant investment platform that aims to use AI tech to empower Quant Research, from exploring ideas to implementing productions. Qlib supports diverse ML modeling paradigms, i…
The Enterprise-Grade Production-Ready Multi-Agent Orchestration Framework. Website: https://swarms.ai
Code and documentation to train Stanford's Alpaca models, and generate the data.
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
A quick guide (especially) for trending instruction finetuning datasets
A powerful tool for creating fine-tuning datasets for LLM
The code used to train and run inference with the ColVision models, e.g. ColPali, ColQwen2, and ColSmol.
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
A minimal tensor processing unit (TPU), inspired by Google's TPU V2 and V1
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Python Backtesting library for trading strategies
stefan-jansen / zipline-reloaded
Forked from quantopian/ziplineZipline, a Pythonic Algorithmic Trading Library
Find your trading edge, using the fastest engine for backtesting, algorithmic trading, and research.
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.

