Run Local LLMs on Any Device. Open-source
Replace OpenAI GPT with another LLM in your app
Phi-3.5 for Mac: Locally-run Vision and Language Models
A high-throughput and memory-efficient inference and serving engine
Tensor search for humans
PyTorch library of curated Transformer models and their components
Operating LLMs in production
Database system for building simpler and faster AI-powered application
Run 100B+ language models at home, BitTorrent-style
Visual Instruction Tuning: Large Language-and-Vision Assistant
State-of-the-art Parameter-Efficient Fine-Tuning
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
A high-performance ML model serving framework, offers dynamic batching
LLM training code for MosaicML foundation models
Implementation of "Tree of Thoughts
The unofficial python package that returns response of Google Bard
Framework that is dedicated to making neural data processing
Low-latency REST API for serving text-embeddings
Implementation of model parallel autoregressive transformers on GPUs