Browse free open source LLM Inference tools and projects for Windows and Linux below. Use the toggles on the left to filter open source LLM Inference tools by OS, license, language, programming language, and project status.
High-performance neural network inference framework for mobile
ONNX Runtime: cross-platform, high performance ML inferencing
AIMET is a library that provides advanced quantization and compression
Protect and discover secrets using Gitleaks
C++ library for high performance inference on NVIDIA GPUs
Easy-to-use deep learning framework with 3 key features
An MLOps framework to package, deploy, monitor and manage models
Superduper: Integrate AI models and machine learning workflows
OpenMMLab Model Deployment Framework
Standardized Serverless ML Inference Platform on Kubernetes
Unified Model Serving Framework
Deep learning inference framework optimized for mobile platforms
A unified framework for scalable computing
Uniform deep learning inference framework for mobile
Self-contained Machine Learning and Natural Language Processing lib
A set of Docker images for training and serving models in TensorFlow
A GPU-accelerated library containing highly optimized building blocks
Official inference library for Mistral models
Neural Network Compression Framework for enhanced OpenVINO
Powering Amazon custom machine learning chips
A general-purpose probabilistic programming system
High-level Deep Learning Framework written in Kotlin
Toolkit for allowing inference and serving with MXNet in SageMaker
Framework that is dedicated to making neural data processing
A real time inference engine for temporal logical specifications