TensorRT Model Optimizer

Getting Started

  • Overview
  • Installation
    • Installation for Linux
    • Installation for Windows
  • Quick Start: Quantization
  • Quick Start: Quantization (Windows)
  • Quick Start: Pruning
  • Quick Start: Distillation
  • Quick Start: Sparsity

Guides

  • Support Matrix
  • Quantization
  • Pruning
  • NAS
  • Distillation
  • Sparsity
  • Saving & Restoring
  • Speculative Decoding

Deployment

  • TensorRT-LLM
  • DirectML
  • Unified HuggingFace Checkpoint

Examples

  • All GitHub Examples

Reference

  • Changelog
  • modelopt API

Support

  • Contact us
  • FAQs
TensorRT Model Optimizer
  • Installation
  • View page source

Installation

  • Installation for Linux
  • Installation for Windows
Previous Next

© Copyright 2023-2025, NVIDIA Corporation.

Built with Sphinx using a theme provided by Read the Docs.