Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
74 commits
Select commit Hold shift + click to select a range
5dc2f10
Auto-merge updates from master branch
invalid-email-address Feb 19, 2025
63f87e0
[Automated Commit] Format Codebase
github-actions[bot] Feb 19, 2025
a629034
Auto-merge updates from master branch
invalid-email-address Feb 20, 2025
bf8c1c2
Auto-merge updates from master branch
invalid-email-address Feb 21, 2025
ec284d2
Updated tags for submission checker command in docs
arjunsuresh Feb 24, 2025
5335553
Update mobilenets docs
arjunsuresh Feb 25, 2025
b9767aa
Update main.py
arjunsuresh Feb 25, 2025
f42aeeb
Update main.py
arjunsuresh Feb 25, 2025
d377848
Auto-merge updates from master branch
invalid-email-address Feb 25, 2025
c699ce3
update dataset download commands - waymo calib (#2130)
anandhu-eng Feb 25, 2025
c35459b
Auto-merge updates from master branch
invalid-email-address Feb 25, 2025
7c09a73
Merge branch 'master' into dev
arjunsuresh Feb 25, 2025
74ed21c
Merge branch 'master' into dev
arjunsuresh Feb 25, 2025
0eee795
Merge branch 'master' into dev
nathanwasson Feb 25, 2025
6a90ae4
Auto-merge updates from master branch
invalid-email-address Feb 27, 2025
7bf2c5e
Merge from Master (#2155)
arjunsuresh Mar 13, 2025
2a73202
[Automated Commit] Format Codebase
github-actions[bot] Mar 13, 2025
7ad9af1
Auto-merge updates from master branch
invalid-email-address Apr 15, 2025
2fb1057
Inference docs - Update model and dataset download commands (#2153)
anandhu-eng Apr 22, 2025
d804837
add powershell command to get result folder structure (#2156)
anandhu-eng Apr 22, 2025
f09fb9d
Auto-merge updates from master branch
invalid-email-address Apr 22, 2025
046b57d
Auto-merge updates from master branch
invalid-email-address Apr 22, 2025
5b65f7b
[Automated Commit] Format Codebase
github-actions[bot] Apr 22, 2025
ee16c1a
Merge branch 'master' into dev
arjunsuresh Jun 10, 2025
70fcbe0
[Automated Commit] Format Codebase
github-actions[bot] Jun 10, 2025
a5f12c7
Auto-merge updates from master branch
invalid-email-address Jun 10, 2025
3868556
Auto-merge updates from master branch
invalid-email-address Jun 10, 2025
18d49b4
Merge branch 'master' into dev
arjunsuresh Jun 10, 2025
a91453b
Auto-merge updates from master branch
invalid-email-address Jun 11, 2025
d288662
Auto-merge updates from master branch
invalid-email-address Jun 11, 2025
88324f3
Auto-merge updates from master branch
invalid-email-address Jun 16, 2025
ac42a2f
[Automated Commit] Format Codebase
github-actions[bot] Jun 16, 2025
53ae823
Auto-merge updates from master branch
invalid-email-address Jun 16, 2025
2b371c9
[Automated Commit] Format Codebase
github-actions[bot] Jun 16, 2025
9d5abe7
Auto-merge updates from master branch
invalid-email-address Jun 16, 2025
b896d58
Auto-merge updates from master branch
invalid-email-address Jun 17, 2025
92d37ac
Auto-merge updates from master branch
invalid-email-address Jun 19, 2025
72d52bb
[Automated Commit] Format Codebase
github-actions[bot] Jun 19, 2025
3af0227
Auto-merge updates from master branch
invalid-email-address Jun 23, 2025
c8ed997
Auto-merge updates from master branch
invalid-email-address Jun 24, 2025
fa3987f
Auto-merge updates from master branch
invalid-email-address Jun 24, 2025
dbe9632
Auto-merge updates from master branch
invalid-email-address Jun 24, 2025
ace438c
Auto-merge updates from master branch
invalid-email-address Jun 24, 2025
f841519
Fix Typo in Interactive Latencies (#2147) (#2225)
arjunsuresh Jun 26, 2025
4e77177
Fix Typo in Interactive Latencies (#2147) (#2226)
arjunsuresh Jun 26, 2025
56f8acf
Merge pull request #2227 from mlcommons/master
arjunsuresh Jun 26, 2025
868dd2b
Auto-merge updates from master branch
invalid-email-address Jul 2, 2025
350032a
Update MLCFlow commands for v5.1 (#2237)
anandhu-eng Jul 5, 2025
3a8595a
[Automated Commit] Format Codebase
github-actions[bot] Jul 5, 2025
184f0e8
Auto-merge updates from master branch
invalid-email-address Jul 7, 2025
381d239
Auto-merge updates from master branch
invalid-email-address Jul 8, 2025
00bb55f
Auto-merge updates from master branch
invalid-email-address Jul 8, 2025
a79004e
Auto-merge updates from master branch
invalid-email-address Jul 8, 2025
ed94d7a
Auto-merge updates from master branch
invalid-email-address Jul 8, 2025
1f4d76a
Auto-merge updates from master branch
invalid-email-address Jul 8, 2025
4e178b6
Auto-merge updates from master branch
invalid-email-address Jul 8, 2025
7c4f919
Auto-merge updates from master branch
invalid-email-address Jul 8, 2025
f77d066
Auto-merge updates from master branch
invalid-email-address Jul 15, 2025
906c0fc
[Automated Commit] Format Codebase
github-actions[bot] Jul 15, 2025
111063c
Auto-merge updates from master branch
invalid-email-address Jul 15, 2025
1d252a9
Auto-merge updates from master branch
invalid-email-address Jul 16, 2025
8980b1a
Auto-merge updates from master branch
invalid-email-address Jul 16, 2025
024e4ad
Update main.py
anandhu-eng Jul 17, 2025
9e2c7a9
[Automated Commit] Format Codebase
github-actions[bot] Jul 17, 2025
0928e11
updating for 5.1-dev (inference doc)
anandhu-eng Jul 17, 2025
7069a9e
[Automated Commit] Format Codebase
github-actions[bot] Jul 17, 2025
253854f
fix typo
anandhu-eng Jul 17, 2025
2f64e8b
Auto-merge updates from master branch
invalid-email-address Jul 18, 2025
7710c4d
Auto-merge updates from master branch
invalid-email-address Jul 21, 2025
371d583
[Automated Commit] Format Codebase
github-actions[bot] Jul 21, 2025
49db182
Auto-merge updates from master branch
invalid-email-address Jul 22, 2025
c8eb259
Auto-merge updates from master branch
invalid-email-address Jul 22, 2025
8ded855
Auto-merge updates from master branch
invalid-email-address Jul 22, 2025
ea86fc0
Update main.py
anandhu-eng Jul 23, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 6 additions & 1 deletion compliance/nvidia/TEST06/run_verification.py
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,12 @@ def get_args():
"--scenario",
"-s",
required=True,
choices=["Offline", "Server", "Interactive", "SingleStream", "MultiStream"],
choices=[
"Offline",
"Server",
"Interactive",
"SingleStream",
"MultiStream"],
)
args = parser.parse_args()
return args
Expand Down
11 changes: 11 additions & 0 deletions docs/benchmarks/language/deepseek-r1.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
---
hide:
- toc
---

# Reasoning using DeepSeek-R1

=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "deepseek-r1", "reference", devices=["CUDA"]) }}
24 changes: 24 additions & 0 deletions docs/benchmarks/language/get-deepseek-r1-data.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,24 @@
---
hide:
- toc
---

# Reasoning using DeepSeek R1

## Dataset

The benchmark implementation run command will automatically download the validation and calibration datasets and do the necessary preprocessing. In case you want to download only the datasets, you can use the below commands.

=== "Validation"

### Get Validation Dataset
```
mlcr get,preprocessed,dataset,deepseek-r1,_validation,_mlc,_rclone --outdirname=<path to download> -j
```

=== "Calibration"

### Get Calibration Dataset
```
mlcr get,preprocessed,dataset,deepseek-r1,_calibration,_mlc,_rclone --outdirname=<path to download> -j
```
60 changes: 60 additions & 0 deletions docs/benchmarks/language/get-llama3_1-8b-data.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,60 @@
---
hide:
- toc
---

# Text Summarization using LLAMA3.1-8b

## Dataset

The benchmark implementation run command will automatically download the validation and calibration datasets and do the necessary preprocessing. In case you want to download only the datasets, you can use the below commands.

=== "Validation"

=== "Full dataset (Datacenter)"

### Get Validation Dataset
```
mlcr get,dataset,cnndm,_validation,_datacenter,_llama3,_mlc,_rclone --outdirname=<path to download> -j
```

=== "5000 samples (Edge)"

### Get Validation Dataset
```
mlcr get,dataset,cnndm,_validation,_edge,_llama3,_mlc,_rclone --outdirname=<path to download> -j
```

=== "Calibration"

### Get Calibration Dataset
```
mlcr get,dataset,cnndm,_calibration,_llama3,_mlc,_rclone --outdirname=<path to download> -j
```

- `--outdirname=<PATH_TO_DOWNLOAD_LLAMA3_405B_DATASET>` could be provided to download the dataset to a specific location.

## Model
The benchmark implementation run command will automatically download the required model and do the necessary conversions. In case you want to only download the official model, you can use the below commands.

=== "Pytorch"

=== "From MLCOMMONS Google Drive"

> **Note:** One has to accept the [MLCommons Llama 3.1 License Confidentiality Notice](http://llama3-1.mlcommons.org/) to access the model files in MLCOMMONS Google Drive.

### Get the Official MLPerf LLAMA3.1-405B model from MLCOMMONS Cloudfare R2
```
TBD
```

=== "From Hugging Face repo"

> **Note:** Access to the HuggingFace model could be requested [here](https://ai.meta.com/resources/models-and-libraries/llama-downloads/).

### Get model from HuggingFace repo
```
mlcr get,ml-model,llama3,_hf,_meta-llama/Llama-3.1-8B-Instruct --hf_token=<huggingface access token> -j
```

- `--outdirname=<PATH_TO_DOWNLOAD_LLAMA3_8B_MODEL>` could be provided to download the model to a specific location.
11 changes: 11 additions & 0 deletions docs/benchmarks/language/llama3_1-8b.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
---
hide:
- toc
---

# Text Summarization using LLAMA3_1-8b

=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "llama3_1-8b", "reference", devices=["CPU","CUDA"]) }}
40 changes: 40 additions & 0 deletions docs/benchmarks/speech_to_text/get-whisper-data.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,40 @@
---
hide:
- toc
---

# Speech to Text using Whisper

## Dataset

The benchmark implementation run command will automatically download the validation and calibration datasets and do the necessary preprocessing. In case you want to download only the datasets, you can use the below commands.

=== "Validation"

=== "Preprocessed"

### Get Preprocessed Validation Dataset
```
mlcr get,dataset,whisper,_preprocessed,_mlc,_rclone --outdirname=<path to download> -j
```

=== "Unprocessed"

### Get Unprocessed Validation Dataset
```
mlcr get,dataset,whisper,_unprocessed --outdirname=<path to download> -j
```

## Model
The benchmark implementation run command will automatically download the required model and do the necessary conversions if any. In case you want to only download the official model, you can use the below commands.

=== "Pytorch"

=== "From MLCOMMONS"

### Get the Official MLPerf Whisper model from MLCOMMONS Cloudflare R2
```
mlcr get,ml-model,whisper,_rclone,_mlc s-j
```

- `--outdirname=<PATH_TO_DOWNLOAD_WHISPER_MODEL>` could be provided to download the model to a specific location.
11 changes: 11 additions & 0 deletions docs/benchmarks/speech_to_text/whisper.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
---
hide:
- toc
---

# Speech to Text using Whisper

=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "whisper", "reference", devices=["CPU","CUDA"]) }}
30 changes: 30 additions & 0 deletions language/deepseek-r1/README.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,11 @@
# Mlperf Inference DeepSeek Reference Implementation

## Automated command to run the benchmark via MLFlow

Please see the [new docs site](https://docs.mlcommons.org/inference/benchmarks/language/deepseek-r1/) for an automated way to run this benchmark across different available implementations and do an end-to-end submission with or without docker.

You can also do pip install mlc-scripts and then use `mlcr` commands for downloading the model and datasets using the commands given in the later sections.

## Model & Dataset Download

> **Model**: [deepseek-ai/DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1) (revision: `56d4cbbb4d29f4355bab4b9a39ccb717a14ad5ad`)
Expand All @@ -19,6 +25,14 @@ The dataset is an ensemble of the datasets: AIME, MATH500, gpqa, MMLU-Pro, livec

### Preprocessed

**Using MLCFlow Automation**

```
mlcr get,dataset,whisper,_preprocessed,_mlc,_rclone --outdirname=<path to download> -j
```

**Using Native method**

You can use Rclone to download the preprocessed dataset from a Cloudflare R2 bucket.

To run Rclone on Windows, you can download the executable [here](https://rclone.org/install/#windows).
Expand All @@ -38,6 +52,14 @@ rclone copy mlc-inference:mlcommons-inference-wg-public/deepseek_r1/datasets/mlp

### Calibration

**Using MLCFlow Automation**

```
mlcr get,preprocessed,dataset,deepseek-r1,_calibration,_mlc,_rclone --outdirname=<path to download> -j
```

**Using Native method**

Download and install Rclone as described in the previous section.

Then navigate in the terminal to your desired download directory and run the following command to download the dataset:
Expand Down Expand Up @@ -179,6 +201,14 @@ The following table shows which backends support different evaluation and MLPerf

## Accuracy Evaluation

**Using MLCFlow Automation**

```
TBD
```

**Using Native method**

Accuracy evaluation is handled uniformly across all backends:

```bash
Expand Down
52 changes: 40 additions & 12 deletions language/llama3.1-8b/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@

## Automated command to run the benchmark via MLFlow

Please see the [new docs site](https://docs.mlcommons.org/inference/benchmarks/language/llama3_1-8b/) (TBD) for an automated way to run this benchmark across different available implementations and do an end-to-end submission with or without docker.
Please see the [new docs site](https://docs.mlcommons.org/inference/benchmarks/language/llama3_1-8b/) for an automated way to run this benchmark across different available implementations and do an end-to-end submission with or without docker.

You can also do pip install mlc-scripts and then use `mlcr` commands for downloading the model and datasets using the commands given in the later sections.

Expand Down Expand Up @@ -99,7 +99,10 @@ pip install -e ../../loadgen
## Get Model
### MLCommons Members Download (Recommended for official submission)

You need to request for access to [MLCommons](http://llama3-1.mlcommons.org/) and you'll receive an email with the download instructions. You can download the model automatically via the below command
You need to request for access to [MLCommons](http://llama3-1.mlcommons.org/) and you'll receive an email with the download instructions.

**Official Model download using MLCFlow Automation**
You can download the model automatically via the below command
```
TBD
```
Expand All @@ -115,6 +118,12 @@ git clone https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct ${CHECKPOINT_P
cd ${CHECKPOINT_PATH} && git checkout be673f326cab4cd22ccfef76109faf68e41aa5f1
```

**External Model download using MLCFlow Automation**
You can download the model automatically via the below command
```
mlcr get,ml-model,llama3,_hf,_meta-llama/Llama-3.1-8B-Instruct --hf_token=<huggingface access token> -j
```

### Download huggingface model through MLC

```
Expand Down Expand Up @@ -142,24 +151,39 @@ rclone config create mlc-inference s3 provider=Cloudflare access_key_id=f65ba5ee
You can then navigate in the terminal to your desired download directory and run the following command to download the dataset:

#### Full dataset (datacenter)

**Using MLCFlow Automation**
```
mlcr get,dataset,cnndm,_validation,_datacenter,_llama3,_mlc,_rclone --outdirname=<path to download> -j
```

**Native method**
```
rclone copy mlc-inference:mlcommons-inference-wg-public/llama3.1_8b/datasets/cnn_eval.json ./ -P
```

#### 5000 samples (edge)

**Using MLCFlow Automation**
```
mlcr get,dataset,cnndm,_validation,_edge,_llama3,_mlc,_rclone --outdirname=<path to download> -j
```

**Native method**
```
rclone copy mlc-inference:mlcommons-inference-wg-public/llama3.1_8b/datasets/sample_cnn_eval_5000.json ./ -P
```

#### Calibration

**Using MLCFlow Automation**
```
rclone copy mlc-inference:mlcommons-inference-wg-public/llama3.1_8b/datasets/cnn_dailymail_calibration.json ./ -P
mlcr get,dataset,cnndm,_calibration,_llama3,_mlc,_rclone --outdirname=<path to download> -j
```

**MLC Command**

**Native method**
```
TBD
rclone copy mlc-inference:mlcommons-inference-wg-public/llama3.1_8b/datasets/cnn_dailymail_calibration.json ./ -P
```

You can also download the calibration dataset from the Cloudflare R2 bucket by running the following command:
Expand All @@ -168,11 +192,6 @@ You can also download the calibration dataset from the Cloudflare R2 bucket by r
rclone copy mlc-inference:mlcommons-inference-wg-public/llama3.1_8b/cnn_eval.json ./ -P
```

**MLC Command**
```
TBD
```


## Run Performance Benchmarks

Expand Down Expand Up @@ -265,8 +284,17 @@ The ServerSUT was not tested for GPU runs.

### Evaluate the accuracy using MLCFlow
You can also evaulate the accuracy from the generated accuracy log by using the following MLC command

**Full dataset (datacenter)**

```
TBD
mlcr run,accuracy,mlperf,_cnndm_llama_3,_edge --result_dir=<Path to directory where files are generated after the benchmark run>
```

**5000 samples (edge)**

```
mlcr run,accuracy,mlperf,_cnndm_llama_3,_datacenter --result_dir=<Path to directory where files are generated after the benchmark run>
```

## Accuracy Target
Expand Down
Loading