Skip to content

Conversation

@keithachorn-intel
Copy link
Contributor

The updated value represents the dataset size, not the 'performance_sample_count_override' value set in mlperf.conf. After adding the new submission checker test, this mismatch raises an error which was not previously present.

@keithachorn-intel keithachorn-intel requested a review from a team as a code owner June 24, 2025 21:36
@github-actions
Copy link
Contributor

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@keithachorn-intel
Copy link
Contributor Author

If this could possibly be merged today, it would be very helpful in shipping our workload containers. Thanks! @arjunsuresh

@attafosu
Copy link
Contributor

@arjunsuresh There seem to be a mismatch between the dataset size (204k) from mlcommons rules table, and what is generated (330k) from multihot_criteo.py's num_aggregated_samples
Note that it is also this count which is used as the total_samples_count by loadgen, and it makes its way into the mlperf_log_detail.txt as the qsl_reported_total_count.

@arjunsuresh
Copy link
Contributor

@arjunsuresh
Copy link
Contributor

@pgmpablo157321 can you please make the relevant change in the rules?

@arjunsuresh arjunsuresh merged commit 064db01 into mlcommons:master Jun 24, 2025
21 checks passed
@github-actions github-actions bot locked and limited conversation to collaborators Jun 24, 2025
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants