Skip to content

Commit

Permalink
Merge branch 'master' into amdenable
Browse files Browse the repository at this point in the history
  • Loading branch information
anandhu-eng authored Oct 23, 2024
2 parents d3c332d + fe99768 commit 2caa6e2
Show file tree
Hide file tree
Showing 4 changed files with 544 additions and 442 deletions.
2 changes: 1 addition & 1 deletion docs/benchmarks/language/reproducibility/indyscc24-bert.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ All the needed files are automatically pushed to the GitHub repository if you ma
=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "bert-99", "reference", extra_variation_tags="", scenarios=["Offline"],categories=["Edge"], setup_tips=False) }}
{{ mlperf_inference_implementation_readme (4, "bert-99", "reference", extra_variation_tags="", fixed_scenarios=["Offline"],categories=["Edge"], setup_tips=False) }}


## Submission Commands
Expand Down
4 changes: 2 additions & 2 deletions docs/benchmarks/text_to_image/reproducibility/scc24.md
Original file line number Diff line number Diff line change
Expand Up @@ -46,11 +46,11 @@ or supporting multi-node execution) useful for the community and [MLCommons](htt
=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "sdxl", "reference", extra_variation_tags=",_short,_scc24-base", devices=["ROCm", "CUDA"],scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, skip_test_query_count=True, extra_input_string="--precision=float16") }}
{{ mlperf_inference_implementation_readme (4, "sdxl", "reference", extra_variation_tags=",_short,_scc24-base", devices=["ROCm", "CUDA"],fixed_scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, skip_test_query_count=True, extra_input_string="--precision=float16") }}

=== "Nvidia"
## Nvidia MLPerf Implementation
{{ mlperf_inference_implementation_readme (4, "sdxl", "nvidia", extra_variation_tags=",_short,_scc24-base", scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, implementation_tips=False, skip_test_query_count=True) }}
{{ mlperf_inference_implementation_readme (4, "sdxl", "nvidia", extra_variation_tags=",_short,_scc24-base", fixed_scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, implementation_tips=False, skip_test_query_count=True) }}

!!! info
Once the above run is successful, you can change `_scc24-base` to `_scc24-main` to run the main variant.
Expand Down
Loading

0 comments on commit 2caa6e2

Please sign in to comment.