-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Addition of 4 pytorch onnx-export models. #278
base: main
Are you sure you want to change the base?
Changes from all commits
a45c418
9b9f7b1
10d44e7
a5c0f77
a088614
a0d2d52
7501f7b
8411a4a
bb90c56
3d9bd2c
34aff57
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -2,20 +2,19 @@ | |
"config_name": "cpu_llvm_task", | ||
"iree_compile_flags" : [ | ||
"--iree-hal-target-backends=llvm-cpu", | ||
"--iree-llvmcpu-target-cpu-features=host" | ||
"--iree-llvmcpu-target-cpu-features=host", | ||
"--iree-input-demote-i64-to-i32" | ||
], | ||
"iree_run_module_flags": [ | ||
"--device=local-task" | ||
], | ||
"skip_compile_tests": [ | ||
"pytorch/models/sdxl-scheduled-unet-3-tank", | ||
"pytorch/models/sdxl-vae-decode-tank", | ||
"pytorch/models/sdxl-prompt-encoder-tank" | ||
], | ||
"skip_compile_tests": [], | ||
"skip_run_tests": [], | ||
"expected_compile_failures": [ | ||
"pytorch/models/opt-125M", // TODO(#17344): need to regenerate .mlirbc | ||
"pytorch/models/resnet50", // TODO(#17344): need to regenerate .mlirbc | ||
"pytorch/models/direct/opt-125M", // TODO(#17344): need to regenerate .mlirbc | ||
"pytorch/models/direct/resnet50", // TODO(#17344): need to regenerate .mlirbc | ||
], | ||
"expected_run_failures": [] | ||
"expected_run_failures": [ | ||
"pytorch/models/onnx-export/mobilebert-uncased", | ||
] | ||
Comment on lines
+17
to
+19
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
Can we add models that we think are stable too? :) I'd like to prefetch all models from e2eshark and get them running here if possible, rather than cherrypick a few after we notice issues. |
||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -2,25 +2,23 @@ | |
"config_name": "gpu_rocm", | ||
"iree_compile_flags" : [ | ||
"--iree-hal-target-backends=rocm", | ||
"--iree-rocm-target-chip=gfx90a", | ||
"--iree-opt-const-eval=false", | ||
"--iree-codegen-transform-dialect-library=${IREE_TEST_PATH_EXTENSION}/attention_and_matmul_spec.mlir" | ||
"--iree-rocm-target-chip=gfx90a" | ||
], | ||
"iree_run_module_flags": [ | ||
"--device=hip" | ||
], | ||
"skip_compile_tests": [ | ||
"pytorch/models/sdxl-scheduled-unet-3-tank", | ||
"pytorch/models/sdxl-vae-decode-tank", | ||
"pytorch/models/sdxl-prompt-encoder-tank" | ||
], | ||
"skip_compile_tests": [], | ||
"skip_run_tests": [], | ||
"expected_compile_failures": [ | ||
"pytorch/models/opt-125M", // TODO(#17344): need to regenerate .mlirbc | ||
"pytorch/models/resnet50", | ||
"pytorch/models/direct/opt-125M", // TODO(#17344): need to regenerate .mlirbc | ||
"pytorch/models/direct/resnet50", | ||
// error: 'builtin.module' op failed to run transform dialect passes | ||
// (might need to drop the iree-codegen-transform-dialect-library flag) | ||
Comment on lines
15
to
16
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Can drop this comment since the How about we just leave the reason comments off? Also on my PR there, if we run with |
||
"sharktank/llama/open-llama-3b-v2-f16" | ||
"sharktank/llama/open-llama-3b-v2-f16", | ||
"pytorch/models/onnx-export/mit-b0", | ||
"pytorch/models/onnx-export/mobilebert-uncased", | ||
"pytorch/models/onnx-export/t5-base", | ||
"pytorch/models/onnx-export/t5-large" | ||
], | ||
"expected_run_failures": [] | ||
} |
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
That seems rather roundabout... I guess ONNX there is used as a serialization format? IMO if we're adding tests for these models we should add both the direct and onnx-export versions. |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,3 @@ | ||
--input=1x3x512x512xf32=@inference_input.0.bin | ||
--expected_output=1x1000xf32=@inference_output.0.bin | ||
--expected_f32_threshold=0.01f |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,13 @@ | ||
{ | ||
"file_format": "test_cases_v0", | ||
"test_cases": [ | ||
{ | ||
"name": "real_weights", | ||
"runtime_flagfile": "real_weights_data_flags.txt", | ||
"remote_files": [ | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/mit-b0/inference_input.0.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/mit-b0/inference_output.0.bin" | ||
] | ||
} | ||
] | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,3 @@ | ||
--input=1x128xi64=@inference_input.0.bin | ||
--expected_output=1x2xf32=@inference_output.0.bin | ||
--expected_f32_threshold=0.01f |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,14 @@ | ||
{ | ||
"file_format": "test_cases_v0", | ||
"test_cases": [ | ||
{ | ||
"name": "real_weights", | ||
"runtime_flagfile": "real_weights_data_flags.txt", | ||
"remote_files": [ | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/mobilebert-uncased/inference_input.0.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/mobilebert-uncased/inference_output.0.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/mobilebert-uncased/model.mlirbc" | ||
] | ||
} | ||
] | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,54 @@ | ||
--input=1x512xi64=@inference_input.0.bin | ||
--input=1x512xf32=@inference_input.1.bin | ||
--input=1x512xi64=@inference_input.2.bin | ||
--expected_output=1x512x768xf32=@inference_output.0.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.1.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.2.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.3.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.4.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.5.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.6.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.7.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.8.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.9.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.10.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.11.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.12.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.13.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.14.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.15.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.16.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.17.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.18.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.19.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.20.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.21.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.22.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.23.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.24.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.25.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.26.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.27.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.28.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.29.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.30.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.31.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.32.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.33.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.34.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.35.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.36.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.37.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.38.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.39.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.40.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.41.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.42.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.43.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.44.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.45.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.46.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.47.bin | ||
--expected_output=1x12x512x64xf32=@inference_output.48.bin | ||
--expected_output=1x512x768xf32=@inference_output.49.bin | ||
--expected_f32_threshold=0.01f |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,65 @@ | ||
{ | ||
"file_format": "test_cases_v0", | ||
"test_cases": [ | ||
{ | ||
"name": "real_weights", | ||
"runtime_flagfile": "real_weights_data_flags.txt", | ||
"remote_files": [ | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_input.0.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_input.1.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_input.2.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.0.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.1.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.2.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.3.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.4.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.5.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.6.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.7.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.8.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.9.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.10.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.11.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.12.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.13.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.14.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.15.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.16.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.17.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.18.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.19.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.20.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.21.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.22.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.23.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.24.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.25.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.26.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.27.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.28.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.29.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.30.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.31.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.32.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.33.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.34.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.35.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.36.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.37.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.38.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.39.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.40.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.41.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.42.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.43.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.44.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.45.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.46.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.47.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.48.bin", | ||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/inference_output.49.bin", | ||
Comment on lines
+57
to
+60
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. o_o that's a lot of outputs. Fine for now, but let's file an issue to use archive files (.zip or .tar.gz). If we use something like Hugging Face repositories, we could also lean on their download API to fetch an entire repo |
||
"https://sharkpublic.blob.core.windows.net/sharkpublic/shark-test-suite/iree-tests/pytorch/models/onnx-export/t5-base/model.mlirbc" | ||
] | ||
} | ||
] | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not sure I agree with this. I'm worried about the number of flags we have developers using, and I'd like to have more visibility into which models require which flags and why. One way we can do that is by having users of the test suite test with default flags.
In this case, the
--iree-input-
flags are sort of special in that they are generic across backends so I do see a case for using some of them... maybe in a_compatibility_flags
test suite config, then we could diff the XFAILs between that and the_default_flags
config.