From 80a4c224cc9546e6427ec2326c84e7c31e1f847e Mon Sep 17 00:00:00 2001 From: Ian Date: Thu, 7 Dec 2023 16:39:50 +0000 Subject: [PATCH] Fixes cpu flag update for stateless llama from iree bump --- python/turbine_models/custom_models/stateless_llama.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/turbine_models/custom_models/stateless_llama.py b/python/turbine_models/custom_models/stateless_llama.py index 0d3d749d2..61727d673 100644 --- a/python/turbine_models/custom_models/stateless_llama.py +++ b/python/turbine_models/custom_models/stateless_llama.py @@ -237,7 +237,7 @@ def forward(token0: torch.Tensor, *state0_flat): "--iree-opt-const-expr-hoisting=False", ] if device == "cpu": - flags.append("--iree-llvmcpu-enable-microkernels") + flags.append("--iree-llvmcpu-enable-ukernels=all") device = "llvm-cpu" elif device == "vulkan": flags.extend(