Update runtime configuration on MTL (#12720)
This commit is contained in:
parent
bda87c21eb
commit
9aa4be8ced
1 changed files with 0 additions and 1 deletions
|
|
@ -34,7 +34,6 @@ def optimize_llm_pre(model: torch.nn.Module, qtype, mixed_precision,
|
|||
quantization_group_size=0, load=False, max_prompt_len=512):
|
||||
if os.environ.get("IPEX_LLM_NPU_MTL", "0") == "1":
|
||||
# For MTL support
|
||||
os.environ["IPEX_LLM_NPU_USE_LEVEL0"] = "0"
|
||||
os.environ["IPEX_LLM_NPU_DISABLE_COMPILE_OPT"] = "1"
|
||||
|
||||
if os.environ.get("IPEX_LLM_NPU_ARL", "0") == "1":
|
||||
|
|
|
|||
Loading…
Reference in a new issue