ipex-llm/python
Ruonan Wang a00efa0564 LLM: add mlp & qkv fusion for FP16 Llama-7B (#9932)
* add mlp fusion for llama

* add mlp fusion

* fix style

* update

* add mm_qkv_out

* fix style

* update

* meet code review

* meet code review
2024-01-26 11:50:38 +08:00
..
llm LLM: add mlp & qkv fusion for FP16 Llama-7B (#9932) 2024-01-26 11:50:38 +08:00