ipex-llm/python/llm/example
Ruonan Wang 0819fad34e
support Llama2-7B / Llama3-8B for NPU C++ (#12431)
* support llama2

* update

* support fused_layers=4 for Llama2-7B
2024-11-22 18:47:19 +08:00
..
CPU Limit trl version in example (#12332) 2024-11-05 14:50:10 +08:00
GPU add optimization to openjourney (#12423) 2024-11-21 15:23:51 +08:00
NPU/HF-Transformers-AutoModels support Llama2-7B / Llama3-8B for NPU C++ (#12431) 2024-11-22 18:47:19 +08:00