ipex-llm/python/llm/example
Yang Wang 51bcac1229
follow up on experimental support of fused decoder layer for llama2 (#11785)
* clean up and support transpose value cache

* refine

* fix style

* fix style
2024-08-13 18:53:55 -07:00
..
CPU upgrade glm-4v example transformers version (#11719) 2024-08-06 14:55:09 +08:00
GPU deepspeed zero3 QLoRA finetuning (#11625) 2024-08-13 16:15:29 +08:00
NPU/HF-Transformers-AutoModels follow up on experimental support of fused decoder layer for llama2 (#11785) 2024-08-13 18:53:55 -07:00