ipex-llm/python/llm/example/GPU/HF-Transformers-AutoModels/Advanced-Quantizations
Ruonan Wang a9fd20b6ba LLM: Update qkv fusion for GGUF-IQ2 (#10271)
* first commit

* update mistral

* fix transformers==4.36.0

* fix

* disable qk for mixtral now

* fix style
2024-02-29 12:49:53 +08:00
..
AWQ [LLM] Support llm-awq vicuna-7b-1.5 on arc (#9874) 2024-01-10 14:28:39 +08:00
GGUF Fix Mixtral GGUF Wrong Output Issue (#9930) 2024-01-18 14:11:27 +08:00
GGUF-IQ2 LLM: Update qkv fusion for GGUF-IQ2 (#10271) 2024-02-29 12:49:53 +08:00
GPTQ Update llm gpu xpu default related info to PyTorch 2.1 (#9866) 2024-01-09 15:38:47 +08:00