ipex-llm/python/llm/example/GPU/HF-Transformers-AutoModels/Advanced-Quantizations
Wang, Jian4 496bb2e845 LLM: Support load BaiChuan model family gguf model (#9685)
* support baichuan model family gguf model

* update gguf generate.py

* add verify models

* add support model_family

* update

* update style

* update type

* update readme

* update

* remove support model_family
2023-12-15 13:34:33 +08:00
..
AWQ [LLM]Add Yi-34B-AWQ to verified AWQ model. (#9676) 2023-12-14 09:55:47 +08:00
GGUF LLM: Support load BaiChuan model family gguf model (#9685) 2023-12-15 13:34:33 +08:00
GPTQ Support directly loading gptq models from huggingface (#9391) 2023-11-13 20:48:12 -08:00