ipex-llm/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations
Wang, Jian4 0193f29411 LLM : Enable gguf float16 and Yuan2 model (#10372)
* enable float16

* add yun files

* enable yun

* enable set low_bit on yuan2

* update

* update license

* update generate

* update readme

* update python style

* update
2024-03-13 10:19:18 +08:00
..
AWQ Support llm-awq backend (#9856) 2024-01-09 13:07:32 +08:00
GGUF LLM : Enable gguf float16 and Yuan2 model (#10372) 2024-03-13 10:19:18 +08:00
GPTQ Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00