ipex-llm/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations
Wang, Jian4 fe8976a00f LLM: Support gguf models use low_bit and fix no json(#10408)
* support others model use low_bit

* update readme

* update to add *.json
2024-03-15 09:34:18 +08:00
..
AWQ Support llm-awq backend (#9856) 2024-01-09 13:07:32 +08:00
GGUF LLM: Support gguf models use low_bit and fix no json(#10408) 2024-03-15 09:34:18 +08:00
GPTQ Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00