ipex-llm/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations
Heyang Sun 1fa7793fc0 Load Mixtral GGUF Model (#9690)
* Load Mixtral GGUF Model

* refactor

* fix empty tensor when to cpu

* update gpu and cpu readmes

* add dtype when set tensor into module
2023-12-19 13:54:38 +08:00
..
AWQ [LLM]Add Yi-34B-AWQ to verified AWQ model. (#9676) 2023-12-14 09:55:47 +08:00
GGUF Load Mixtral GGUF Model (#9690) 2023-12-19 13:54:38 +08:00
GPTQ Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00