ipex-llm/python/llm/src/ipex_llm/ggml
Cengguang Zhang 9930351112
LLM: add new qtype woq_int4 to support gemm int4 temporary. (#12706)
This PR add temporary qtype woq_int4 to avoid affecting other qtype and models.

Co-authored-by: leonardozcm <leonardo1997zcm@gmail.com>
2025-01-15 14:41:33 +08:00
..
model Support imatrix-guided quantization for NPU CW (#12468) 2024-12-02 11:31:26 +08:00
__init__.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
convert.py Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
convert_model.py Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
quantize.py LLM: add new qtype woq_int4 to support gemm int4 temporary. (#12706) 2025-01-15 14:41:33 +08:00