ipex-llm/python/llm/src/ipex_llm
Zijie Li c410d9cf73
[NPU] support asym_int4 for baichuan (#12576)
* add npu support for baichuan

* Update baichuan_mp.py

* Update baichuan_mp.py
2024-12-24 09:17:50 +08:00
..
cli Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
ggml [NPU] initial support of asym_int4_rtn (#12484) 2024-12-05 17:40:36 +08:00
gptq Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
langchain Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
llamaindex Llamaindex: add tokenizer_id and support chat (#10590) 2024-04-07 13:51:34 +08:00
serving Upgrade to vllm 0.6.2 (#12338) 2024-11-12 20:35:34 +08:00
transformers [NPU] support asym_int4 for baichuan (#12576) 2024-12-24 09:17:50 +08:00
utils [NPU] Support glm-edge models (#12511) 2024-12-09 14:06:27 +08:00
vllm add vLLM glm4 fix (#12474) 2024-12-02 14:05:16 +08:00
__init__.py IPEX Duplicate importer V2 (#11310) 2024-06-19 16:29:19 +08:00
convert_model.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
format.sh Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
llm_patching.py Upgrade Peft version to 0.10.0 for LLM finetune (#10886) 2024-05-07 15:09:14 +08:00
models.py Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
optimize.py fix and optimize sd (#12436) 2024-11-25 14:09:48 +08:00