ipex-llm/python/llm/src/ipex_llm
2024-12-12 17:56:30 +08:00
..
cli Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
ggml [NPU] initial support of asym_int4_rtn (#12484) 2024-12-05 17:40:36 +08:00
gptq Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
langchain Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
llamaindex Llamaindex: add tokenizer_id and support chat (#10590) 2024-04-07 13:51:34 +08:00
serving Upgrade to vllm 0.6.2 (#12338) 2024-11-12 20:35:34 +08:00
transformers [NPU] Fix abnormal output with latest driver (#12530) 2024-12-12 17:56:30 +08:00
utils [NPU] Support glm-edge models (#12511) 2024-12-09 14:06:27 +08:00
vllm add vLLM glm4 fix (#12474) 2024-12-02 14:05:16 +08:00
__init__.py IPEX Duplicate importer V2 (#11310) 2024-06-19 16:29:19 +08:00
convert_model.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
format.sh Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
llm_patching.py Upgrade Peft version to 0.10.0 for LLM finetune (#10886) 2024-05-07 15:09:14 +08:00
models.py Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
optimize.py fix and optimize sd (#12436) 2024-11-25 14:09:48 +08:00