ipex-llm/python/llm/src/ipex_llm
Xin Qiu 183e0c6cf5
glm-4v-9b support (#11327)
* chatglm4v support

* fix style check

* update glm4v
2024-06-17 13:52:37 +08:00
..
cli Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
ggml Add quantization scale search switch (#11326) 2024-06-14 18:46:52 +08:00
gptq Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
langchain Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
llamaindex Llamaindex: add tokenizer_id and support chat (#10590) 2024-04-07 13:51:34 +08:00
serving Expose timeout for streamer for fastchat worker (#11288) 2024-06-12 17:02:40 +08:00
transformers glm-4v-9b support (#11327) 2024-06-17 13:52:37 +08:00
utils Remove duplicate check for ipex (#11281) 2024-06-12 13:52:02 +08:00
vllm LLM: Fix vLLM CPU model convert mismatch (#11254) 2024-06-07 15:54:34 +08:00
__init__.py Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00
convert_model.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
format.sh Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
llm_patching.py Upgrade Peft version to 0.10.0 for LLM finetune (#10886) 2024-05-07 15:09:14 +08:00
models.py Remove chatglm_C Module to Eliminate LGPL Dependency (#11178) 2024-05-31 17:03:11 +08:00
optimize.py Update tests for transformers 4.36 (#10858) 2024-05-24 10:26:38 +08:00