| .. |
|
cli
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
ggml
|
Init NPU quantize method and support q8_0_rtn (#11452)
|
2024-07-01 13:45:07 +08:00 |
|
gptq
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
langchain
|
Remove chatglm_C Module to Eliminate LGPL Dependency (#11178)
|
2024-05-31 17:03:11 +08:00 |
|
llamaindex
|
Llamaindex: add tokenizer_id and support chat (#10590)
|
2024-04-07 13:51:34 +08:00 |
|
serving
|
add new benchmark_util.py (#11713)
|
2024-08-05 16:18:48 +08:00 |
|
transformers
|
support and optimize minicpm-v-2_6 (#11738)
|
2024-08-07 18:21:16 +08:00 |
|
utils
|
Add benchmark util for transformers 4.42 (#11725)
|
2024-08-07 08:48:07 +08:00 |
|
vllm
|
Fix vLLM CPU /chat endpoint (#11748)
|
2024-08-09 10:33:52 +08:00 |
|
__init__.py
|
IPEX Duplicate importer V2 (#11310)
|
2024-06-19 16:29:19 +08:00 |
|
convert_model.py
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
format.sh
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
llm_patching.py
|
Upgrade Peft version to 0.10.0 for LLM finetune (#10886)
|
2024-05-07 15:09:14 +08:00 |
|
models.py
|
Remove chatglm_C Module to Eliminate LGPL Dependency (#11178)
|
2024-05-31 17:03:11 +08:00 |
|
optimize.py
|
Update tests for transformers 4.36 (#10858)
|
2024-05-24 10:26:38 +08:00 |