| .. |
|
cli
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
ggml
|
support gguf_q4k_m / gguf_q4k_s (#10887)
|
2024-05-17 14:30:09 +08:00 |
|
gptq
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
langchain
|
add langchain vllm interface (#11121)
|
2024-05-24 17:19:27 +08:00 |
|
llamaindex
|
Llamaindex: add tokenizer_id and support chat (#10590)
|
2024-04-07 13:51:34 +08:00 |
|
serving
|
[Docker] Fix image using two cards error (#11144)
|
2024-05-27 16:20:13 +08:00 |
|
transformers
|
Support new fp8 e4m3 (#11158)
|
2024-05-29 14:27:14 +08:00 |
|
utils
|
Update benchmark util for example using (#11027)
|
2024-05-15 14:16:35 +08:00 |
|
vllm
|
LLM: Add CPU vLLM entrypoint (#11083)
|
2024-05-24 09:16:59 +08:00 |
|
__init__.py
|
Update setup.py and add new actions and add compatible mode (#25)
|
2024-03-22 15:44:59 +08:00 |
|
convert_model.py
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
format.sh
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
llm_patching.py
|
Upgrade Peft version to 0.10.0 for LLM finetune (#10886)
|
2024-05-07 15:09:14 +08:00 |
|
models.py
|
Refactor bigdl.llm to ipex_llm (#24)
|
2024-03-22 15:41:21 +08:00 |
|
optimize.py
|
Update tests for transformers 4.36 (#10858)
|
2024-05-24 10:26:38 +08:00 |