ipex-llm/python/llm/src/ipex_llm
2024-05-10 17:33:46 +08:00
..
cli Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
ggml Support q4k in ipex-llm (#10796) 2024-04-18 18:55:28 +08:00
gptq Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
langchain Add tokenizer_id in Langchain (#10588) 2024-04-03 14:25:35 +08:00
llamaindex Llamaindex: add tokenizer_id and support chat (#10590) 2024-04-07 13:51:34 +08:00
serving Add embeddings (#10931) 2024-05-07 09:07:02 +08:00
transformers optimize phi3 1st token performance (#10981) 2024-05-10 17:33:46 +08:00
utils Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
vllm Add tensor parallel for vLLM (#10879) 2024-04-26 17:10:49 +08:00
__init__.py Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00
convert_model.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
format.sh Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
llm_patching.py Upgrade Peft version to 0.10.0 for LLM finetune (#10886) 2024-05-07 15:09:14 +08:00
models.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
optimize.py Add vLLM[xpu] related code (#10779) 2024-04-18 15:29:20 +08:00