ipex-llm/python/llm/src/ipex_llm
Cengguang Zhang 763413b7e1
LLM: support llama split tensor for long context in transformers>=4.36. (#10844)
* LLm: support llama split tensor for long context in transformers>=4.36.

* fix dtype.

* fix style.

* fix style.

* fix style.

* fix style.

* fix dtype.

* fix style.
2024-04-23 16:13:25 +08:00
..
cli Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
ggml Support q4k in ipex-llm (#10796) 2024-04-18 18:55:28 +08:00
gptq Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
langchain Add tokenizer_id in Langchain (#10588) 2024-04-03 14:25:35 +08:00
llamaindex Llamaindex: add tokenizer_id and support chat (#10590) 2024-04-07 13:51:34 +08:00
serving [vLLM]Remove vllm-v1, refactor v2 (#10842) 2024-04-22 17:51:32 +08:00
transformers LLM: support llama split tensor for long context in transformers>=4.36. (#10844) 2024-04-23 16:13:25 +08:00
utils Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
vllm [vLLM]Remove vllm-v1, refactor v2 (#10842) 2024-04-22 17:51:32 +08:00
__init__.py Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00
convert_model.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
format.sh Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
llm_patching.py Axolotl v0.4.0 support (#10773) 2024-04-17 09:49:11 +08:00
models.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
optimize.py Add vLLM[xpu] related code (#10779) 2024-04-18 15:29:20 +08:00