ipex-llm/python/llm/src/ipex_llm
Cengguang Zhang 7c43ac0164
LLM: optimize llama natvie sdp for split qkv tensor (#10693)
* LLM: optimize llama natvie sdp for split qkv tensor.

* fix block real size.

* fix comment.

* fix style.

* refactor.
2024-04-08 17:48:11 +08:00
..
cli Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
ggml LLM: support iq1_s (#10564) 2024-03-29 09:43:55 +08:00
gptq Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
langchain Add tokenizer_id in Langchain (#10588) 2024-04-03 14:25:35 +08:00
llamaindex Llamaindex: add tokenizer_id and support chat (#10590) 2024-04-07 13:51:34 +08:00
serving Replace ipex with ipex-llm (#10554) 2024-03-28 13:54:40 +08:00
transformers LLM: optimize llama natvie sdp for split qkv tensor (#10693) 2024-04-08 17:48:11 +08:00
utils Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
vllm Remove not-imported MistralConfig (#10670) 2024-04-07 10:32:05 +08:00
__init__.py Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00
convert_model.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
format.sh Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
llm_patching.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
models.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
optimize.py LLM: fix torch_dtype setting of apply fp16 optimization through optimize_model (#10556) 2024-03-27 14:18:45 +08:00