ipex-llm/python/llm/src/ipex_llm
2024-04-12 15:40:25 +08:00
..
cli Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
ggml LLM: support iq1_s (#10564) 2024-03-29 09:43:55 +08:00
gptq Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
langchain Add tokenizer_id in Langchain (#10588) 2024-04-03 14:25:35 +08:00
llamaindex Llamaindex: add tokenizer_id and support chat (#10590) 2024-04-07 13:51:34 +08:00
serving Replace ipex with ipex-llm (#10554) 2024-03-28 13:54:40 +08:00
transformers LLM: add bs limitation for llama softmax upcast to fp32 (#10752) 2024-04-12 15:40:25 +08:00
utils Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
vllm Remove not-imported MistralConfig (#10670) 2024-04-07 10:32:05 +08:00
__init__.py Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00
convert_model.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
format.sh Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
llm_patching.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
models.py Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
optimize.py LLM: fix torch_dtype setting of apply fp16 optimization through optimize_model (#10556) 2024-03-27 14:18:45 +08:00