ipex-llm/python/llm
2024-04-26 15:28:11 +08:00
..
dev Add benchmark script for pipeline parallel inference (#10873) 2024-04-26 15:28:11 +08:00
example LLM: Enable batch generate (world_size>1) in Deepspeed-AutoTP-FastAPI example (#10876) 2024-04-26 13:24:28 +08:00
portable-zip Fix baichuan-13b issue on portable zip under transformers 4.36 (#10746) 2024-04-12 16:27:01 -07:00
scripts Update scripts readme (#10725) 2024-04-25 17:24:37 +08:00
src/ipex_llm use new quantize kv (#10888) 2024-04-26 14:42:17 +08:00
test Add phi-3 to perf (#10883) 2024-04-25 20:21:56 +08:00
.gitignore [LLM] add chatglm pybinding binary file release (#8677) 2023-08-04 11:45:27 +08:00
setup.py Support llama-index install option for upstreaming purposes (#10866) 2024-04-23 19:08:29 +08:00
version.txt Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00