ipex-llm/python/llm
SONG Ge 0c67639539
Add more examples for pipeline parallel inference (#11372)
* add more model exampels for pipelien parallel inference

* add mixtral and vicuna models

* add yi model and past_kv supprot for chatglm family

* add docs

* doc update

* add license

* update
2024-06-21 17:55:16 +08:00
..
dev Fix run.py run_ipex_fp16_gpu (#11361) 2024-06-20 10:29:32 +08:00
example Add more examples for pipeline parallel inference (#11372) 2024-06-21 17:55:16 +08:00
portable-zip Fix null pointer dereferences error. (#11125) 2024-05-30 16:16:10 +08:00
scripts Miniconda/Anaconda -> Miniforge update in examples (#11194) 2024-06-04 10:14:02 +08:00
src/ipex_llm Add more examples for pipeline parallel inference (#11372) 2024-06-21 17:55:16 +08:00
test Add qwen-moe batch1 to nightly perf (#11369) 2024-06-20 14:17:41 +08:00
tpp OSPDT: add tpp licenses (#11165) 2024-06-06 10:59:06 +08:00
.gitignore [LLM] add chatglm pybinding binary file release (#8677) 2023-08-04 11:45:27 +08:00
setup.py Upgrade accelerate to 0.23.0 (#11331) 2024-06-17 15:03:11 +08:00
version.txt Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00