ipex-llm/python/llm
Yina Chen f24352aef9
llama 3.1/3.2 support compresskv (#12347)
* llama 3.1/3.2 support compresskv

* update

* fix transformers 4.45 error

* fix style

* fix typo

* disable llama3.2 1b compresskv
2024-11-06 17:33:43 +08:00
..
dev Add dummy model in iGPU perf (#12341) 2024-11-05 17:56:10 +08:00
example [NPU] Llama3, Qwen2 1.5b, MiniCPM 1/2B groupwise support (#12327) 2024-11-05 15:51:31 +08:00
portable-zip Fix null pointer dereferences error. (#11125) 2024-05-30 16:16:10 +08:00
scripts fix typo in python/llm/scripts/README.md (#11536) 2024-07-09 09:53:14 +08:00
src/ipex_llm llama 3.1/3.2 support compresskv (#12347) 2024-11-06 17:33:43 +08:00
test Add MiniCPM-V-2_6 to arc perf test (#12349) 2024-11-06 16:32:28 +08:00
tpp OSPDT: add tpp licenses (#11165) 2024-06-06 10:59:06 +08:00
.gitignore [LLM] add chatglm pybinding binary file release (#8677) 2023-08-04 11:45:27 +08:00
setup.py Support cpp release for ARL on Windows (#12189) 2024-10-14 17:20:31 +08:00
version.txt Update pypi tag to 2.2.0.dev0 (#11895) 2024-08-22 16:48:09 +08:00