* LLM: support llama2 8k input with w4a16. * fix comment and style. * fix style. * fix comments and split tensor to quantized attention forward. * fix style. * refactor name. * fix style. * fix style. * fix style. * refactor checker name. * refactor native sdp split qkv tensor name. * fix style. * fix comment rename variables. * fix co-exist of intermedia results. |
||
|---|---|---|
| .. | ||
| dev | ||
| example | ||
| portable-zip | ||
| scripts | ||
| src/ipex_llm | ||
| test | ||
| .gitignore | ||
| setup.py | ||
| version.txt | ||