ipex-llm/python/llm
Yang Wang 5a1f446d3c
support fp8 in xetla (#10555)
* support fp8 in xetla

* change name

* adjust model file

* support convert back to cpu

* factor

* fix bug

* fix style
2024-04-08 13:22:09 -07:00
..
dev add test api transformer_int4_fp16_gpu (#10627) 2024-04-07 15:47:17 +08:00
example LLM: upgrade deepspeed in AutoTP on GPU (#10647) 2024-04-07 14:05:19 +08:00
portable-zip Migrate portable zip to ipex-llm (#10617) 2024-04-07 13:58:58 +08:00
scripts LLM: check user env (#10580) 2024-03-29 17:19:34 +08:00
src/ipex_llm support fp8 in xetla (#10555) 2024-04-08 13:22:09 -07:00
test fix UT threshold (#10689) 2024-04-08 14:58:20 +08:00
.gitignore [LLM] add chatglm pybinding binary file release (#8677) 2023-08-04 11:45:27 +08:00
setup.py Update pip install to use --extra-index-url for ipex package (#10557) 2024-03-28 09:56:23 +08:00
version.txt Update setup.py and add new actions and add compatible mode (#25) 2024-03-22 15:44:59 +08:00