ipex-llm/python
Yang Wang 25428b22b4 Fix chatglm2 attention and kv cache (#8924)
* fix chatglm2 attention

* fix bf16 bug

* make model stateless

* add utils

* cleanup

* fix style
2023-09-07 18:54:29 -07:00
..
llm Fix chatglm2 attention and kv cache (#8924) 2023-09-07 18:54:29 -07:00