ipex-llm/python
Cengguang Zhang 30d009bca7 LLM: support quantized kv cache for Mistral in transformers >=4.36.0 (#10326)
* support quantize kv for mistral in transformers 4.36

* update mistral support.

* fix style.
2024-03-05 16:23:50 +08:00
..
llm LLM: support quantized kv cache for Mistral in transformers >=4.36.0 (#10326) 2024-03-05 16:23:50 +08:00