ipex-llm/python/llm/src/bigdl
SONG Ge d5b81af7bd Support mixtral attention optimization on transformers-v4.36.0 (#9674)
* add example code to support mistral/mixtral attention on transformers v4.36.0

* update

* style fix

* add update for seen-tokens

* support mixtral

* rm mistral change

* small fix

* add more comments and remove use_cache part

---------

Co-authored-by: plusbang <binbin1.deng@intel.com>
2023-12-15 14:30:23 +08:00
..
llm Support mixtral attention optimization on transformers-v4.36.0 (#9674) 2023-12-15 14:30:23 +08:00
__init__.py LLM: add first round files (#8225) 2023-05-25 11:29:18 +08:00