ipex-llm/python
Ruonan Wang 16433dd959 LLM: fix first token judgement of flash attention (#9841)
* fix flash attention

* meet code review

* fix
2024-01-05 13:49:37 +08:00
..
llm LLM: fix first token judgement of flash attention (#9841) 2024-01-05 13:49:37 +08:00