ipex-llm/python/llm/example/GPU/Speculative-Decoding
Ruonan Wang 3e601f9a5d LLM: Support speculative decoding in bigdl-llm (#9951)
* first commit

* fix error, add llama example

* hidden print

* update api usage

* change to api v3

* update

* meet code review

* meet code review, fix style

* add reference, fix style

* fix style

* fix first token time
2024-01-22 19:14:56 +08:00
..
llama2 LLM: Support speculative decoding in bigdl-llm (#9951) 2024-01-22 19:14:56 +08:00