ipex-llm/python/llm/dev/benchmark/LongBench/config/model2maxlen.json
Xu, Shuo ee33b93464
Longbench: NV code to ipex-llm (#11662)
* add nv longbench

* LongBench: NV code to ipex-llm

* ammend

* add more models support

* ammend

* optimize LongBench's user experience

* ammend

* ammend

* fix typo

* ammend

* remove cuda related information & add a readme

* add license to python scripts & polish the readme

* ammend

* ammend

---------

Co-authored-by: cyita <yitastudy@gmail.com>
Co-authored-by: ATMxsp01 <shou.xu@intel.com>
Co-authored-by: leonardozcm <leonardo1997zcm@gmail.com>
2024-09-18 15:55:14 +08:00

18 lines
499 B
JSON

{
"llama2-7b-chat-4k": 4096,
"longchat-v1.5-7b-32k": 4096,
"xgen-7b-8k": 4096,
"internlm-7b-8k": 4096,
"chatglm2-6b": 4096,
"chatglm2-6b-32k": 4096,
"chatglm3-6b-32k": 4096,
"chatglm4-9b": 4096,
"vicuna-v1.5-7b-16k": 4096,
"mistral-7B-instruct-v0.2": 4096,
"mistral-7B-instruct-v0.1": 4096,
"mixtral-8x7B-instruct-v0.1": 4096,
"llama-2-7B-32k-instruct": 4096,
"lwm-text-chat-1m": 4096,
"lwm-text-1m": 4096,
"qwen2-7b-instruct": 4096
}