From 2fbd375a94b4e8e753c34b154265b3d0592f3f1d Mon Sep 17 00:00:00 2001 From: RyuKosei <70006706+RyuKosei@users.noreply.github.com> Date: Wed, 24 Jul 2024 23:06:08 -0700 Subject: [PATCH] update several models for nightly perf test (#11643) Co-authored-by: Yishuo Wang --- python/llm/test/benchmark/arc-perf-test-batch2.yaml | 5 ++++- python/llm/test/benchmark/arc-perf-test-batch4.yaml | 7 +++++-- python/llm/test/benchmark/arc-perf-test.yaml | 5 ++++- .../test/benchmark/arc-perf-transformers-437-batch2.yaml | 1 + .../test/benchmark/arc-perf-transformers-437-batch4.yaml | 1 + python/llm/test/benchmark/arc-perf-transformers-437.yaml | 1 + 6 files changed, 16 insertions(+), 4 deletions(-) diff --git a/python/llm/test/benchmark/arc-perf-test-batch2.yaml b/python/llm/test/benchmark/arc-perf-test-batch2.yaml index 88f3c8b5..e5863465 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch2.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch2.yaml @@ -8,9 +8,12 @@ repo_id: - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - #- 'Qwen/Qwen-VL-Chat' + - 'Qwen/Qwen-VL-Chat' #- 'SmerkyG/rwkv-5-world-7b' #this model only fp32 is supported for now, fp16 and bf16 are not supported - '01-ai/Yi-6B-Chat' + - 'mistralai/Mistral-7B-Instruct-v0.2' + - 'deepseek-ai/deepseek-coder-7b-instruct-v1.5' + - '01-ai/Yi-1.5-6B-Chat' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-test-batch4.yaml b/python/llm/test/benchmark/arc-perf-test-batch4.yaml index b382bed9..0c83cc9b 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch4.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch4.yaml @@ -8,9 +8,12 @@ repo_id: - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - #- 'Qwen/Qwen-VL-Chat' + - 'Qwen/Qwen-VL-Chat' #- 'SmerkyG/rwkv-5-world-7b' #this model only fp32 is supported for now, fp16 and bf16 are not supported - '01-ai/Yi-6B-Chat' + - 'mistralai/Mistral-7B-Instruct-v0.2' + - 'deepseek-ai/deepseek-coder-7b-instruct-v1.5' + - '01-ai/Yi-1.5-6B-Chat' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 @@ -29,7 +32,7 @@ exclude: - 'baichuan-inc/Baichuan2-7B-Chat:2048' - 'baichuan-inc/Baichuan2-13B-Chat-4bit:1024' - 'baichuan-inc/Baichuan2-13B-Chat-4bit:2048' -# - 'Qwen/Qwen-VL-Chat:2048' + - 'Qwen/Qwen-VL-Chat:2048' # - 'fnlp/moss-moon-003-sft-4bit:1024' # - 'fnlp/moss-moon-003-sft-4bit:2048' task: 'continuation' # task can be 'continuation', 'QA' and 'summarize' diff --git a/python/llm/test/benchmark/arc-perf-test.yaml b/python/llm/test/benchmark/arc-perf-test.yaml index 09cf7bfd..5298514b 100644 --- a/python/llm/test/benchmark/arc-perf-test.yaml +++ b/python/llm/test/benchmark/arc-perf-test.yaml @@ -8,9 +8,12 @@ repo_id: - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - #- 'Qwen/Qwen-VL-Chat' + - 'Qwen/Qwen-VL-Chat' #- 'SmerkyG/rwkv-5-world-7b' #this model only fp32 is supported for now, fp16 and bf16 are not supported - '01-ai/Yi-6B-Chat' + - 'mistralai/Mistral-7B-Instruct-v0.2' + - 'deepseek-ai/deepseek-coder-7b-instruct-v1.5' + - '01-ai/Yi-1.5-6B-Chat' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml b/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml index 3baea9c8..d675d506 100644 --- a/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml +++ b/python/llm/test/benchmark/arc-perf-transformers-437-batch2.yaml @@ -5,6 +5,7 @@ repo_id: - 'meta-llama/Meta-Llama-3-8B-Instruct' - 'microsoft/phi-3-vision-128k-instruct' - 'Qwen/Qwen2-7B-Instruct' + - 'microsoft/Phi-3-mini-128k-instruct' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml b/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml index 26796f08..f3d55c83 100644 --- a/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml +++ b/python/llm/test/benchmark/arc-perf-transformers-437-batch4.yaml @@ -5,6 +5,7 @@ repo_id: - 'meta-llama/Meta-Llama-3-8B-Instruct' # mwj: need to test - 'microsoft/phi-3-vision-128k-instruct' - 'Qwen/Qwen2-7B-Instruct' + - 'microsoft/Phi-3-mini-128k-instruct' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3 diff --git a/python/llm/test/benchmark/arc-perf-transformers-437.yaml b/python/llm/test/benchmark/arc-perf-transformers-437.yaml index 878b747b..1c775344 100644 --- a/python/llm/test/benchmark/arc-perf-transformers-437.yaml +++ b/python/llm/test/benchmark/arc-perf-transformers-437.yaml @@ -5,6 +5,7 @@ repo_id: - 'meta-llama/Meta-Llama-3-8B-Instruct' - 'microsoft/phi-3-vision-128k-instruct' - 'Qwen/Qwen2-7B-Instruct' + - 'microsoft/Phi-3-mini-128k-instruct' local_model_hub: '/mnt/disk1/models' warm_up: 1 num_trials: 3