diff --git a/python/llm/test/benchmark/arc-perf-test-batch2.yaml b/python/llm/test/benchmark/arc-perf-test-batch2.yaml index e5863465..70447fd7 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch2.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch2.yaml @@ -4,8 +4,6 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' - - 'mistralai/Mistral-7B-v0.1' - - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - 'Qwen/Qwen-VL-Chat' @@ -29,4 +27,4 @@ test_api: cpu_embedding: False # whether put embedding to CPU (only avaiable now for gpu win related test_api) exclude: - 'baichuan-inc/Baichuan2-13B-Chat-4bit:2048' -task: 'continuation' # task can be 'continuation', 'QA' and 'summarize' \ No newline at end of file +task: 'continuation' # task can be 'continuation', 'QA' and 'summarize' diff --git a/python/llm/test/benchmark/arc-perf-test-batch4.yaml b/python/llm/test/benchmark/arc-perf-test-batch4.yaml index 0c83cc9b..3bfd4796 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch4.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch4.yaml @@ -4,8 +4,6 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' - - 'mistralai/Mistral-7B-v0.1' #mwj: need to check - - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - 'Qwen/Qwen-VL-Chat' diff --git a/python/llm/test/benchmark/arc-perf-test.yaml b/python/llm/test/benchmark/arc-perf-test.yaml index 5298514b..890b8dbf 100644 --- a/python/llm/test/benchmark/arc-perf-test.yaml +++ b/python/llm/test/benchmark/arc-perf-test.yaml @@ -4,8 +4,6 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' - - 'mistralai/Mistral-7B-v0.1' - - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - 'Qwen/Qwen-VL-Chat'