From 985213614b1dd2b1fde6a138d4d785b1fc37ecf2 Mon Sep 17 00:00:00 2001 From: SichengStevenLi <144295301+SichengStevenLi@users.noreply.github.com> Date: Tue, 6 Aug 2024 16:12:00 +0800 Subject: [PATCH] Removed no longer needed models for Arc nightly perf (#11722) * removed LLMs that are no longer needed Removed: mistralai/Mistral-7B-v0.1 deepseek-ai/deepseek-coder-6.7b-instruct * Update arc-perf-test-batch4.yaml Removed: deepseek-ai/deepseek-coder-6.7b-instruct mistralai/Mistral-7B-v0.1 * Update arc-perf-test.yaml Removed: deepseek-ai/deepseek-coder-6.7b-instruct mistralai/Mistral-7B-v0.1 * Create arc-perf-transformers-438.yaml * Moved arc-perf-transformers-438.yaml location * Create arc-perf-transformers-438-batch2.yaml * Create arc-perf-transformers-438-batch4.yaml * Delete python/llm/test/benchmark/arc-perf-transformers-438-batch2.yaml * Delete python/llm/test/benchmark/arc-perf-transformers-438-batch4.yaml * Delete python/llm/test/benchmark/arc-perf-transformers-438.yaml --- python/llm/test/benchmark/arc-perf-test-batch2.yaml | 4 +--- python/llm/test/benchmark/arc-perf-test-batch4.yaml | 2 -- python/llm/test/benchmark/arc-perf-test.yaml | 2 -- 3 files changed, 1 insertion(+), 7 deletions(-) diff --git a/python/llm/test/benchmark/arc-perf-test-batch2.yaml b/python/llm/test/benchmark/arc-perf-test-batch2.yaml index e5863465..70447fd7 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch2.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch2.yaml @@ -4,8 +4,6 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' - - 'mistralai/Mistral-7B-v0.1' - - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - 'Qwen/Qwen-VL-Chat' @@ -29,4 +27,4 @@ test_api: cpu_embedding: False # whether put embedding to CPU (only avaiable now for gpu win related test_api) exclude: - 'baichuan-inc/Baichuan2-13B-Chat-4bit:2048' -task: 'continuation' # task can be 'continuation', 'QA' and 'summarize' \ No newline at end of file +task: 'continuation' # task can be 'continuation', 'QA' and 'summarize' diff --git a/python/llm/test/benchmark/arc-perf-test-batch4.yaml b/python/llm/test/benchmark/arc-perf-test-batch4.yaml index 0c83cc9b..3bfd4796 100644 --- a/python/llm/test/benchmark/arc-perf-test-batch4.yaml +++ b/python/llm/test/benchmark/arc-perf-test-batch4.yaml @@ -4,8 +4,6 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' - - 'mistralai/Mistral-7B-v0.1' #mwj: need to check - - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - 'Qwen/Qwen-VL-Chat' diff --git a/python/llm/test/benchmark/arc-perf-test.yaml b/python/llm/test/benchmark/arc-perf-test.yaml index 5298514b..890b8dbf 100644 --- a/python/llm/test/benchmark/arc-perf-test.yaml +++ b/python/llm/test/benchmark/arc-perf-test.yaml @@ -4,8 +4,6 @@ repo_id: - 'THUDM/chatglm3-6b-4bit' - 'baichuan-inc/Baichuan2-7B-Chat' - 'baichuan-inc/Baichuan2-13B-Chat-4bit' - - 'mistralai/Mistral-7B-v0.1' - - 'deepseek-ai/deepseek-coder-6.7b-instruct' - 'THUDM/glm-4-9b-chat' - 'openbmb/MiniCPM-2B-sft-bf16' - 'Qwen/Qwen-VL-Chat'