From 08018a18dfbe4d2035b38a6e29e2dcc149e2b37b Mon Sep 17 00:00:00 2001 From: Xiangyu Tian <109123695+xiangyuT@users.noreply.github.com> Date: Sun, 7 Apr 2024 10:32:05 +0800 Subject: [PATCH] Remove not-imported MistralConfig (#10670) --- python/llm/src/ipex_llm/vllm/config.py | 9 --------- 1 file changed, 9 deletions(-) diff --git a/python/llm/src/ipex_llm/vllm/config.py b/python/llm/src/ipex_llm/vllm/config.py index a5f4762b..2f1806d2 100644 --- a/python/llm/src/ipex_llm/vllm/config.py +++ b/python/llm/src/ipex_llm/vllm/config.py @@ -318,15 +318,6 @@ def _get_and_verify_max_len( def get_config(model: str, trust_remote_code: bool, revision: Optional[str] = None) -> PretrainedConfig: - # NOTE: Because the Mistral model in HF hub does not have - # `configuration_mistral.py`, we cannot use `AutoConfig` to load the - # config. Instead, we use `MistralConfig` directly. - # NOTE: This is a hack. This does not work for local models. - # FIXME: Remove this once the Mistral model is available in the stable - # version of HF transformers. - if "mistral" in model.lower(): - return MistralConfig.from_pretrained(model, revision=revision) - try: config = AutoConfig.from_pretrained( model, trust_remote_code=trust_remote_code, revision=revision)