diff --git a/python/llm/test/llamaindex/test_llamaindex.py b/python/llm/test/llamaindex/test_llamaindex.py index 1e7c8cfe..3864860f 100644 --- a/python/llm/test/llamaindex/test_llamaindex.py +++ b/python/llm/test/llamaindex/test_llamaindex.py @@ -17,7 +17,7 @@ import pytest from unittest import TestCase import os -from ipex_llm.llamaindex.llms import BigdlLLM +from ipex_llm.llamaindex.llms import IpexLLM class Test_LlamaIndex_Transformers_API(TestCase): def setUp(self): @@ -50,7 +50,7 @@ class Test_LlamaIndex_Transformers_API(TestCase): return prompt def test_bigdl_llm(self): - llm = BigdlLLM( + llm = IpexLLM.from_model_id( model_name=self.llama_model_path, tokenizer_name=self.llama_model_path, context_window=512, diff --git a/python/llm/test/llamaindex_gpu/test_llamaindex.py b/python/llm/test/llamaindex_gpu/test_llamaindex.py index b894a37f..37c7f53a 100644 --- a/python/llm/test/llamaindex_gpu/test_llamaindex.py +++ b/python/llm/test/llamaindex_gpu/test_llamaindex.py @@ -18,7 +18,7 @@ import torch import pytest from unittest import TestCase import os -from ipex_llm.llamaindex.llms import BigdlLLM +from ipex_llm.llamaindex.llms import IpexLLM class Test_LlamaIndex_Transformers_API(TestCase): def setUp(self): @@ -51,7 +51,7 @@ class Test_LlamaIndex_Transformers_API(TestCase): return prompt def test_bigdl_llm(self): - llm = BigdlLLM( + llm = IpexLLM.from_model_id( model_name=self.llama_model_path, tokenizer_name=self.llama_model_path, context_window=512,