From 1571ba642595838e70c0911e2e6d1abc50b10072 Mon Sep 17 00:00:00 2001 From: Yina Chen <33650826+cyita@users.noreply.github.com> Date: Thu, 8 Jun 2023 11:04:47 +0800 Subject: [PATCH] remove unused import gptneox_cpp (#8293) --- python/llm/src/bigdl/llm/ggml/model/generation/utils.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/python/llm/src/bigdl/llm/ggml/model/generation/utils.py b/python/llm/src/bigdl/llm/ggml/model/generation/utils.py index 5e8da4f0..18ffa271 100644 --- a/python/llm/src/bigdl/llm/ggml/model/generation/utils.py +++ b/python/llm/src/bigdl/llm/ggml/model/generation/utils.py @@ -22,7 +22,6 @@ from typing import Optional, Union, Sequence, List from bigdl.llm.utils.common import invalidInputError -from bigdl.llm.ggml.model.gptneox import gptneox_cpp class GenerationMixin: @@ -57,7 +56,7 @@ class GenerationMixin: def generate( self, - inputs: Union[Optional[Sequence[int]], Sequence[gptneox_cpp.gptneox_token]]=None, + inputs: Optional[Sequence[int]]=None, max_new_tokens: int = 128, top_k: int = 40, top_p: float = 0.95, @@ -72,7 +71,7 @@ class GenerationMixin: mirostat_eta: float = 0.1, stop: Optional[Union[str, List[str]]]=[], # TODO: rebase to support stopping_criteria **kwargs, - ) -> Union[Optional[Sequence[int]], Optional[Sequence[gptneox_cpp.gptneox_token]], None]: + ) -> Union[Optional[Sequence[int]], None]: # TODO: modify docs """Create a generator of tokens from a prompt.