LLM: avoid unnecessary import torch except converting process (#8297)
This commit is contained in:
parent
f9e2bda04a
commit
08bdfce2d8
1 changed files with 4 additions and 2 deletions
|
|
@ -64,9 +64,7 @@ from pathlib import Path
|
|||
from typing import (IO, TYPE_CHECKING, Any, Callable, Dict, Iterable, List,
|
||||
Literal, Optional, Sequence, Tuple, TypeVar, Union)
|
||||
import numpy as np
|
||||
import torch
|
||||
from sentencepiece import SentencePieceProcessor
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer, AutoConfig
|
||||
from bigdl.llm.utils.common import invalidInputError
|
||||
|
||||
if TYPE_CHECKING:
|
||||
|
|
@ -1238,6 +1236,8 @@ def bytes_to_unicode():
|
|||
|
||||
|
||||
def _convert_gptneox_hf_to_ggml(model_path, outfile_dir, outtype):
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||
import torch
|
||||
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
||||
model = AutoModelForCausalLM.from_pretrained(model_path,
|
||||
torch_dtype=torch.float16
|
||||
|
|
@ -1317,6 +1317,8 @@ def _convert_gptneox_hf_to_ggml(model_path, outfile_dir, outtype):
|
|||
|
||||
|
||||
def _convert_bloom_hf_to_ggml(model_path, outfile_dir, outtype):
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer, AutoConfig
|
||||
import torch
|
||||
conv_map = {'word_embeddings': 'tok_embeddings',
|
||||
'word_embeddings_layernorm': 'norm',
|
||||
'input_layernorm': 'attention_norm',
|
||||
|
|
|
|||
Loading…
Reference in a new issue