parent
d4027d7164
commit
c256cd136b
1 changed files with 7 additions and 7 deletions
|
|
@ -41,13 +41,12 @@ def llm_convert(model,
|
||||||
check, _used_args = _special_kwarg_check(kwargs=kwargs,
|
check, _used_args = _special_kwarg_check(kwargs=kwargs,
|
||||||
check_args=["tmp_path"])
|
check_args=["tmp_path"])
|
||||||
invalidInputError(check, f"Invaid input kwargs found: {_used_args}")
|
invalidInputError(check, f"Invaid input kwargs found: {_used_args}")
|
||||||
ggml_convert_model(input_path=model,
|
return ggml_convert_model(input_path=model,
|
||||||
output_path=outfile,
|
output_path=outfile,
|
||||||
model_family=model_family,
|
model_family=model_family,
|
||||||
dtype=outtype,
|
dtype=outtype,
|
||||||
**_used_args,
|
**_used_args,
|
||||||
)
|
)
|
||||||
return outfile
|
|
||||||
elif model_format == "gptq":
|
elif model_format == "gptq":
|
||||||
invalidInputError(model.endswith(".pt"), "only support pytorch's .pt format now.")
|
invalidInputError(model.endswith(".pt"), "only support pytorch's .pt format now.")
|
||||||
invalidInputError(model_family == "llama" and outtype == 'int4',
|
invalidInputError(model_family == "llama" and outtype == 'int4',
|
||||||
|
|
@ -68,6 +67,7 @@ def llm_convert(model,
|
||||||
else:
|
else:
|
||||||
invalidInputError(False, f"Unsupported input model_type: {model_format}")
|
invalidInputError(False, f"Unsupported input model_type: {model_format}")
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
parser = argparse.ArgumentParser(description='Model Convert Parameters')
|
parser = argparse.ArgumentParser(description='Model Convert Parameters')
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue