update example to reduce peak memory usage (#12274)
This commit is contained in:
parent
e713296090
commit
854398f6e0
1 changed files with 1 additions and 0 deletions
|
|
@ -60,6 +60,7 @@ if __name__ == "__main__":
|
|||
optimize_model=True,
|
||||
pipeline=True,
|
||||
max_output_len=args.max_output_len,
|
||||
torch_dtype=torch.float16,
|
||||
attn_implementation="eager")
|
||||
|
||||
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
|
||||
|
|
|
|||
Loading…
Reference in a new issue