Hotfix of benchmark script (#12467)
This commit is contained in:
		
							parent
							
								
									c911026f03
								
							
						
					
					
						commit
						f99f188023
					
				
					 1 changed files with 1 additions and 1 deletions
				
			
		| 
						 | 
					@ -641,7 +641,7 @@ def transformers_int4_npu_win(repo_id,
 | 
				
			||||||
        model = AutoModelForCausalLM.from_pretrained(model_path, load_in_low_bit=low_bit, trust_remote_code=True, torch_dtype=torch.float16,
 | 
					        model = AutoModelForCausalLM.from_pretrained(model_path, load_in_low_bit=low_bit, trust_remote_code=True, torch_dtype=torch.float16,
 | 
				
			||||||
                                                     optimize_model=optimize_model, max_context_len=max_context_len, max_prompt_len=int(in_out_len[0]), 
 | 
					                                                     optimize_model=optimize_model, max_context_len=max_context_len, max_prompt_len=int(in_out_len[0]), 
 | 
				
			||||||
                                                     quantization_group_size=npu_group_size, transpose_value_cache=transpose_value_cache,
 | 
					                                                     quantization_group_size=npu_group_size, transpose_value_cache=transpose_value_cache,
 | 
				
			||||||
                                                     save_directory=save_directory, use_cache=True, attn_implementation="eager").eval()
 | 
					                                                     mixed_precision=True, save_directory=save_directory, use_cache=True, attn_implementation="eager").eval()
 | 
				
			||||||
        tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
 | 
					        tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
 | 
				
			||||||
    end = time.perf_counter()
 | 
					    end = time.perf_counter()
 | 
				
			||||||
    load_time = end - st
 | 
					    load_time = end - st
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue