update batch kernel condition (#12421)
This commit is contained in:
		
							parent
							
								
									7288c759ce
								
							
						
					
					
						commit
						145e8b480f
					
				
					 1 changed files with 10 additions and 2 deletions
				
			
		| 
						 | 
				
			
			@ -362,14 +362,22 @@ def use_batch_forward(x: torch.Tensor, qtype: int, output_len: int):
 | 
			
		|||
    batch_size = x.shape[0]
 | 
			
		||||
    hard_condition = (
 | 
			
		||||
        x.dtype in [torch.float, torch.half]
 | 
			
		||||
        and batch_size <= 48
 | 
			
		||||
        and x.shape[1] % 128 == 0
 | 
			
		||||
        and (
 | 
			
		||||
            (
 | 
			
		||||
                qtype in [SYM_INT4, ASYM_INT4, FP8E5, FP8E4]
 | 
			
		||||
                and x.shape[1] % 128 == 0
 | 
			
		||||
                and (
 | 
			
		||||
                    batch_size <= 48
 | 
			
		||||
                    or (
 | 
			
		||||
                        batch_size <= 64
 | 
			
		||||
                        and x.shape[1] % 256 == 0
 | 
			
		||||
                        and output_len % 64 == 0
 | 
			
		||||
                    )
 | 
			
		||||
                )
 | 
			
		||||
            )
 | 
			
		||||
            or (
 | 
			
		||||
                qtype in [SYM_INT8, FP4, FP6, Q4_K, Q6_K]
 | 
			
		||||
                and batch_size <= 48
 | 
			
		||||
                and device in ["arc", "flex", "pvc", "mtl"]
 | 
			
		||||
                and x.shape[1] % 256 == 0
 | 
			
		||||
                and output_len % 32 == 0
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue