* add nv longbench * LongBench: NV code to ipex-llm * ammend * add more models support * ammend * optimize LongBench's user experience * ammend * ammend * fix typo * ammend * remove cuda related information & add a readme * add license to python scripts & polish the readme * ammend * ammend --------- Co-authored-by: cyita <yitastudy@gmail.com> Co-authored-by: ATMxsp01 <shou.xu@intel.com> Co-authored-by: leonardozcm <leonardo1997zcm@gmail.com>
		
			
				
	
	
		
			29 lines
		
	
	
	
		
			474 B
		
	
	
	
		
			YAML
		
	
	
	
	
	
			
		
		
	
	
			29 lines
		
	
	
	
		
			474 B
		
	
	
	
		
			YAML
		
	
	
	
	
	
model_name:
 | 
						|
  # - "mistral-7B-instruct-v0.2"
 | 
						|
  - "llama2-7b-chat-4k"
 | 
						|
  # - "chatglm4-9b"
 | 
						|
  # - "qwen2-7b-instruct"
 | 
						|
 | 
						|
full_kv: True
 | 
						|
optimize_model: True
 | 
						|
dtype: 'fp16'
 | 
						|
low_bit: 'sym_int4'
 | 
						|
 | 
						|
e: False
 | 
						|
 | 
						|
compress_kv:
 | 
						|
  - "ablation_c512_w32_k7_maxpool"
 | 
						|
  - "ablation_c1024_w32_k7_maxpool"
 | 
						|
 | 
						|
datasets:
 | 
						|
  - "multi_news"
 | 
						|
  - "qasper"
 | 
						|
  - "hotpotqa"
 | 
						|
  - "trec"
 | 
						|
  - "passage_count"
 | 
						|
  - "lcc"
 | 
						|
  # - "multifieldqa_zh"
 | 
						|
  # - "dureader"
 | 
						|
  # - "vcsum"
 | 
						|
  # - "lsht"
 | 
						|
  # - "passage_retrieval_zh"
 |