Shaojun Liu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								7f8c5b410b 
								
							 
						 
						
							
							
								
								Quickstart: Run PyTorch Inference on Intel GPU using Docker (on Linux or WSL) ( #10970 )  
							
							 
							
							... 
							
							
							
							* add entrypoint.sh
* add quickstart
* remove entrypoint
* update
* Install related library of benchmarking
* update
* print out results
* update docs
* minor update
* update
* update quickstart
* update
* update
* update
* update
* update
* update
* add chat & example section
* add more details
* minor update
* rename quickstart
* update
* minor update
* update
* update config.yaml
* update readme
* use --gpu
* add tips
* minor update
* update 
							
						 
						
							2024-05-14 12:58:31 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									ZehuaCao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								99255fe36e 
								
							 
						 
						
							
							
								
								fix ppl ( #10996 )  
							
							 
							
							
							
						 
						
							2024-05-13 13:57:19 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								dfa3147278 
								
							 
						 
						
							
							
								
								update ( #10944 )  
							
							 
							
							
							
						 
						
							2024-05-08 14:28:05 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								0edef1f94c 
								
							 
						 
						
							
							
								
								LLM: add min_new_tokens to all in one benchmark. ( #10911 )  
							
							 
							
							
							
						 
						
							2024-05-06 09:32:59 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								1a8a93d5e0 
								
							 
						 
						
							
							
								
								Further fix nightly perf ( #10901 )  
							
							 
							
							
							
						 
						
							2024-04-28 10:18:58 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								ddfdaec137 
								
							 
						 
						
							
							
								
								Fix nightly perf ( #10899 )  
							
							 
							
							... 
							
							
							
							* Fix nightly perf by adding default value in benchmark for use_fp16_torch_dtype
* further fixes 
							
						 
						
							2024-04-28 09:39:29 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								f51bf018eb 
								
							 
						 
						
							
							
								
								Add benchmark script for pipeline parallel inference ( #10873 )  
							
							 
							
							
							
						 
						
							2024-04-26 15:28:11 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								cd369c2715 
								
							 
						 
						
							
							
								
								LLM: add device id to benchmark utils. ( #10877 )  
							
							 
							
							
							
						 
						
							2024-04-25 14:01:51 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								eb39c61607 
								
							 
						 
						
							
							
								
								LLM: add min new token to perf test. ( #10869 )  
							
							 
							
							
							
						 
						
							2024-04-24 14:32:02 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c9dee6cd0e 
								
							 
						 
						
							
							
								
								Update 8192.txt ( #10824 )  
							
							 
							
							... 
							
							
							
							* Update 8192.txt
* Update 8192.txt with original text 
							
						 
						
							2024-04-23 14:02:09 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								23c6a52fb0 
								
							 
						 
						
							
							
								
								LLM: Fix ipex torchscript=True error ( #10832 )  
							
							 
							
							... 
							
							
							
							* remove
* update
* remove torchscript 
							
						 
						
							2024-04-22 15:53:09 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Kai Huang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								053ec30737 
								
							 
						 
						
							
							
								
								Transformers ppl evaluation on wikitext ( #10784 )  
							
							 
							
							... 
							
							
							
							* tranformers code
* cache 
							
						 
						
							2024-04-18 15:27:18 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									hxsz1997 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								0d518aab8d 
								
							 
						 
						
							
							
								
								Merge pull request  #10697  from MargarettMao/ceval  
							
							 
							
							... 
							
							
							
							combine english and chinese, remove nan 
							
						 
						
							2024-04-12 14:37:47 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									jenniew 
								
							 
						 
						
							
							
							
							
								
							
							
								cdbb1de972 
								
							 
						 
						
							
							
								
								Mark Color Modification  
							
							 
							
							
							
						 
						
							2024-04-12 14:00:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								2685c41318 
								
							 
						 
						
							
							
								
								Modify all-in-one benchmark ( #10726 )  
							
							 
							
							... 
							
							
							
							* Update 8192 prompt in all-in-one
* Add cpu_embedding param for linux api
* Update run.py
* Update README.md 
							
						 
						
							2024-04-11 13:38:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wenjing Margaret Mao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								289cc99cd6 
								
							 
						 
						
							
							
								
								Update README.md ( #10700 )  
							
							 
							
							... 
							
							
							
							Edit "summarize the results" 
							
						 
						
							2024-04-09 16:01:12 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wenjing Margaret Mao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d3116de0db 
								
							 
						 
						
							
							
								
								Update README.md ( #10701 )  
							
							 
							
							... 
							
							
							
							edit "summarize the results" 
							
						 
						
							2024-04-09 15:50:25 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d59e0cce5c 
								
							 
						 
						
							
							
								
								Migrate harness to ipexllm ( #10703 )  
							
							 
							
							... 
							
							
							
							* migrate to ipexlm
* fix workflow
* fix run_multi
* fix precision map
* rename ipexlm to ipexllm
* rename bigdl to ipex  in comments 
							
						 
						
							2024-04-09 15:48:53 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									jenniew 
								
							 
						 
						
							
							
							
							
								
							
							
								591bae092c 
								
							 
						 
						
							
							
								
								combine english and chinese, remove nan  
							
							 
							
							
							
						 
						
							2024-04-08 19:37:51 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								2d88bb9b4b 
								
							 
						 
						
							
							
								
								add test api transformer_int4_fp16_gpu ( #10627 )  
							
							 
							
							... 
							
							
							
							* add test api transformer_int4_fp16_gpu
* update config.yaml and README.md in all-in-one
* modify run.py in all-in-one
* re-order test-api
* re-order test-api in config
* modify README.md in all-in-one
* modify README.md in all-in-one
* modify config.yaml
---------
Co-authored-by: pengyb2001 <arda@arda-arc21.sh.intel.com>
Co-authored-by: ivy-lv11 <zhicunlv@gmail.com> 
							
						 
						
							2024-04-07 15:47:17 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								9ad4b29697 
								
							 
						 
						
							
							
								
								LLM: CPU benchmark using tcmalloc ( #10675 )  
							
							 
							
							
							
						 
						
							2024-04-07 14:17:01 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d9a1153b4e 
								
							 
						 
						
							
							
								
								LLM: upgrade deepspeed in AutoTP on GPU ( #10647 )  
							
							 
							
							
							
						 
						
							2024-04-07 14:05:19 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								27be448920 
								
							 
						 
						
							
							
								
								LLM: add cpu_embedding and peak memory record for deepspeed autotp script ( #10621 )  
							
							 
							
							
							
						 
						
							2024-04-02 17:32:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								a10f5a1b8d 
								
							 
						 
						
							
							
								
								add python style check ( #10620 )  
							
							 
							
							... 
							
							
							
							* add python style check
* fix style checks
* update runner
* add ipex-llm-finetune-qlora-cpu-k8s to manually_build workflow
* update tag to 2.1.0-SNAPSHOT 
							
						 
						
							2024-04-02 16:17:56 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ruonan Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d6af4877dd 
								
							 
						 
						
							
							
								
								LLM: remove ipex.optimize for gpt-j ( #10606 )  
							
							 
							
							... 
							
							
							
							* remove ipex.optimize
* fix
* fix 
							
						 
						
							2024-04-01 12:21:49 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								fbeb10c796 
								
							 
						 
						
							
							
								
								LLM: Set different env based on different Linux kernels ( #10566 )  
							
							 
							
							
							
						 
						
							2024-03-27 17:56:33 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ruonan Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								ea4bc450c4 
								
							 
						 
						
							
							
								
								LLM: add esimd sdp for pvc ( #10543 )  
							
							 
							
							... 
							
							
							
							* add esimd sdp for pvc
* update
* fix
* fix batch 
							
						 
						
							2024-03-26 19:04:40 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c563b41491 
								
							 
						 
						
							
							
								
								add nightly_build workflow ( #10533 )  
							
							 
							
							... 
							
							
							
							* add nightly_build workflow
* add create-job-status-badge action
* update
* update
* update
* update setup.py
* release
* revert 
							
						 
						
							2024-03-26 12:47:38 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								16b2ef49c6 
								
							 
						 
						
							
							
								
								Update_document by heyang ( #30 )  
							
							 
							
							
							
						 
						
							2024-03-25 10:06:02 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								9df70d95eb 
								
							 
						 
						
							
							
								
								Refactor bigdl.llm to  ipex_llm ( #24 )  
							
							 
							
							... 
							
							
							
							* Rename bigdl/llm to ipex_llm
* rm python/llm/src/bigdl
* from bigdl.llm to from ipex_llm 
							
						 
						
							2024-03-22 15:41:21 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
							
							
								
							
							
								85ef3f1d99 
								
							 
						 
						
							
							
								
								LLM: add empty cache in deepspeed autotp benchmark script ( #10488 )  
							
							 
							
							
							
						 
						
							2024-03-21 10:51:23 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xiangyu Tian 
								
							 
						 
						
							
							
							
							
								
							
							
								5a5fd5af5b 
								
							 
						 
						
							
							
								
								LLM: Add speculative benchmark on CPU/XPU ( #10464 )  
							
							 
							
							... 
							
							
							
							Add speculative benchmark on CPU/XPU. 
							
						 
						
							2024-03-21 09:51:06 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xiangyu Tian 
								
							 
						 
						
							
							
							
							
								
							
							
								cbe24cc7e6 
								
							 
						 
						
							
							
								
								LLM: Enable BigDL IPEX Int8 ( #10480 )  
							
							 
							
							... 
							
							
							
							Enable BigDL IPEX Int8 
							
						 
						
							2024-03-20 15:59:54 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
							
							
								
							
							
								e41d556436 
								
							 
						 
						
							
							
								
								LLM: change fp16 benchmark to model.half ( #10477 )  
							
							 
							
							... 
							
							
							
							* LLM: change fp16 benchmark to model.half
* fix 
							
						 
						
							2024-03-20 13:38:39 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
							
							
								
							
							
								e9055c32f9 
								
							 
						 
						
							
							
								
								LLM: fix fp16 mem record in benchmark ( #10461 )  
							
							 
							
							... 
							
							
							
							* LLM: fix fp16 mem record in benchmark
* change style 
							
						 
						
							2024-03-19 16:17:23 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
							
							
								
							
							
								0451103a43 
								
							 
						 
						
							
							
								
								LLM: add int4+fp16 benchmark script for windows benchmarking ( #10449 )  
							
							 
							
							... 
							
							
							
							* LLM: add fp16 for benchmark script
* remove transformer_int4_fp16_loadlowbit_gpu_win 
							
						 
						
							2024-03-19 11:11:25 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuxuan Xia 
								
							 
						 
						
							
							
							
							
								
							
							
								f36224aac4 
								
							 
						 
						
							
							
								
								Fix ceval run.sh ( #10410 )  
							
							 
							
							
							
						 
						
							2024-03-14 10:57:25 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
							
							
								
							
							
								0193f29411 
								
							 
						 
						
							
							
								
								LLM : Enable  gguf float16 and Yuan2 model ( #10372 )  
							
							 
							
							... 
							
							
							
							* enable float16
* add yun files
* enable yun
* enable set low_bit on yuan2
* update
* update license
* update generate
* update readme
* update python style
* update 
							
						 
						
							2024-03-13 10:19:18 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xiangyu Tian 
								
							 
						 
						
							
							
							
							
								
							
							
								0ded0b4b13 
								
							 
						 
						
							
							
								
								LLM: Enable BigDL IPEX optimization for int4 ( #10319 )  
							
							 
							
							... 
							
							
							
							Enable BigDL IPEX optimization for int4 
							
						 
						
							2024-03-12 17:08:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Lilac09 
								
							 
						 
						
							
							
							
							
								
							
							
								5809a3f5fe 
								
							 
						 
						
							
							
								
								Add run-hbm.sh & add user guide for spr and hbm ( #10357 )  
							
							 
							
							... 
							
							
							
							* add run-hbm.sh
* add spr and hbm guide
* only support quad mode
* only support quad mode
* update special cases
* update special cases 
							
						 
						
							2024-03-12 16:15:27 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
							
							
								
							
							
								5d996a5caf 
								
							 
						 
						
							
							
								
								LLM: add benchmark script for deepspeed autotp on gpu ( #10380 )  
							
							 
							
							
							
						 
						
							2024-03-12 15:19:57 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								17bdb1a60b 
								
							 
						 
						
							
							
								
								LLM: add whisper models into nightly test ( #10193 )  
							
							 
							
							... 
							
							
							
							* LLM: add whisper models into nightly test
* small fix
* small fix
* add more whisper models
* test all cases
* test specific cases
* collect the csv
* store the resut
* to html
* small fix
* small test
* test all cases
* modify whisper_csv_to_html 
							
						 
						
							2024-03-11 20:00:47 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuxuan Xia 
								
							 
						 
						
							
							
							
							
								
							
							
								0c8d3c9830 
								
							 
						 
						
							
							
								
								Add C-Eval HTML report ( #10294 )  
							
							 
							
							... 
							
							
							
							* Add C-Eval HTML report
* Fix C-Eval workflow pr trigger path
* Fix C-Eval workflow typos
* Add permissions to C-Eval workflow
* Fix C-Eval workflow typo
* Add pandas dependency
* Fix C-Eval workflow typo 
							
						 
						
							2024-03-07 16:44:49 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
							
							
								
							
							
								178eea5009 
								
							 
						 
						
							
							
								
								upload bigdl-llm wheel to sourceforge for backup ( #10321 )  
							
							 
							
							... 
							
							
							
							* test: upload to sourceforge
* update scripts
* revert 
							
						 
						
							2024-03-05 16:36:01 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								fd81d66047 
								
							 
						 
						
							
							
								
								LLM: Compress some models to save space ( #10315 )  
							
							 
							
							... 
							
							
							
							* LLM: compress some models to save space
* add deleted comments 
							
						 
						
							2024-03-04 17:53:03 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
							
							
								
							
							
								27d9a14989 
								
							 
						 
						
							
							
								
								[LLM] all-on-one update: memory optimize and streaming output ( #10302 )  
							
							 
							
							... 
							
							
							
							* Memory saving for continous in-out pair run and add support for streaming output on MTL iGPU
* Small fix
* Small fix
* Add things back 
							
						 
						
							2024-03-01 18:02:30 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Keyan (Kyrie) Zhang 
								
							 
						 
						
							
							
							
							
								
							
							
								59861f73e5 
								
							 
						 
						
							
							
								
								Add Deepseek-6.7B ( #9991 )  
							
							 
							
							... 
							
							
							
							* Add new example Deepseek
* Add new example Deepseek
* Add new example Deepseek
* Add new example Deepseek
* Add new example Deepseek
* modify deepseek
* modify deepseek
* Add verified model in README
* Turn cpu_embedding=True in Deepseek example
---------
Co-authored-by: Shengsheng Huang <shengsheng.huang@intel.com> 
							
						 
						
							2024-02-28 11:36:39 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									hxsz1997 
								
							 
						 
						
							
							
							
							
								
							
							
								cba61a2909 
								
							 
						 
						
							
							
								
								Add html report of ppl ( #10218 )  
							
							 
							
							... 
							
							
							
							* remove include and language option, select the corresponding dataset based on the model name in Run
* change the nightly test time
* change the nightly test time of harness and ppl
* save the ppl result to json file
* generate csv file and print table result
* generate html
* modify the way to get parent folder
* update html in parent folder
* add llm-ppl-summary and llm-ppl-summary-html
* modify echo single result
* remove download fp16.csv
* change model name of PR
* move ppl nightly related files to llm/test folder
* reformat
* seperate make_table from make_table_and_csv.py
* separate make_csv from make_table_and_csv.py
* update llm-ppl-html
* remove comment
* add Download fp16.results 
							
						 
						
							2024-02-27 17:37:08 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								213ef06691 
								
							 
						 
						
							
							
								
								fix readme  
							
							 
							
							
							
						 
						
							2024-02-24 00:38:08 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								6fe5344fa6 
								
							 
						 
						
							
							
								
								separate make_csv from the file  
							
							 
							
							
							
						 
						
							2024-02-23 16:33:38 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								bfa98666a6 
								
							 
						 
						
							
							
								
								fall back to make_table.py  
							
							 
							
							
							
						 
						
							2024-02-23 16:33:38 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								f315c7f93a 
								
							 
						 
						
							
							
								
								Move harness nightly related files to llm/test folder ( #10209 )  
							
							 
							
							... 
							
							
							
							* move harness nightly files to test folder
* change workflow file path accordingly
* use arc01 when pr
* fix path
* fix fp16 csv path 
							
						 
						
							2024-02-23 11:12:36 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
							
							
								
							
							
								21de2613ce 
								
							 
						 
						
							
							
								
								[LLM] Add model loading time record for all-in-one benchmark ( #10201 )  
							
							 
							
							... 
							
							
							
							* Add model loading time record in csv for all-in-one benchmark
* Small fix
* Small fix to number after . 
							
						 
						
							2024-02-22 13:57:18 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuxuan Xia 
								
							 
						 
						
							
							
							
							
								
							
							
								7cbc2429a6 
								
							 
						 
						
							
							
								
								Fix C-Eval ChatGLM loading issue ( #10206 )  
							
							 
							
							... 
							
							
							
							* Add c-eval workflow and modify running files
* Modify the chatglm evaluator file
* Modify the ceval workflow for triggering test
* Modify the ceval workflow file
* Modify the ceval workflow file
* Modify ceval workflow
* Adjust the ceval dataset download
* Add ceval workflow dependencies
* Modify ceval workflow dataset download
* Add ceval test dependencies
* Add ceval test dependencies
* Correct the result print
* Fix the nightly test trigger time
* Fix ChatGLM loading issue 
							
						 
						
							2024-02-22 10:00:43 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
							
							
								
							
							
								b1a97b71a9 
								
							 
						 
						
							
							
								
								Harness eval: Add is_last parameter and fix logical operator in highlight_vals ( #10192 )  
							
							 
							
							... 
							
							
							
							* Add is_last parameter and fix logical operator in highlight_vals
* Add script to update HTML files in parent folder
* Add running update_html_in_parent_folder.py in summarize step
* Add licence info
* Remove update_html_in_parent_folder.py in Summarize the results for pull request 
							
						 
						
							2024-02-21 14:45:32 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								39d37bd042 
								
							 
						 
						
							
							
								
								upgrade harness package version in workflow ( #10188 )  
							
							 
							
							... 
							
							
							
							* upgrade harness
* update readme 
							
						 
						
							2024-02-21 11:21:30 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
							
							
								
							
							
								001c13243e 
								
							 
						 
						
							
							
								
								[LLM] Add support for low_low_bit benchmark on Windows GPU ( #10167 )  
							
							 
							
							... 
							
							
							
							* Add support for low_low_bit performance test on Windows GPU
* Small fix
* Small fix
* Save memory during converting model process
* Drop the results for first time when loading in low bit on mtl igpu for better performance
* Small fix 
							
						 
						
							2024-02-21 10:51:52 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
							
							
								
							
							
								de3dc609ee 
								
							 
						 
						
							
							
								
								Modify harness evaluation workflow ( #10174 )  
							
							 
							
							... 
							
							
							
							* Modify table head in harness
* Specify the file path of fp16.csv
* change run to run nightly and run pr to debug
* Modify the way to get fp16.csv to downloading from github
* Change the method to calculate diff in html table
* Change the method to calculate diff in html table
* Re-arrange job order
* Re-arrange job order
* Change limit
* Change fp16.csv  path
* Change highlight rules
* Change limit 
							
						 
						
							2024-02-20 18:55:43 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									hxsz1997 
								
							 
						 
						
							
							
							
							
								
							
							
								6e10d98a8d 
								
							 
						 
						
							
							
								
								Fix some typos ( #10175 )  
							
							 
							
							... 
							
							
							
							* add llm-ppl workflow
* update the DATASET_DIR
* test multiple precisions
* modify nightly test
* match the updated ppl code
* add matrix.include
* fix the include error
* update the include
* add more model
* update the precision of include
* update nightly time and add more models
* fix the workflow_dispatch description, change default model of pr and modify the env
* modify workflow_dispatch language options
* modify options
* modify language options
* modeify workflow_dispatch type
* modify type
* modify the type of language
* change seq_len type
* fix some typos
* revert changes to stress_test.txt 
							
						 
						
							2024-02-20 14:14:53 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
							
							
								
							
							
								e31210ba00 
								
							 
						 
						
							
							
								
								Modify html table style and add fp16.csv in harness ( #10169 )  
							
							 
							
							... 
							
							
							
							* Specify the version of pandas in harness evaluation workflow
* Specify the version of pandas in harness evaluation workflow
* Modify html table style and add fp16.csv in harness
* Modify comments 
							
						 
						
							2024-02-19 18:13:40 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuxuan Xia 
								
							 
						 
						
							
							
							
							
								
							
							
								209122559a 
								
							 
						 
						
							
							
								
								Add Ceval workflow and modify the result printing ( #10140 )  
							
							 
							
							... 
							
							
							
							* Add c-eval workflow and modify running files
* Modify the chatglm evaluator file
* Modify the ceval workflow for triggering test
* Modify the ceval workflow file
* Modify the ceval workflow file
* Modify ceval workflow
* Adjust the ceval dataset download
* Add ceval workflow dependencies
* Modify ceval workflow dataset download
* Add ceval test dependencies
* Add ceval test dependencies
* Correct the result print 
							
						 
						
							2024-02-19 17:06:53 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
							
							
								
							
							
								b4dc33def6 
								
							 
						 
						
							
							
								
								In harness-evaluation workflow, add statistical tables ( #10118 )  
							
							 
							
							... 
							
							
							
							* chnage storage
* fix typo
* change label
* change label to arc03
* change needs in the last step
* add generate csv in harness/make_table_results.py
* modify needs in the last job
* add csv to html
* mfix path issue in llm-harness-summary-nightly
* modify output_path
* modify args in make_table_results.py
* modify make table command in summary
* change pr env label
* remove irrelevant code in summary; add set output path step; add limit in harness run
* re-organize code structure
* modify limit in run harness
* modify csv_to_html input path
* modify needs in summary-nightly 
							
						 
						
							2024-02-08 19:01:05 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuxuan Xia 
								
							 
						 
						
							
							
							
							
								
							
							
								3832eb0ce0 
								
							 
						 
						
							
							
								
								Add ChatGLM C-Eval Evaluator ( #10095 )  
							
							 
							
							... 
							
							
							
							* Add ChatGLM ceval evaluator
* Modify ChatGLM Evaluator Reference 
							
						 
						
							2024-02-07 11:27:06 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ovo233 
								
							 
						 
						
							
							
							
							
								
							
							
								2aaa21c41d 
								
							 
						 
						
							
							
								
								LLM: Update ppl tests ( #10092 )  
							
							 
							
							... 
							
							
							
							* update ppl tests
* use load_dataset api
* add exception handling
* add language argument
* address comments 
							
						 
						
							2024-02-06 17:31:48 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									dingbaorong 
								
							 
						 
						
							
							
							
							
								
							
							
								36c9442c6d 
								
							 
						 
						
							
							
								
								Arc Stable version test ( #10087 )  
							
							 
							
							... 
							
							
							
							* add batch_size in stable version test
* add batch_size in excludes
* add excludes for batch_size
* fix ci
* triger regression test
* fix xpu version
* disable ci
* address kai's comment
---------
Co-authored-by: Ariadne <wyn2000330@126.com> 
							
						 
						
							2024-02-06 10:23:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								c2e562d037 
								
							 
						 
						
							
							
								
								LLM: add batch_size to the csv and html ( #10080 )  
							
							 
							
							... 
							
							
							
							* LLM: add batch_size to the csv and html
* small fix 
							
						 
						
							2024-02-04 16:35:44 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								d2d3f6b091 
								
							 
						 
						
							
							
								
								LLM: ensure the result of daily arc perf test ( #10016 )  
							
							 
							
							... 
							
							
							
							* ensure the result of daily arc perf test
* small fix
* small fix
* small fix
* small fix
* small fix
* small fix
* small fix
* small fix
* small fix
* small fix
* concat more csvs
* small fix
* revert some files 
							
						 
						
							2024-01-31 18:26:21 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ovo233 
								
							 
						 
						
							
							
							
							
								
							
							
								226f398c2a 
								
							 
						 
						
							
							
								
								fix ppl test errors ( #10036 )  
							
							 
							
							
							
						 
						
							2024-01-30 16:26:21 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
							
							
								
							
							
								13e61738c5 
								
							 
						 
						
							
							
								
								hide detail memory for each token in benchmark_utils.py ( #10037 )  
							
							 
							
							
							
						 
						
							2024-01-30 16:04:17 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
							
							
								
							
							
								7952bbc919 
								
							 
						 
						
							
							
								
								add conf batch_size to run_model ( #10010 )  
							
							 
							
							
							
						 
						
							2024-01-26 15:48:48 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								762adc4f9d 
								
							 
						 
						
							
							
								
								Reformat summary table ( #9942 )  
							
							 
							
							... 
							
							
							
							* reformat the table
* refactor the file
* read result.json only 
							
						 
						
							2024-01-25 23:49:00 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ziteng Zhang 
								
							 
						 
						
							
							
							
							
								
							
							
								8b08ad408b 
								
							 
						 
						
							
							
								
								Add batch_size in all_in_one ( #9999 )  
							
							 
							
							... 
							
							
							
							Add batch_size in all_in_one, except run_native_int4 
							
						 
						
							2024-01-25 17:43:49 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								86055d76d5 
								
							 
						 
						
							
							
								
								fix optimize_model not working ( #9995 )  
							
							 
							
							
							
						 
						
							2024-01-25 16:39:05 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								301425e377 
								
							 
						 
						
							
							
								
								harness tests on pvc multiple xpus ( #9908 )  
							
							 
							
							... 
							
							
							
							* add run_multi_llb.py
* update readme
* add job hint 
							
						 
						
							2024-01-23 13:20:37 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cheen Hau, 俊豪 
								
							 
						 
						
							
							
							
							
								
							
							
								947b1e27b7 
								
							 
						 
						
							
							
								
								Add readme for Whisper Test ( #9944 )  
							
							 
							
							... 
							
							
							
							* Fix local data path
* Remove non-essential files
* Add readme
* Minor fixes to script
* Bugfix, refactor
* Add references to original source. Bugfixes.
* Reviewer comments
* Properly print and explain output
* Move files to dev/benchmark
* Fixes 
							
						 
						
							2024-01-22 15:11:33 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
							
							
								
							
							
								6fb3f40f7e 
								
							 
						 
						
							
							
								
								fix error for benchmark_util.py running on cpu ( #9949 )  
							
							 
							
							
							
						 
						
							2024-01-22 10:14:40 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
							
							
								
							
							
								db8e90796a 
								
							 
						 
						
							
							
								
								LLM: add avg token latency information and benchmark guide of autotp ( #9940 )  
							
							 
							
							
							
						 
						
							2024-01-19 15:09:57 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
							
							
								
							
							
								610b5226be 
								
							 
						 
						
							
							
								
								move reserved memory to benchmark_utils.py ( #9907 )  
							
							 
							
							... 
							
							
							
							* move reserved memory to benchmark_utils.py
* meet code review 
							
						 
						
							2024-01-19 09:44:30 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								a8c866c32b 
								
							 
						 
						
							
							
								
								add ppl benchmark ( #9914 )  
							
							 
							
							... 
							
							
							
							* add ppl benchmark
* add license
* add readme
* add dataset argument
* add dataset usage
* fixed low bit args
* correct result
* fix terminal display
* fix ppl update
* enable fp16 fp32 bf16
* format the desc
* fix model_kwargs
* add more readme 
							
						 
						
							2024-01-18 17:54:28 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								100e0a87e5 
								
							 
						 
						
							
							
								
								LLM: add compressed chatglm3 model ( #9892 )  
							
							 
							
							... 
							
							
							
							* LLM: add compressed chatglm3 model
* small fix
* revert github action 
							
						 
						
							2024-01-18 17:48:15 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ruonan Wang 
								
							 
						 
						
							
							
							
							
								
							
							
								b059a32fff 
								
							 
						 
						
							
							
								
								LLM: add benchmark api for bigdl-llm fp16 on GPU ( #9919 )  
							
							 
							
							... 
							
							
							
							* add bmk for bigdl fp16
* fix 
							
						 
						
							2024-01-17 14:24:35 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
							
							
								
							
							
								511cbcf773 
								
							 
						 
						
							
							
								
								LLM: add Ceval benchmark test. ( #9872 )  
							
							 
							
							... 
							
							
							
							* init ceval benchmark test.
* upload dataset.
* add other tests.
* add qwen evaluator.
* fix qwen evaluator style.
* fix qwen evaluator style.
* update qwen evaluator.
* add llama evaluator.
* update eval
* fix typo.
* fix
* fix typo.
* fix llama evaluator.
* fix bug.
* fix style.
* delete dataset.
* fix style.
* fix style.
* add README.md and fix typo.
* fix comments.
* remove run scripts 
							
						 
						
							2024-01-16 19:14:26 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								0e69bfe6b0 
								
							 
						 
						
							
							
								
								LLM: fix the performance drop of starcoder ( #9889 )  
							
							 
							
							... 
							
							
							
							* LLM: fix the performance drop of starcoder
* small fix
* small fix 
							
						 
						
							2024-01-12 09:14:15 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ziteng Zhang 
								
							 
						 
						
							
							
							
							
								
							
							
								4f4ce73f31 
								
							 
						 
						
							
							
								
								[LLM] Add transformer_autocast_bf16 into all-in-one ( #9890 )  
							
							 
							
							... 
							
							
							
							* Add transformer_autocast_bf16 into all-in-one 
							
						 
						
							2024-01-11 17:51:07 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								33fd1f9c76 
								
							 
						 
						
							
							
								
								LLM: fix input length logic for run_transformer_int4_gpu ( #9864 )  
							
							 
							
							... 
							
							
							
							* LLM: fix input length logic for run_transformer_int4_gpu
* small fix
* small fix
* small fix 
							
						 
						
							2024-01-10 18:20:14 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cheen Hau, 俊豪 
								
							 
						 
						
							
							
							
							
								
							
							
								b2aa267f50 
								
							 
						 
						
							
							
								
								Enhance LLM GPU installation document ( #9828 )  
							
							 
							
							... 
							
							
							
							* Improve gpu install doc
* Add troubleshooting - setvars.sh not done properly.
* Further improvements
* 2024.x.x -> 2024.0
* Fixes
* Fix Install BigDL-LLM From Wheel : bigdl-llm[xpu_2.0]
* Remove "export USE_XETLA=OFF" for Max GPU 
							
						 
						
							2024-01-09 16:30:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									dingbaorong 
								
							 
						 
						
							
							
							
							
								
							
							
								f6bb4ab313 
								
							 
						 
						
							
							
								
								Arc stress test ( #9795 )  
							
							 
							
							... 
							
							
							
							* add arc stress test
* triger ci
* triger CI
* triger ci
* disable ci 
							
						 
						
							2023-12-27 21:02:41 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
							
							
								
							
							
								6c75c689ea 
								
							 
						 
						
							
							
								
								bigdl-llm stress test for stable version ( #9781 )  
							
							 
							
							... 
							
							
							
							* 1k-512 2k-512 baseline
* add cpu stress test
* update yaml name
* update
* update
* clean up
* test
* update
* update
* update
* test
* update 
							
						 
						
							2023-12-27 15:40:53 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									dingbaorong 
								
							 
						 
						
							
							
							
							
								
							
							
								5cfb4c4f5b 
								
							 
						 
						
							
							
								
								Arc stable version performance regression test ( #9785 )  
							
							 
							
							... 
							
							
							
							* add arc stable version regression test
* empty gpu mem between different models
* triger ci
* comment spr test
* triger ci
* address kai's comments and disable ci
* merge fp8 and int4
* disable ci 
							
						 
						
							2023-12-27 11:01:56 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								c05d7e1532 
								
							 
						 
						
							
							
								
								LLM: add star_corder_15.5b model ( #9772 )  
							
							 
							
							... 
							
							
							
							* LLM: add star_corder_15.5b model
* revert llm_performance_tests.yml 
							
						 
						
							2023-12-26 18:55:56 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									dingbaorong 
								
							 
						 
						
							
							
							
							
								
							
							
								64d05e581c 
								
							 
						 
						
							
							
								
								add peak gpu mem stats in transformer_int4_gpu ( #9766 )  
							
							 
							
							... 
							
							
							
							* add peak gpu mem stats in transformer_int4_gpu
* address weiguang's comments 
							
						 
						
							2023-12-26 15:38:28 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								7fd7c37e1b 
								
							 
						 
						
							
							
								
								Enable fp8e5 harness ( #9761 )  
							
							 
							
							... 
							
							
							
							* fix precision format like fp8e5
* match fp8_e5m2 
							
						 
						
							2023-12-22 16:59:48 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								474c099559 
								
							 
						 
						
							
							
								
								LLM: using separate threads to do inference ( #9727 )  
							
							 
							
							... 
							
							
							
							* using separate threads to do inference
* resolve some comments
* resolve some comments
* revert llm_performance_tests.yml file 
							
						 
						
							2023-12-21 17:56:43 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								b06a3146c8 
								
							 
						 
						
							
							
								
								Fix 70b oom ( #9738 )  
							
							 
							
							... 
							
							
							
							* add default value to bigdl llm
* fix model oom 
							
						 
						
							2023-12-21 10:40:52 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								3e8d198b57 
								
							 
						 
						
							
							
								
								LLM: add eval func ( #9662 )  
							
							 
							
							... 
							
							
							
							* Add eval func
* add left eval 
							
						 
						
							2023-12-14 14:59:02 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
							
							
								
							
							
								cbdd49f229 
								
							 
						 
						
							
							
								
								[LLM] win igpu performance for ipex 2.1 and oneapi 2024.0 ( #9679 )  
							
							 
							
							... 
							
							
							
							* Change igpu win tests for ipex 2.1 and oneapi 2024.0
* Qwen model repo id updates; updates model list for 512-64
* Add .eval for win igpu all-in-one benchmark for best performance 
							
						 
						
							2023-12-13 18:52:29 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Mingyu Wei 
								
							 
						 
						
							
							
							
							
								
							
							
								16febc949c 
								
							 
						 
						
							
							
								
								[LLM] Add exclude option in all-in-one performance test ( #9632 )  
							
							 
							
							... 
							
							
							
							* add exclude option in all-in-one perf test
* update arc-perf-test.yaml
* Exclude in_out_pairs in main function
* fix some bugs
* address Kai's comments
* define excludes at the beginning
* add bloomz:2048 to exclude 
							
						 
						
							2023-12-13 18:13:06 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
							
							
								
							
							
								968d99e6f5 
								
							 
						 
						
							
							
								
								Remove empty cache between each iteration of generation ( #9660 )  
							
							 
							
							
							
						 
						
							2023-12-12 17:24:06 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
							
							
								
							
							
								972cdb9992 
								
							 
						 
						
							
							
								
								gsm8k OOM workaround ( #9597 )  
							
							 
							
							... 
							
							
							
							* update bigdl_llm.py
* update the installation of harness
* fix partial function
* import ipex
* force seq len in decrease order
* put func outside class
* move comments
* default 'trust_remote_code' as True
* Update llm-harness-evaluation.yml 
							
						 
						
							2023-12-08 18:47:25 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
							
							
								
							
							
								e9299adb3b 
								
							 
						 
						
							
							
								
								LLM: Highlight some values in the html ( #9635 )  
							
							 
							
							... 
							
							
							
							* highlight some values in the html
* revert the llm_performance_tests.yml 
							
						 
						
							2023-12-07 19:02:41 +08:00