Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c9e6d42ad1 
								
							 
						 
						
							
							
								
								LLM: Fix chatglm3-6b-32k error ( #10719 )  
							
							 
							
							... 
							
							
							
							* fix chatglm3-6b-32k
* update style 
							
						 
						
							2024-04-10 11:24:06 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Keyan (Kyrie) Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								585c174e92 
								
							 
						 
						
							
							
								
								Read the value of KV_CACHE_ALLOC_BLOCK_LENGTH from the environment variables ( #10707 )  
							
							 
							
							... 
							
							
							
							* Read the value of KV_CACHE_ALLOC_BLOCK_LENGTH from the environment variables.
* Fix style 
							
						 
						
							2024-04-10 10:48:46 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d1eaea509f 
								
							 
						 
						
							
							
								
								update chatglm readme ( #10659 )  
							
							 
							
							
							
						 
						
							2024-04-09 14:24:46 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								878a97077b 
								
							 
						 
						
							
							
								
								Fix llava example to support transformerds 4.36 ( #10614 )  
							
							 
							
							... 
							
							
							
							* fix llava example
* update 
							
						 
						
							2024-04-09 13:47:07 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								1e817926ba 
								
							 
						 
						
							
							
								
								Fix low memory generation example issue in transformers 4.36 ( #10702 )  
							
							 
							
							... 
							
							
							
							* update cache in low memory generate
* update 
							
						 
						
							2024-04-09 09:56:52 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								97db2492c8 
								
							 
						 
						
							
							
								
								Update setup.py for bigdl-core-xe-esimd-21 on Windows ( #10705 )  
							
							 
							
							... 
							
							
							
							* Support bigdl-core-xe-esimd-21 for windows in setup.py
* Update setup-llm-env accordingly 
							
						 
						
							2024-04-09 18:21:21 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Zhicun 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								b4147a97bb 
								
							 
						 
						
							
							
								
								Fix dtype mismatch error ( #10609 )  
							
							 
							
							... 
							
							
							
							* fix llama
* fix
* fix code style
* add torch type in model.py
---------
Co-authored-by: arda <arda@arda-arc19.sh.intel.com> 
							
						 
						
							2024-04-09 17:50:33 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								f37a1f2a81 
								
							 
						 
						
							
							
								
								Upgrade to python 3.11 ( #10711 )  
							
							 
							
							... 
							
							
							
							* create conda env with python 3.11
* recommend to use Python 3.11
* update 
							
						 
						
							2024-04-09 17:41:17 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								8f45e22072 
								
							 
						 
						
							
							
								
								fix llama2 ( #10710 )  
							
							 
							
							
							
						 
						
							2024-04-09 17:28:37 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								e438f941f2 
								
							 
						 
						
							
							
								
								disable rwkv5 fp16 ( #10699 )  
							
							 
							
							
							
						 
						
							2024-04-09 16:42:11 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								6a32216269 
								
							 
						 
						
							
							
								
								LLM: add llama2 8k input example. ( #10696 )  
							
							 
							
							... 
							
							
							
							* LLM: add llama2-32K example.
* refactor name.
* fix comments.
* add IPEX_LLM_LOW_MEM notes and update sample output. 
							
						 
						
							2024-04-09 16:02:37 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wenjing Margaret Mao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								289cc99cd6 
								
							 
						 
						
							
							
								
								Update README.md ( #10700 )  
							
							 
							
							... 
							
							
							
							Edit "summarize the results" 
							
						 
						
							2024-04-09 16:01:12 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wenjing Margaret Mao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d3116de0db 
								
							 
						 
						
							
							
								
								Update README.md ( #10701 )  
							
							 
							
							... 
							
							
							
							edit "summarize the results" 
							
						 
						
							2024-04-09 15:50:25 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Chen, Zhentao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d59e0cce5c 
								
							 
						 
						
							
							
								
								Migrate harness to ipexllm ( #10703 )  
							
							 
							
							... 
							
							
							
							* migrate to ipexlm
* fix workflow
* fix run_multi
* fix precision map
* rename ipexlm to ipexllm
* rename bigdl to ipex  in comments 
							
						 
						
							2024-04-09 15:48:53 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Keyan (Kyrie) Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								1e27e08322 
								
							 
						 
						
							
							
								
								Modify example from fp32 to fp16 ( #10528 )  
							
							 
							
							... 
							
							
							
							* Modify example from fp32 to fp16
* Remove Falcon from fp16 example for now
* Remove MPT from fp16 example 
							
						 
						
							2024-04-09 15:45:49 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								44922bb5c2 
								
							 
						 
						
							
							
								
								LLM: support baichuan2-13b using AutoTP ( #10691 )  
							
							 
							
							
							
						 
						
							2024-04-09 14:06:01 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yina Chen 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c7422712fc 
								
							 
						 
						
							
							
								
								mistral 4.36 use fp16 sdp ( #10704 )  
							
							 
							
							
							
						 
						
							2024-04-09 13:50:33 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ovo233 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								dcb2038aad 
								
							 
						 
						
							
							
								
								Enable optimization for sentence_transformers ( #10679 )  
							
							 
							
							... 
							
							
							
							* enable optimization for sentence_transformers
* fix python style check failure 
							
						 
						
							2024-04-09 12:33:46 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yang Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								5a1f446d3c 
								
							 
						 
						
							
							
								
								support fp8 in xetla ( #10555 )  
							
							 
							
							... 
							
							
							
							* support fp8 in xetla
* change name
* adjust model file
* support convert back to cpu
* factor
* fix bug
* fix style 
							
						 
						
							2024-04-08 13:22:09 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									jenniew 
								
							 
						 
						
							
							
							
							
								
							
							
								591bae092c 
								
							 
						 
						
							
							
								
								combine english and chinese, remove nan  
							
							 
							
							
							
						 
						
							2024-04-08 19:37:51 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								7c43ac0164 
								
							 
						 
						
							
							
								
								LLM: optimize llama natvie sdp for split qkv tensor ( #10693 )  
							
							 
							
							... 
							
							
							
							* LLM: optimize llama natvie sdp for split qkv tensor.
* fix block real size.
* fix comment.
* fix style.
* refactor. 
							
						 
						
							2024-04-08 17:48:11 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								1274cba79b 
								
							 
						 
						
							
							
								
								stablelm fp8 kv cache ( #10672 )  
							
							 
							
							... 
							
							
							
							* stablelm fp8 kvcache
* update
* fix
* change to fp8 matmul
* fix style
* fix
* fix
* meet code review
* add comment 
							
						 
						
							2024-04-08 15:16:46 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								65127622aa 
								
							 
						 
						
							
							
								
								fix UT threshold ( #10689 )  
							
							 
							
							
							
						 
						
							2024-04-08 14:58:20 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c0cd238e40 
								
							 
						 
						
							
							
								
								LLM: support llama2 8k input with w4a16. ( #10677 )  
							
							 
							
							... 
							
							
							
							* LLM: support llama2 8k input with w4a16.
* fix comment and style.
* fix style.
* fix comments and split tensor to quantized attention forward.
* fix style.
* refactor name.
* fix style.
* fix style.
* fix style.
* refactor checker name.
* refactor native sdp split qkv tensor name.
* fix style.
* fix comment rename variables.
* fix co-exist of intermedia results. 
							
						 
						
							2024-04-08 11:43:15 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Zhicun 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								321bc69307 
								
							 
						 
						
							
							
								
								Fix llamaindex ut ( #10673 )  
							
							 
							
							... 
							
							
							
							* fix llamaindex ut
* add GPU ut 
							
						 
						
							2024-04-08 09:47:51 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									yb-peng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								2d88bb9b4b 
								
							 
						 
						
							
							
								
								add test api transformer_int4_fp16_gpu ( #10627 )  
							
							 
							
							... 
							
							
							
							* add test api transformer_int4_fp16_gpu
* update config.yaml and README.md in all-in-one
* modify run.py in all-in-one
* re-order test-api
* re-order test-api in config
* modify README.md in all-in-one
* modify README.md in all-in-one
* modify config.yaml
---------
Co-authored-by: pengyb2001 <arda@arda-arc21.sh.intel.com>
Co-authored-by: ivy-lv11 <zhicunlv@gmail.com> 
							
						 
						
							2024-04-07 15:47:17 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								47cabe8fcc 
								
							 
						 
						
							
							
								
								LLM: Fix no return_last_logit running bigdl_ipex chatglm3 ( #10678 )  
							
							 
							
							... 
							
							
							
							* fix no return_last_logits
* update only for chatglm 
							
						 
						
							2024-04-07 15:27:58 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								9ad4b29697 
								
							 
						 
						
							
							
								
								LLM: CPU benchmark using tcmalloc ( #10675 )  
							
							 
							
							
							
						 
						
							2024-04-07 14:17:01 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d9a1153b4e 
								
							 
						 
						
							
							
								
								LLM: upgrade deepspeed in AutoTP on GPU ( #10647 )  
							
							 
							
							
							
						 
						
							2024-04-07 14:05:19 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								56dfcb2ade 
								
							 
						 
						
							
							
								
								Migrate portable zip to ipex-llm ( #10617 )  
							
							 
							
							... 
							
							
							
							* change portable zip prompt to ipex-llm
* fix chat with ui
* add no proxy 
							
						 
						
							2024-04-07 13:58:58 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Zhicun 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								9d8ba64c0d 
								
							 
						 
						
							
							
								
								Llamaindex: add tokenizer_id and support chat ( #10590 )  
							
							 
							
							... 
							
							
							
							* add tokenizer_id
* fix
* modify
* add from_model_id and from_mode_id_low_bit
* fix typo and add comment
* fix python code style
---------
Co-authored-by: pengyb2001 <284261055@qq.com> 
							
						 
						
							2024-04-07 13:51:34 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								10ee786920 
								
							 
						 
						
							
							
								
								Replace with IPEX-LLM in example comments ( #10671 )  
							
							 
							
							... 
							
							
							
							* Replace with IPEX-LLM in example comments
* More replacement
* revert some changes 
							
						 
						
							2024-04-07 13:29:51 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xiangyu Tian 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								08018a18df 
								
							 
						 
						
							
							
								
								Remove not-imported MistralConfig ( #10670 )  
							
							 
							
							
							
						 
						
							2024-04-07 10:32:05 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								1a9b8204a4 
								
							 
						 
						
							
							
								
								LLM: support int4 fp16 chatglm2-6b 8k input. ( #10648 )  
							
							 
							
							
							
						 
						
							2024-04-07 09:39:21 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								69bdbf5806 
								
							 
						 
						
							
							
								
								Fix vllm print error message issue ( #10664 )  
							
							 
							
							... 
							
							
							
							* update chatglm readme
* Add condition to invalidInputError
* update
* update
* style 
							
						 
						
							2024-04-05 15:08:13 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jason Dai 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								29d97e4678 
								
							 
						 
						
							
							
								
								Update readme ( #10665 )  
							
							 
							
							
							
						 
						
							2024-04-05 18:01:57 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								4c3e493b2d 
								
							 
						 
						
							
							
								
								fix stablelm2 1.6b ( #10656 )  
							
							 
							
							... 
							
							
							
							* fix stablelm2 1.6b
* meet code review 
							
						 
						
							2024-04-03 22:15:32 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								cc8b3be11c 
								
							 
						 
						
							
							
								
								Add GPU and CPU example for stablelm-zephyr-3b ( #10643 )  
							
							 
							
							... 
							
							
							
							* Add example for StableLM
* fix
* add to readme 
							
						 
						
							2024-04-03 16:28:31 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Heyang Sun 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								6000241b10 
								
							 
						 
						
							
							
								
								Add Deepspeed Example of FLEX Mistral ( #10640 )  
							
							 
							
							
							
						 
						
							2024-04-03 16:04:17 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d18dbfb097 
								
							 
						 
						
							
							
								
								update spr perf test ( #10644 )  
							
							 
							
							
							
						 
						
							2024-04-03 15:53:55 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								702e686901 
								
							 
						 
						
							
							
								
								optimize starcoder normal kv cache ( #10642 )  
							
							 
							
							
							
						 
						
							2024-04-03 15:27:02 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								3a9ab8f1ae 
								
							 
						 
						
							
							
								
								fix stablelm logits diff ( #10636 )  
							
							 
							
							... 
							
							
							
							* fix logits diff
* Small fixes
---------
Co-authored-by: Yuwen Hu <yuwen.hu@intel.com> 
							
						 
						
							2024-04-03 15:08:12 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Zhicun 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								b827f534d5 
								
							 
						 
						
							
							
								
								Add tokenizer_id in Langchain ( #10588 )  
							
							 
							
							... 
							
							
							
							* fix low-bit
* fix
* fix style
---------
Co-authored-by: arda <arda@arda-arc12.sh.intel.com> 
							
						 
						
							2024-04-03 14:25:35 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Zhicun 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								f6fef09933 
								
							 
						 
						
							
							
								
								fix prompt format for llama-2 in langchain ( #10637 )  
							
							 
							
							
							
						 
						
							2024-04-03 14:17:34 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								330d4b4f4b 
								
							 
						 
						
							
							
								
								update readme ( #10631 )  
							
							 
							
							
							
						 
						
							2024-04-02 23:08:02 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Kai Huang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c875b3c858 
								
							 
						 
						
							
							
								
								Add seq len check for llama softmax upcast to fp32 ( #10629 )  
							
							 
							
							
							
						 
						
							2024-04-03 12:05:13 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								4431134ec5 
								
							 
						 
						
							
							
								
								update readme ( #10632 )  
							
							 
							
							
							
						 
						
							2024-04-02 19:54:30 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								23e33a0ca1 
								
							 
						 
						
							
							
								
								Fix qwen-vl style ( #10633 )  
							
							 
							
							... 
							
							
							
							* update
* update 
							
						 
						
							2024-04-02 18:41:38 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								2bbd8a1548 
								
							 
						 
						
							
							
								
								LLM: fix llama2 FP16 & bs>1 & autotp on PVC and ARC ( #10611 )  
							
							 
							
							
							
						 
						
							2024-04-03 09:28:04 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jiao Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								654dc5ba57 
								
							 
						 
						
							
							
								
								Fix Qwen-VL example problem ( #10582 )  
							
							 
							
							... 
							
							
							
							* update
* update
* update
* update 
							
						 
						
							2024-04-02 12:17:30 -07:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								fd384ddfb8 
								
							 
						 
						
							
							
								
								Optimize StableLM ( #10619 )  
							
							 
							
							... 
							
							
							
							* Initial commit for stablelm optimizations
* Small style fix
* add dependency
* Add mlp optimizations
* Small fix
* add attention forward
* Remove quantize kv for now as head_dim=80
* Add merged qkv
* fix lisence
* Python style fix
---------
Co-authored-by: qiuxin2012 <qiuxin2012cs@gmail.com> 
							
						 
						
							2024-04-02 18:58:38 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								27be448920 
								
							 
						 
						
							
							
								
								LLM: add cpu_embedding and peak memory record for deepspeed autotp script ( #10621 )  
							
							 
							
							
							
						 
						
							2024-04-02 17:32:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								ba8cc6bd68 
								
							 
						 
						
							
							
								
								optimize starcoder2-3b ( #10625 )  
							
							 
							
							
							
						 
						
							2024-04-02 17:16:29 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								a10f5a1b8d 
								
							 
						 
						
							
							
								
								add python style check ( #10620 )  
							
							 
							
							... 
							
							
							
							* add python style check
* fix style checks
* update runner
* add ipex-llm-finetune-qlora-cpu-k8s to manually_build workflow
* update tag to 2.1.0-SNAPSHOT 
							
						 
						
							2024-04-02 16:17:56 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								58b57177e3 
								
							 
						 
						
							
							
								
								LLM: support bigdl quantize kv cache env and add warning. ( #10623 )  
							
							 
							
							... 
							
							
							
							* LLM: support bigdl quantize kv cache env and add warnning.
* fix style.
* fix comments. 
							
						 
						
							2024-04-02 15:41:08 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Kai Huang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								0a95c556a1 
								
							 
						 
						
							
							
								
								Fix starcoder first token perf ( #10612 )  
							
							 
							
							... 
							
							
							
							* add bias check
* update 
							
						 
						
							2024-04-02 09:21:38 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								e567956121 
								
							 
						 
						
							
							
								
								LLM: add memory optimization for llama. ( #10592 )  
							
							 
							
							... 
							
							
							
							* add initial memory optimization.
* fix logic.
* fix logic,
* remove env var check in mlp split. 
							
						 
						
							2024-04-02 09:07:50 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Keyan (Kyrie) Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								01f491757a 
								
							 
						 
						
							
							
								
								Modify the link in Langchain-upstream ut ( #10608 )  
							
							 
							
							... 
							
							
							
							* Modify the link in Langchain-upstream ut
* fix langchain-upstream ut 
							
						 
						
							2024-04-01 17:03:40 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ruonan Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								bfc1caa5e5 
								
							 
						 
						
							
							
								
								LLM: support iq1s for llama2-70b-hf ( #10596 )  
							
							 
							
							
							
						 
						
							2024-04-01 13:13:13 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ruonan Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d6af4877dd 
								
							 
						 
						
							
							
								
								LLM: remove ipex.optimize for gpt-j ( #10606 )  
							
							 
							
							... 
							
							
							
							* remove ipex.optimize
* fix
* fix 
							
						 
						
							2024-04-01 12:21:49 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								437a349dd6 
								
							 
						 
						
							
							
								
								fix rwkv with pip installer ( #10591 )  
							
							 
							
							
							
						 
						
							2024-03-29 17:56:45 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								9a83f21b86 
								
							 
						 
						
							
							
								
								LLM: check user env ( #10580 )  
							
							 
							
							... 
							
							
							
							* LLM: check user env
* small fix
* small fix
* small fix 
							
						 
						
							2024-03-29 17:19:34 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Keyan (Kyrie) Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								848fa04dd6 
								
							 
						 
						
							
							
								
								Fix typo in Baichuan2 example ( #10589 )  
							
							 
							
							
							
						 
						
							2024-03-29 13:31:47 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ruonan Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								0136fad1d4 
								
							 
						 
						
							
							
								
								LLM: support iq1_s ( #10564 )  
							
							 
							
							... 
							
							
							
							* init version
* update utils
* remove unsed code 
							
						 
						
							2024-03-29 09:43:55 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Qiyuan Gong 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								f4537798c1 
								
							 
						 
						
							
							
								
								Enable kv cache quantization by default for flex when 1 < batch <= 8 ( #10584 )  
							
							 
							
							... 
							
							
							
							* Enable kv cache quantization by default for flex when 1 < batch <= 8.
* Change up bound from <8 to <=8. 
							
						 
						
							2024-03-29 09:43:42 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								b44f7adbad 
								
							 
						 
						
							
							
								
								LLM: Disable esimd sdp for PVC GPU when batch size>1 ( #10579 )  
							
							 
							
							... 
							
							
							
							* llm: disable esimd sdp for pvc bz>1.
* fix logic.
* fix: avoid call get device name twice. 
							
						 
						
							2024-03-28 22:55:48 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								5963239b46 
								
							 
						 
						
							
							
								
								Fix qwen's position_ids no enough ( #10572 )  
							
							 
							
							... 
							
							
							
							* fix position_ids
* fix position_ids 
							
						 
						
							2024-03-28 17:05:49 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									ZehuaCao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								52a2135d83 
								
							 
						 
						
							
							
								
								Replace ipex with ipex-llm ( #10554 )  
							
							 
							
							... 
							
							
							
							* fix ipex with ipex_llm
* fix ipex with ipex_llm
* update
* update
* update
* update
* update
* update
* update
* update 
							
						 
						
							2024-03-28 13:54:40 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cheen Hau, 俊豪 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								1c5eb14128 
								
							 
						 
						
							
							
								
								Update pip install to use --extra-index-url for ipex package ( #10557 )  
							
							 
							
							... 
							
							
							
							* Change to 'pip install .. --extra-index-url' for readthedocs
* Change to 'pip install .. --extra-index-url' for examples
* Change to 'pip install .. --extra-index-url' for remaining files
* Fix URL for ipex
* Add links for ipex US and CN servers
* Update ipex cpu url
* remove readme
* Update for github actions
* Update for dockerfiles 
							
						 
						
							2024-03-28 09:56:23 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								92dfed77be 
								
							 
						 
						
							
							
								
								LLM: fix abnormal output of fp16 deepspeed autotp ( #10558 )  
							
							 
							
							
							
						 
						
							2024-03-28 09:35:48 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jason Dai 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c450c85489 
								
							 
						 
						
							
							
								
								Delete llm/readme.md ( #10569 )  
							
							 
							
							
							
						 
						
							2024-03-27 20:06:40 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xiangyu Tian 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								51d34ca68e 
								
							 
						 
						
							
							
								
								Fix wrong import in speculative ( #10562 )  
							
							 
							
							
							
						 
						
							2024-03-27 18:21:07 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cheen Hau, 俊豪 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								f239bc329b 
								
							 
						 
						
							
							
								
								Specify oneAPI minor version in documentation ( #10561 )  
							
							 
							
							
							
						 
						
							2024-03-27 17:58:57 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									WeiguangHan 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								fbeb10c796 
								
							 
						 
						
							
							
								
								LLM: Set different env based on different Linux kernels ( #10566 )  
							
							 
							
							
							
						 
						
							2024-03-27 17:56:33 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									hxsz1997 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								d86477f14d 
								
							 
						 
						
							
							
								
								Remove native_int4 in LangChain examples ( #10510 )  
							
							 
							
							... 
							
							
							
							* rebase the modify to ipex-llm
* modify the typo 
							
						 
						
							2024-03-27 17:48:16 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Guancheng Fu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								04baac5a2e 
								
							 
						 
						
							
							
								
								Fix fastchat top_k ( #10560 )  
							
							 
							
							... 
							
							
							
							* fix -1 top_k
* fix
* done 
							
						 
						
							2024-03-27 16:01:58 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								fc8c7904f0 
								
							 
						 
						
							
							
								
								LLM: fix torch_dtype setting of apply fp16 optimization through optimize_model ( #10556 )  
							
							 
							
							
							
						 
						
							2024-03-27 14:18:45 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Ruonan Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								ea4bc450c4 
								
							 
						 
						
							
							
								
								LLM: add esimd sdp for pvc ( #10543 )  
							
							 
							
							... 
							
							
							
							* add esimd sdp for pvc
* update
* fix
* fix batch 
							
						 
						
							2024-03-26 19:04:40 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								b78289a595 
								
							 
						 
						
							
							
								
								Remove ipex-llm dependency in readme ( #10544 )  
							
							 
							
							
							
						 
						
							2024-03-26 18:25:14 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xiangyu Tian 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								11550d3f25 
								
							 
						 
						
							
							
								
								LLM: Add length check for IPEX-CPU speculative decoding  ( #10529 )  
							
							 
							
							... 
							
							
							
							Add length check for IPEX-CPU speculative decoding. 
							
						 
						
							2024-03-26 17:47:10 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Guancheng Fu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								a3b007f3b1 
								
							 
						 
						
							
							
								
								[Serving] Fix fastchat breaks ( #10548 )  
							
							 
							
							... 
							
							
							
							* fix fastchat
* fix doc 
							
						 
						
							2024-03-26 17:03:52 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								69a28d6b4c 
								
							 
						 
						
							
							
								
								fix chatglm ( #10540 )  
							
							 
							
							
							
						 
						
							2024-03-26 16:01:00 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Shaojun Liu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								c563b41491 
								
							 
						 
						
							
							
								
								add nightly_build workflow ( #10533 )  
							
							 
							
							... 
							
							
							
							* add nightly_build workflow
* add create-job-status-badge action
* update
* update
* update
* update setup.py
* release
* revert 
							
						 
						
							2024-03-26 12:47:38 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								0a3e4e788f 
								
							 
						 
						
							
							
								
								LLM: fix mistral hidden_size setting for deepspeed autotp ( #10527 )  
							
							 
							
							
							
						 
						
							2024-03-26 10:55:44 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								1dd40b429c 
								
							 
						 
						
							
							
								
								enable fp4 fused mlp and qkv ( #10531 )  
							
							 
							
							... 
							
							
							
							* enable fp4 fused mlp and qkv
* update qwen
* update qwen2 
							
						 
						
							2024-03-26 08:34:00 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								16b2ef49c6 
								
							 
						 
						
							
							
								
								Update_document by heyang ( #30 )  
							
							 
							
							
							
						 
						
							2024-03-25 10:06:02 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								a1048ca7f6 
								
							 
						 
						
							
							
								
								Update setup.py and add new actions and add compatible mode ( #25 )  
							
							 
							
							... 
							
							
							
							* update setup.py
* add new action
* add compatible mode 
							
						 
						
							2024-03-22 15:44:59 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
								
								
							
							
							
								
							
							
								9df70d95eb 
								
							 
						 
						
							
							
								
								Refactor bigdl.llm to  ipex_llm ( #24 )  
							
							 
							
							... 
							
							
							
							* Rename bigdl/llm to ipex_llm
* rm python/llm/src/bigdl
* from bigdl.llm to from ipex_llm 
							
						 
						
							2024-03-22 15:41:21 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Jin Qiao 
								
							 
						 
						
							
							
							
							
								
							
							
								cc5806f4bc 
								
							 
						 
						
							
							
								
								LLM: add save/load example for hf-transformers ( #10432 )  
							
							 
							
							
							
						 
						
							2024-03-22 13:57:47 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Wang, Jian4 
								
							 
						 
						
							
							
							
							
								
							
							
								34d0a9328c 
								
							 
						 
						
							
							
								
								LLM: Speed-up mixtral in pipeline parallel inference ( #10472 )  
							
							 
							
							... 
							
							
							
							* speed-up mixtral
* fix style 
							
						 
						
							2024-03-22 11:06:28 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Cengguang Zhang 
								
							 
						 
						
							
							
							
							
								
							
							
								b9d4280892 
								
							 
						 
						
							
							
								
								LLM: fix baichuan7b quantize kv abnormal output. ( #10504 )  
							
							 
							
							... 
							
							
							
							* fix abnormal output.
* fix style.
* fix style. 
							
						 
						
							2024-03-22 10:00:08 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yishuo Wang 
								
							 
						 
						
							
							
							
							
								
							
							
								f0f317b6cf 
								
							 
						 
						
							
							
								
								fix a typo in yuan ( #10503 )  
							
							 
							
							
							
						 
						
							2024-03-22 09:40:04 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Guancheng Fu 
								
							 
						 
						
							
							
							
							
								
							
							
								3a3756b51d 
								
							 
						 
						
							
							
								
								Add FastChat bigdl_worker ( #10493 )  
							
							 
							
							... 
							
							
							
							* done
* fix format
* add licence
* done
* fix doc
* refactor folder
* add license 
							
						 
						
							2024-03-21 18:35:05 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Xin Qiu 
								
							 
						 
						
							
							
							
							
								
							
							
								dba7ddaab3 
								
							 
						 
						
							
							
								
								add sdp fp8 for qwen llama436 baichuan mistral baichuan2 ( #10485 )  
							
							 
							
							... 
							
							
							
							* add sdp fp8
* fix style
* fix qwen
* fix baichuan 13
* revert baichuan 13b and baichuan2-13b
* fix style
* update 
							
						 
						
							2024-03-21 17:23:05 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Kai Huang 
								
							 
						 
						
							
							
							
							
								
							
							
								30f111cd32 
								
							 
						 
						
							
							
								
								lm_head empty_cache for more models ( #10490 )  
							
							 
							
							... 
							
							
							
							* modify constraint
* fix style 
							
						 
						
							2024-03-21 17:11:43 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Yuwen Hu 
								
							 
						 
						
							
							
							
							
								
							
							
								1579ee4421 
								
							 
						 
						
							
							
								
								[LLM] Add nightly igpu perf test for INT4+FP16 1024-128 ( #10496 )  
							
							 
							
							
							
						 
						
							2024-03-21 16:07:06 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
							
							
								
							
							
								2958ca49c0 
								
							 
						 
						
							
							
								
								LLM: add patching function for llm finetuning ( #10247 )  
							
							 
							
							
							
						 
						
							2024-03-21 16:01:01 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									Zhicun 
								
							 
						 
						
							
							
							
							
								
							
							
								5b97fdb87b 
								
							 
						 
						
							
							
								
								update deepseek example readme ( #10420 )  
							
							 
							
							... 
							
							
							
							* update readme
* update
* update readme 
							
						 
						
							2024-03-21 15:21:48 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									hxsz1997 
								
							 
						 
						
							
							
							
							
								
							
							
								a5f35757a4 
								
							 
						 
						
							
							
								
								Migrate langchain rag cpu example to gpu ( #10450 )  
							
							 
							
							... 
							
							
							
							* add langchain rag on gpu
* add rag example in readme
* add trust_remote_code in TransformersEmbeddings.from_model_id
* add trust_remote_code in TransformersEmbeddings.from_model_id in cpu 
							
						 
						
							2024-03-21 15:20:46 +08:00  
						
						
							 
							
							
								 
							 
							
						 
					 
				
					
						
							
								
								
									 
									binbin Deng 
								
							 
						 
						
							
							
							
							
								
							
							
								85ef3f1d99 
								
							 
						 
						
							
							
								
								LLM: add empty cache in deepspeed autotp benchmark script ( #10488 )  
							
							 
							
							
							
						 
						
							2024-03-21 10:51:23 +08:00