Ruonan Wang
ca3e59a1dc
LLM: support stop for starcoder native int4 stream ( #8734 )
2023-08-11 14:51:30 +08:00
Song Jiaming
e292dfd970
[WIP] LLM transformers api for langchain ( #8642 )
2023-08-11 13:32:35 +08:00
Yishuo Wang
3d5a7484a2
[LLM] fix bloom and starcoder memory release ( #8728 )
2023-08-11 11:18:19 +08:00
xingyuan li
33d9ad234f
[LLM] Linux vnni build with ubuntu 18.04 ( #8710 )
...
* move from almalinux
2023-08-10 19:04:03 +09:00
xingyuan li
02ec01cb48
[LLM] Add bigdl-core-xe dependency when installing bigdl-llm[xpu] ( #8716 )
...
* add bigdl-core-xe dependency
2023-08-10 17:41:42 +09:00
Shengsheng Huang
7c56c39e36
Fix GPU examples READ to use bigdl-core-xe ( #8714 )
...
* Update README.md
* Update README.md
2023-08-10 12:53:49 +08:00
Yina Chen
6d1ca88aac
add voice assistant example ( #8711 )
2023-08-10 12:42:14 +08:00
Kai Huang
4573ff7ffe
Update llm api doc ( #8712 )
...
* change to bigdlllm
* update
2023-08-10 12:17:19 +08:00
Song Jiaming
e717e304a6
LLM first example test and template ( #8658 )
2023-08-10 10:03:11 +08:00
Ruonan Wang
1a7b698a83
[LLM] support ipex arc int4 & add basic llama2 example ( #8700 )
...
* first support of xpu
* make it works on gpu
update setup
update
add GPU llama2 examples
add use_optimize flag to disbale optimize for gpu
fix style
update gpu exmaple readme
fix
* update example, and update env
* fix setup to add cpp files
* replace jit with aot to avoid data leak
* rename to bigdl-core-xe
* update installation in example readme
2023-08-09 22:20:32 +08:00
Jason Dai
d03218674a
Update llm readme ( #8703 )
2023-08-09 14:47:26 +08:00
Kai Huang
1b65288bdb
Add api doc for LLM ( #8605 )
...
* api doc initial
* update desc
2023-08-08 18:17:16 +08:00
binbin Deng
4c44153584
LLM: add Qwen transformers int4 example ( #8699 )
2023-08-08 11:23:09 +08:00
Yishuo Wang
710b9b8982
[LLM] add linux chatglm pybinding binary file ( #8698 )
2023-08-08 11:16:30 +08:00
xingyuan li
4482ccb329
[LLM] Change build system from centos7 to ubuntu18.04 ( #8686 )
...
* centos7 to ubuntu18
* ubuntu git version 2.17 need to update
* use almalinux8 to build avx2 binaries
2023-08-07 19:09:58 +09:00
binbin Deng
ea5d7aff5b
LLM: add chatglm native int4 transformers API ( #8695 )
2023-08-07 17:52:47 +08:00
Yishuo Wang
6da830cf7e
[LLM] add chaglm pybinding binary file in setup.py ( #8692 )
2023-08-07 09:41:03 +08:00
Yuwen Hu
06ed70fa43
Update google site verification for readthedocs ( #8689 )
2023-08-04 16:05:28 +08:00
Yuwen Hu
418cb06850
Renew google site verification for readthedocs ( #8687 )
2023-08-04 15:36:17 +08:00
Cengguang Zhang
ebcf75d506
feat: set transformers lib version. ( #8683 )
2023-08-04 15:01:59 +08:00
Yishuo Wang
ef08250c21
[LLM] chatglm pybinding support ( #8672 )
2023-08-04 14:27:29 +08:00
Yishuo Wang
5837cc424a
[LLM] add chatglm pybinding binary file release ( #8677 )
2023-08-04 11:45:27 +08:00
xingyuan li
bc4cdb07c9
Remove conda for llm workflow ( #8671 )
2023-08-04 12:09:42 +09:00
Yang Wang
b6468bac43
optimize chatglm2 long sequence ( #8662 )
...
* add chatglm2
* optimize a little
* optimize chatglm long sequence
* fix style
* address comments and fix style
* fix bug
2023-08-03 17:56:24 -07:00
Yang Wang
3407f87075
Fix llama kv cache bug ( #8674 )
2023-08-03 17:54:55 -07:00
Yina Chen
59903ea668
llm linux support avx & avx2 ( #8669 )
2023-08-03 17:10:59 +08:00
xingyuan li
110cfb5546
[LLM] Remove old windows nightly test code ( #8668 )
...
Remove old Windows nightly test code triggered by task scheduler
Add new Windows nightly workflow for nightly testing
2023-08-03 17:12:23 +09:00
Yina Chen
bd177ab612
[LLM] llm binary build linux add avx & avx2 ( #8665 )
...
* llm add linux avx & avx2 release
* fix name
* update check
2023-08-03 14:38:31 +08:00
xingyuan li
610084e3c0
[LLM] Complete windows unittest ( #8611 )
...
* add windows nightly test workflow
* use github runner to run pr test
* model load should use lowbit
* remove tmp dir after testing
2023-08-03 14:48:42 +09:00
binbin Deng
a15a2516e6
add ( #8659 )
2023-08-03 10:12:10 +08:00
Xin Qiu
0714888705
build windows avx dll ( #8657 )
...
* windows avx
* add to actions
2023-08-03 02:06:24 +08:00
Yina Chen
119bf6d710
[LLM] Support linux cpp dynamic load .so ( #8655 )
...
* support linux cpp dynamic load .so
* update cli
2023-08-02 20:15:45 +08:00
Zhao Changmin
ca998cc6f2
LLM: Mute shape mismatch output ( #8601 )
...
* LLM: Mute shape mismatch output
2023-08-02 16:46:22 +08:00
Yina Chen
15b3adc7ec
[LLM] llm linux binary make -> cmake ( #8656 )
...
* llm linux make -> cmake
* update
* update
2023-08-02 16:41:54 +08:00
Zhao Changmin
04c713ef06
LLM: Disable transformer api pretraining_tp ( #8645 )
...
* disable pretraining_tp
2023-08-02 11:26:01 +08:00
binbin Deng
6fc31bb4cf
LLM: first update descriptions for ChatGLM transformers int4 example ( #8646 )
2023-08-02 11:00:56 +08:00
xingyuan li
769209b7f0
Chatglm unittest disable due to missing instruction ( #8650 )
2023-08-02 10:28:42 +09:00
Yang Wang
cbeae97a26
Optimize Llama Attention to to reduce KV cache memory copy ( #8580 )
...
* Optimize llama attention to reduce KV cache memory copy
* fix bug
* fix style
* remove git
* fix style
* fix style
* fix style
* fix tests
* move llama attention to another file
* revert
* fix style
* remove jit
* fix
2023-08-01 16:37:58 -07:00
binbin Deng
39994738d1
LLM: add chat & stream chat example for ChatGLM2 transformers int4 ( #8636 )
2023-08-01 14:57:45 +08:00
xingyuan li
cdfbe652ca
[LLM] Add chatglm support for llm-cli ( #8641 )
...
* add chatglm build
* add llm-cli support
* update git
* install cmake
* add ut for chatglm
* add files to setup
* fix bug cause permission error when sf lack file
2023-08-01 14:30:17 +09:00
Zhao Changmin
d6cbfc6d2c
LLM: Add requirements in whisper example ( #8644 )
...
* LLM: Add requirements in whisper example
2023-08-01 12:07:14 +08:00
Zhao Changmin
3e10260c6d
LLM: llm-convert support chatglm family ( #8643 )
...
* convert chatglm
2023-08-01 11:16:18 +08:00
Yina Chen
a607972c0b
[LLM]LLM windows load -api.dll ( #8631 )
...
* temp
* update
* revert setup.py
2023-07-31 13:47:20 +08:00
xingyuan li
3361b66449
[LLM] Revert llm-cli to disable selecting executables on Windows ( #8630 )
...
* revert vnni file select
* revert setup.py
* add model-api.dll
2023-07-31 11:15:44 +09:00
binbin Deng
3dbab9087b
LLM: add llama2-7b native int4 example ( #8629 )
2023-07-28 10:56:16 +08:00
binbin Deng
fb32fefcbe
LLM: support tensor input of native int4 generate ( #8620 )
2023-07-27 17:59:49 +08:00
Zhao Changmin
5b484ab48d
LLM: Support load_low_bit loading models in shards format ( #8612 )
...
* shards_model
---------
Co-authored-by: leonardozcm <leonaordo1997zcm@gmail.com>
2023-07-26 13:30:01 +08:00
xingyuan li
919791e406
Add needs to make sure run in order ( #8621 )
2023-07-26 14:16:57 +09:00
xingyuan li
e3418d7e61
[LLM] Remove concurrency group for binary build workflow ( #8619 )
...
* remove concurrency group for nightly test
2023-07-26 12:15:53 +09:00
binbin Deng
fcf8c085e3
LLM: add llama2-13b native int4 example ( #8613 )
2023-07-26 10:12:52 +08:00