Commit graph

2137 commits

Author SHA1 Message Date
Jin Qiao
2f4c754759 LLM: add bark gpu example (#10091)
* add bark gpu example

* fix

* fix license

* add bark

* add example

* fix

* another way
2024-02-07 10:47:11 +08:00
Xiangyu Tian
8953acd7d6 [LLM] Fix log condition for BIGDL_OPT_IPEX (#10115)
Fix log condition for BIGDL_OPT_IPEX
2024-02-07 10:27:10 +08:00
yb-peng
3f60e9df89 Merge pull request #10101 from pengyb2001/eval_stat
Modify harness evaluation workflow
2024-02-07 00:02:57 +08:00
pengyb2001
f63eba6c5a change pr test machine 2024-02-06 23:35:18 +08:00
pengyb2001
e627727b4b change download path 2024-02-06 21:12:51 +08:00
pengyb2001
2c4e610743 remove irrelevant code 2024-02-06 20:12:10 +08:00
Jason Dai
e2233dddef Update README (#10111) 2024-02-06 19:29:07 +08:00
SONG Ge
0eccb94d75 remove text-generation-webui from bigdl repo (#10107) 2024-02-06 17:46:52 +08:00
Ovo233
2aaa21c41d LLM: Update ppl tests (#10092)
* update ppl tests

* use load_dataset api

* add exception handling

* add language argument

* address comments
2024-02-06 17:31:48 +08:00
Yuwen Hu
3a46b57253 [LLM] Add RWKV4 HF GPU Example (#10105)
* Add GPU HF example for RWKV 4

* Add link to rwkv4

* fix
2024-02-06 16:30:24 +08:00
Yuwen Hu
518ef95abc Small fix for Nonetype error (#10104) 2024-02-06 14:58:52 +08:00
Ruonan Wang
d61f4905ac LLM: 2bit quantization initial support (#10042)
* basis quantize support

* fix new module name

* small update

* and mixed int4 with iq2_xxs

* remove print

* code refactor

* fix style

* meet code review
2024-02-06 14:58:32 +08:00
pengyb2001
d11ef0d117 remove retry in llm install part 2024-02-06 14:25:26 +08:00
pengyb2001
94723bb0b1 add retry in run llm install part;test arc05 with llama2 2024-02-06 14:09:14 +08:00
pengyb2001
2c75b5b981 remove mistral in pr job 2024-02-06 13:51:57 +08:00
pengyb2001
5edefe7d8e remove nightly summary job 2024-02-06 13:50:38 +08:00
Jason Dai
f440cb4fba Update Self-Speculative Decoding Readme (#10102) 2024-02-06 12:59:17 +08:00
pengyb2001
bc92dbf7be remove stableml;change schedule;change storage method 2024-02-06 11:20:37 +08:00
dingbaorong
36c9442c6d Arc Stable version test (#10087)
* add batch_size in stable version test

* add batch_size in excludes

* add excludes for batch_size

* fix ci

* triger regression test

* fix xpu version

* disable ci

* address kai's comment

---------

Co-authored-by: Ariadne <wyn2000330@126.com>
2024-02-06 10:23:50 +08:00
Jiao Wang
33b9e7744d fix dimension (#10097) 2024-02-05 15:07:38 -08:00
SONG Ge
4b02ff188b [WebUI] Add prompt format and stopping words for Qwen (#10066)
* add prompt format and stopping_words for qwen mdoel

* performance optimization

* optimize

* update

* meet comments
2024-02-05 18:23:13 +08:00
WeiguangHan
0aecd8637b LLM: small fix for the html script (#10094) 2024-02-05 17:27:34 +08:00
Zhicun
7d2be7994f add phixtral and optimize phi-moe (#10052) 2024-02-05 11:12:47 +08:00
Zhicun
676d6923f2 LLM: modify transformersembeddings.embed() in langchain (#10051) 2024-02-05 10:42:10 +08:00
Jin Qiao
ad050107b3 LLM: fix mpt load_low_bit issue (#10075)
* fix

* retry

* retry
2024-02-05 10:17:07 +08:00
Lilac09
f8dcaff7f4 use default python (#10070) 2024-02-05 09:06:59 +08:00
SONG Ge
9050991e4e fix gradio check issue temply (#10082) 2024-02-04 16:46:29 +08:00
WeiguangHan
c2e562d037 LLM: add batch_size to the csv and html (#10080)
* LLM: add batch_size to the csv and html

* small fix
2024-02-04 16:35:44 +08:00
Yuwen Hu
136f042f84 [LLM] Make sure python 310-311 tests only happen for nightly tests (#10081)
* Make sure python 310-311 tests only happen for nightly tests

* Use default runner for setup-python-version

* Small fixes
2024-02-04 16:14:48 +08:00
binbin Deng
7e49fbc5dd LLM: make finetuning examples more common for other models (#10078) 2024-02-04 16:03:52 +08:00
Heyang Sun
90f004b80b remove benchmarkwrapper form deepspeed example (#10079) 2024-02-04 15:42:15 +08:00
Jin Qiao
f9a468a2c7 LLM: conditionally choose python version for unit test (#10062)
* conditional python version

* retry

* temporary skip llm-cpp-build

* apply on llm-unit-test-on-arc

* fix

* add llm-cpp-build dependency

* use GITHUB_OUTPUT instead of set-output

* check nightly build

* fix quote

* fix quote

* add llm-cpp-build dependency

* test nightly build

* test pull request
2024-02-04 13:37:34 +08:00
Ruonan Wang
8e33cb0f38 LLM: support speecht5_tts (#10077)
* support speecht5_tts

* fix
2024-02-04 13:26:42 +08:00
yb-peng
738275761d In llm-harness-evaluation, add new models and change schedule to nightly (#10072)
* add new models and change schedule to nightly

* correct syntax error

* modify env set up and job

* change label and schedule time

* change schedule time

* change label
2024-02-04 13:12:09 +08:00
Shaojun Liu
698f84648c split stable version tests (#10076)
Co-authored-by: Your Name <Your Email>
2024-02-04 11:08:12 +08:00
ivy-lv11
428b7105f6 Add HF and PyTorch example InternLM2 (#10061) 2024-02-04 10:25:55 +08:00
binbin Deng
91cf9d41d0 LLM: add solutions of some frequently asked questions (#10068) 2024-02-04 09:28:20 +08:00
Yina Chen
77be19bb97 LLM: Support gpt-j in speculative decoding (#10067)
* gptj

* support gptj in speculative decoding

* fix

* update readme

* small fix
2024-02-02 14:54:55 +08:00
Jason Dai
2927c77d7f Update readme (#10071) 2024-02-01 20:40:20 -08:00
SONG Ge
19183ef476 [WebUI] Reset bigdl-llm loader options with default value (#10064)
* reset bigdl-llm loader options with default value

* remove options which maybe complex for naive users
2024-02-01 15:45:39 +08:00
Xin Qiu
6e0f1a1e92 use apply_rotary_pos_emb_cache_freq_xpu in mixtral (#10060)
* use apply_rotary_pos_emb_cache_freq_xpu in mixtral

* fix style
2024-02-01 15:40:49 +08:00
binbin Deng
aae20d728e LLM: Add initial DPO finetuning example (#10021) 2024-02-01 14:18:08 +08:00
Heyang Sun
601024f418 Mistral CPU example of speculative decoding (#10024)
* Mistral CPU example of speculative decoding

* update transformres version

* update example

* Update README.md
2024-02-01 10:52:32 +08:00
Heyang Sun
968e70544d Enable IPEX Mistral in Speculative (#10059) 2024-02-01 10:48:16 +08:00
Yina Chen
3ca03d4e97 Add deepmind sample into bigdl-llm speculative decoding (#10041)
* migrate deepmind sample

* update

* meet comments

* fix style

* fix style
2024-02-01 09:57:02 +08:00
Lilac09
72e67eedbb Add speculative support in docker (#10058)
* add speculative environment

* add speculative environment

* add speculative environment
2024-02-01 09:53:53 +08:00
binbin Deng
4b92235bdb LLM: add initial FAQ page (#10055) 2024-02-01 09:43:39 +08:00
WeiguangHan
d2d3f6b091 LLM: ensure the result of daily arc perf test (#10016)
* ensure the result of daily arc perf test

* small fix

* small fix

* small fix

* small fix

* small fix

* small fix

* small fix

* small fix

* small fix

* small fix

* concat more csvs

* small fix

* revert some files
2024-01-31 18:26:21 +08:00
WeiguangHan
9724939499 temporarily disable bloom 2k input (#10056) 2024-01-31 17:49:12 +08:00
Jin Qiao
8c8fc148c9 LLM: add rwkv 5 (#10048) 2024-01-31 15:54:55 +08:00