Ruonan Wang
|
bf51ec40b2
|
LLM: Fix empty cache (#9024)
* fix
* fix
* update example
|
2023-09-21 17:16:07 +08:00 |
|
Yang Wang
|
c88f6ec457
|
Experiment XPU QLora Finetuning (#8937)
* Support xpu finetuning
* support xpu finetuning
* fix style
* fix style
* fix style
* refine example
* add readme
* refine readme
* refine api
* fix fp16
* fix example
* refactor
* fix style
* fix compute type
* add qlora
* refine training args
* fix example
* fix style
* fast path forinference
* address comments
* refine readme
* revert lint
|
2023-09-19 10:15:44 -07:00 |
|
Ruonan Wang
|
cabe7c0358
|
LLM: add baichuan2 example for arc (#8994)
* add baichuan2 examples
* add link
* small fix
|
2023-09-18 14:32:27 +08:00 |
|
JinBridge
|
c12b8f24b6
|
LLM: add use_cache=True for all gpu examples (#8971)
|
2023-09-15 09:54:38 +08:00 |
|
binbin Deng
|
be29c75c18
|
LLM: refactor gpu examples (#8963)
* restructure
* change to hf-transformers-models/
|
2023-09-13 14:47:47 +08:00 |
|
Ruonan Wang
|
4de73f592e
|
LLM: add gpu example of chinese-llama-2-7b (#8960)
* add gpu example of chinese -llama2
* update model name and link
* update name
|
2023-09-13 10:16:51 +08:00 |
|
Yina Chen
|
bfc71fbc15
|
Add known issue in arc voice assistant example (#8902)
* add known issue in voice assistant example
* update cpu
|
2023-09-07 09:28:26 +08:00 |
|
Yina Chen
|
74a2c2ddf5
|
Update optimize_model=True in llama2 chatglm2 arc examples (#8878)
* add optimize_model=True in llama2 chatglm2 examples
* add ipex optimize in gpt-j example
|
2023-09-05 10:35:37 +08:00 |
|
Ruonan Wang
|
f42c0bad1b
|
LLM: update GPU doc (#8845)
|
2023-08-30 09:24:19 +08:00 |
|
Jason Dai
|
aab7deab1f
|
Reorganize GPU examples (#8844)
|
2023-08-30 08:32:08 +08:00 |
|