Ruonan Wang
9fda7e34f1
LLM: fix version control ( #8342 )
2023-06-15 15:18:50 +08:00
Ruonan Wang
f7f4e65788
LLM: support int8 and tmp_path for from_pretrained ( #8338 )
2023-06-15 14:48:21 +08:00
Yuwen Hu
b30aa49c4e
[LLM] Add Actions for downloading & converting models ( #8320 )
...
* First push to downloading and converting llm models for testing (Gondolin runner, avx2 for now)
* Change yml file name
2023-06-15 13:43:47 +08:00
ZehuaCao
bc11a2b1cd
Update distributed-tuning.md ( #8324 )
2023-06-15 11:29:37 +08:00
Ruonan Wang
8840dadd86
LLM: binary file version control on source forge ( #8329 )
...
* support version control for llm based on date
* update action
2023-06-15 09:53:27 +08:00
Ruonan Wang
5094970175
LLM: update convert_model to support int8 ( #8326 )
...
* update example and convert_model for int8
* reset example
* fix style
2023-06-15 09:25:07 +08:00
binbin Deng
f64e703083
LLM: first add _tokenize, detokenize and _generate for bloom pybinding ( #8316 )
2023-06-14 17:29:57 +08:00
Xin Qiu
5576679a92
add convert-gptq-to-ggml.py to bigdl-llama ( #8298 )
2023-06-14 14:51:51 +08:00
Ruonan Wang
a6c4b733cb
LLM: Update subprocess to show error message ( #8323 )
...
* update subprocess
* fix style
2023-06-13 16:43:37 +08:00
Shengsheng Huang
02c583144c
[LLM] langchain integrations and examples ( #8256 )
...
* langchain intergrations and examples
* add licences and rename
* add licences
* fix license issues and change backbone to model_family
* update examples to use model_family param
* fix linting
* fix code style
* exclude langchain integration from stylecheck
* update langchain examples and update integrations based on latets changes
* update simple llama-cpp-python style API example
* remove bloom in README
* change default n_threads to 2 and remove redundant code
---------
Co-authored-by: leonardozcm <changmin.zhao@intel.com>
2023-06-12 19:22:07 +08:00
Yuwen Hu
f83c48280f
[LLM] Unify transformers-like API example for 3 different model families ( #8315 )
...
* Refactor bigdl-llm transformers-like API to unify them
* Small fix
2023-06-12 17:20:30 +08:00
xingyuan li
c4028d507c
[LLM] Add unified default value for cli programs ( #8310 )
...
* add unified default value for threads and n_predict
2023-06-12 16:30:27 +08:00
Junwei Deng
f41995051b
LLM: add new readme as first version document ( #8296 )
...
* add new readme
* revice
* revice
* change readme
* add python req
2023-06-09 15:52:02 +08:00
Yuwen Hu
c619315131
[LLM] Add examples for gptneox, llama, and bloom family model using transformers-like API ( #8286 )
...
* First push of bigdl-llm example for gptneox model family
* Add some args and other small updates
* Small updates
* Add example for llama family models
* Small fix
* Small fix
* Update for batch_decode api and change default model for llama example
* Small fix
* Small fix
* Small fix
* Small model family name fix and add example for bloom
* Small fix
* Small default prompt fix
* Small fix
* Change default prompt
* Add sample output for inference
* Hide example inference time
2023-06-09 15:48:22 +08:00
binbin Deng
5d5da7b2c7
LLM: optimize namespace and remove unused import logic ( #8302 )
2023-06-09 15:17:49 +08:00
Ruonan Wang
5d0e130605
LLM: fix convert path error of gptneox and bloom on windows ( #8304 )
2023-06-09 10:10:19 +08:00
Yina Chen
7bfa0fcdf9
fix style ( #8300 )
2023-06-08 16:52:17 +08:00
Yina Chen
637b72f2ad
[LLM] llm transformers api support batch actions ( #8288 )
...
* llm transformers api support batch actions
* align with transformer
* meet comment
2023-06-08 15:10:08 +08:00
xingyuan li
ea3cf6783e
LLM: Command line wrapper for llama/bloom/gptneox ( #8239 )
...
* add llama/bloom/gptneox wrapper
* add readme
* upload binary main file
2023-06-08 14:55:22 +08:00
binbin Deng
08bdfce2d8
LLM: avoid unnecessary import torch except converting process ( #8297 )
2023-06-08 14:24:58 +08:00
binbin Deng
f9e2bda04a
LLM: add stop words and enhance output for bloom pybinding ( #8280 )
2023-06-08 14:06:06 +08:00
Yina Chen
6990328e5c
[LLM]Add bloom quantize in setup.py ( #8295 )
...
* add bloom quantize in setup.py
* fix
2023-06-08 11:18:22 +08:00
Yina Chen
1571ba6425
remove unused import gptneox_cpp ( #8293 )
2023-06-08 11:04:47 +08:00
Ruonan Wang
aa91657019
LLM: add bloom dll/exe in setup ( #8284 )
2023-06-08 09:28:28 +08:00
Pingchuan Ma (Henry)
773255e009
[LLM] Add dev wheel building and basic UT script for LLM package on Linux ( #8264 )
...
* add wheel build for linux
* test fix
* test self-hosted runner
* test fix
* update runner
* update runner
* update fix
* init cicd
* init cicd
* test conda
* update fix
* update no need manual python deps
* test fix bugs
* test fix bugs
* test fix bugs
* fix bugs
2023-06-08 00:49:57 +08:00
Yina Chen
2c037e892b
fix-transformers-neox ( #8285 )
2023-06-07 14:44:43 +08:00
Ruonan Wang
39ad68e786
LLM: enhancements for convert_model ( #8278 )
...
* update convert
* change output name
* add discription for input_path, add check for input_values
* basic support for command line
* fix style
* update based on comment
* update based on comment
2023-06-07 13:22:14 +08:00
Junwei Deng
2d14e593f0
LLM: Support generate(max_new_tokens=...), tokenize and decode for transformers-like API ( #8283 )
...
* first push
* fix pep8
2023-06-07 11:50:35 +08:00
Yina Chen
11cd2a07e0
[LLM] llm transformers format interface first part ( #8276 )
...
* llm-transformers-format
* update
* fix style
2023-06-06 17:17:37 +08:00
Pingchuan Ma (Henry)
a3f353b939
[LLM] add long time loading disclaimer for LLM model converting ( #8279 )
2023-06-06 17:15:13 +08:00
Yuwen Hu
64bc123dd3
[LLM] Add transformers-like API from_pretrained ( #8271 )
...
* Init commit for bigdl.llm.transformers.AutoModelForCausalLM
* Temp change to avoid name conflicts with external transformers lib
* Support downloading model from huggingface
* Small python style fix
* Change location of transformers to avoid library conflicts
* Add return value for converted ggml binary ckpt path for convert_model
* Avoid repeated loading of shared library and adding some comments
* Small fix
* Path type fix anddocstring fix
* Small fix
* Small fix
* Change cache dir to pwd
2023-06-06 17:04:16 +08:00
Pingchuan Ma (Henry)
2ed5842448
[LLM] add convert's python deps for LLM ( #8260 )
...
* add python deps for LLM
* update release.sh
* change deps group name
* update all
* fix update
* test fix
* update
2023-06-06 16:01:17 +08:00
xingyuan li
38be471140
[LLM] convert_model bug fix ( #8274 )
...
* Renamed all bloomz to bloom in ggml/model & utls/convert_util.py
* Add an optional parameter for specific the model conversion path to avoid running out of disk space
2023-06-06 15:16:42 +08:00
Ruonan Wang
8bd2992a8d
LLM: accelerate sample of gptneox and update quantize ( #8262 )
...
* update quantize & accelerate sample
* fix style check
* fix style error
2023-06-05 15:36:00 +08:00
Jun Wang
2bc0e7abbb
[llm] Add convert_model api ( #8244 )
...
* add convert_model api
* change the model_path to input_path
* map int4 to q4_0
* fix blank line
* change bloomz to bloom
* remove default model_family
* change dtype to lower first
2023-06-03 10:18:29 +08:00
Yuwen Hu
e290660b20
[LLM] Add so shared library for Bloom family models ( #8258 )
...
* Add so file downloading for bloom family models
* Supports selecting of avx2/avx512 so for bloom
2023-06-02 17:39:40 +08:00
Pingchuan Ma (Henry)
c48d5f7cff
[LLM] Enable UT workflow logics for LLM ( #8243 )
...
* check push connection
* enable UT workflow logics for LLM
* test fix
* add licenses
* test fix according to suggestions
* test fix
* update changes
2023-06-02 17:06:35 +08:00
Yina Chen
657ea0ee50
[LLM] Fix linux load libs for NeoX and llama ( #8257 )
...
* init
* add lisence
* fix style
2023-06-02 17:03:17 +08:00
Yuwen Hu
286b010bf1
[LLM] First push for Bloomz pybinding ( #8252 )
...
* Initial commit to move bloom pybinding to bigdl-llm
* Revise path for shared library
* Small fix
2023-06-02 14:41:04 +08:00
Yina Chen
91a1528fce
[LLM]Support for linux package (llama, NeoX) & quantize (llama) ( #8246 )
...
* temp
* update
* update
* remove cmake
* runtime get platform -> change platform name using sed
* update
* update
* add platform flags(default: current platform) & delete legacy libs & add neox quantize
2023-06-02 13:51:35 +08:00
Junwei Deng
350d31a472
LLM: first push gptneox pybinding ( #8234 )
...
* first push gptneox pybinding
* fix
* fix code style and add license
---------
Co-authored-by: binbin <binbin1.deng@intel.com>
2023-06-02 09:28:00 +08:00
binbin Deng
3a9aa23835
LLM: fix and update related license in llama pybinding ( #8250 )
2023-06-01 17:09:15 +08:00
Pingchuan Ma (Henry)
141febec1f
Add dev wheel building script for LLM package on Windows ( #8238 )
...
* Add dev wheel building script for LLM package on Windows
* delete conda
* delete python version check
* minor adjust
* wheel name fixed
* test check
* test fix
* change wheel name
2023-06-01 11:55:26 +08:00
binbin Deng
e56f24b424
LLM: first push llama pybinding ( #8241 )
...
* first push llama binding
* update dll
2023-06-01 10:59:15 +08:00
Ruonan Wang
3fd716d422
LLM: update setup.py to add a missing data( #8240 )
2023-06-01 10:25:43 +08:00
binbin Deng
8421af51ae
LLM: support converting to ggml format ( #8235 )
...
* add convert
* fix
* fix
* fix
* try
* test
* update check
* fix
* fix
2023-05-31 15:20:06 +08:00
Ruonan Wang
c890609d1e
LLM: Support package/quantize for llama.cpp/redpajama.cpp on Windows ( #8236 )
...
* support windows of llama.cpp
* update quantize
* update version of llama.cp submodule
* add gptneox.dll
* add quantize-gptneox.exe
2023-05-31 14:47:12 +08:00
Yishuo Wang
fa0b0a4555
[nano] fix optuna version ( #8237 )
2023-05-31 13:44:04 +08:00
Pingchuan Ma (Henry)
1f913a6941
[LLM] Add LLM pep8 coding style checking ( #8233 )
...
* add LLM pep8 coding checking
* resolve bugs in testing scripts and code style revision
2023-05-30 15:58:14 +08:00
Ruonan Wang
4638b85f3e
[llm] Initial support of package and quantize ( #8228 )
...
* first commit of CMakeFiles.txt to include llama & gptneox
* initial support of quantize
* update cmake for only consider linux now
* support quantize interface
* update based on comment
2023-05-26 16:36:46 +08:00