Wang, Jian4
|
9df70d95eb
|
Refactor bigdl.llm to ipex_llm (#24)
* Rename bigdl/llm to ipex_llm
* rm python/llm/src/bigdl
* from bigdl.llm to from ipex_llm
|
2024-03-22 15:41:21 +08:00 |
|
Wang, Jian4
|
fe8976a00f
|
LLM: Support gguf models use low_bit and fix no json(#10408)
* support others model use low_bit
* update readme
* update to add *.json
|
2024-03-15 09:34:18 +08:00 |
|
Wang, Jian4
|
0193f29411
|
LLM : Enable gguf float16 and Yuan2 model (#10372)
* enable float16
* add yun files
* enable yun
* enable set low_bit on yuan2
* update
* update license
* update generate
* update readme
* update python style
* update
|
2024-03-13 10:19:18 +08:00 |
|
Heyang Sun
|
5184f400f9
|
Fix Mixtral GGUF Wrong Output Issue (#9930)
* Fix Mixtral GGUF Wrong Output Issue
* fix style
* fix style
|
2024-01-18 14:11:27 +08:00 |
|
ZehuaCao
|
146076bdb5
|
Support llm-awq backend (#9856)
* Support for LLM-AWQ Backend
* fix
* Update README.md
* Add awqconfig
* modify init
* update
* support llm-awq
* fix style
* fix style
* update
* fix AwqBackendPackingMethod not found error
* fix style
* update README
* fix style
---------
Co-authored-by: Uxito-Ada <414416158@qq.com>
Co-authored-by: Heyang Sun <60865256+Uxito-Ada@users.noreply.github.com>
Co-authored-by: cyita <yitastudy@gmail.com>
|
2024-01-09 13:07:32 +08:00 |
|
Wang, Jian4
|
a54cd767b1
|
LLM: Add gguf falcon (#9801)
* init falcon
* update convert.py
* update style
|
2024-01-03 14:49:02 +08:00 |
|
Wang, Jian4
|
7ed9538b9f
|
LLM: support gguf mpt (#9773)
* add gguf mpt
* update
|
2023-12-28 09:22:39 +08:00 |
|
Wang, Jian4
|
984697afe2
|
LLM: Add bloom gguf support (#9734)
* init
* update bloom add merges
* update
* update readme
* update for llama error
* update
|
2023-12-21 14:06:25 +08:00 |
|
Heyang Sun
|
1fa7793fc0
|
Load Mixtral GGUF Model (#9690)
* Load Mixtral GGUF Model
* refactor
* fix empty tensor when to cpu
* update gpu and cpu readmes
* add dtype when set tensor into module
|
2023-12-19 13:54:38 +08:00 |
|
Wang, Jian4
|
b8437a1c1e
|
LLM: Add gguf mistral model support (#9691)
* add mistral support
* need to upgrade transformers version
* update
|
2023-12-15 13:37:39 +08:00 |
|
Wang, Jian4
|
496bb2e845
|
LLM: Support load BaiChuan model family gguf model (#9685)
* support baichuan model family gguf model
* update gguf generate.py
* add verify models
* add support model_family
* update
* update style
* update type
* update readme
* update
* remove support model_family
|
2023-12-15 13:34:33 +08:00 |
|
Jason Dai
|
51b668f229
|
Update GGUF readme (#9611)
|
2023-12-06 18:21:54 +08:00 |
|
dingbaorong
|
a7bc89b3a1
|
remove q4_1 in gguf example (#9610)
* remove q4_1
* fixes
|
2023-12-06 16:00:05 +08:00 |
|
dingbaorong
|
89069d6173
|
Add gpu gguf example (#9603)
* add gpu gguf example
* some fixes
* address kai's comments
* address json's comments
|
2023-12-06 15:17:54 +08:00 |
|