Zijie Li
|
a644e9409b
|
Miniconda/Anaconda -> Miniforge update in examples (#11194)
* Change installation address
Change former address: "https://docs.conda.io/en/latest/miniconda.html#" to new address: "https://conda-forge.org/download/" for 63 occurrences under python\llm\example
* Change Prompt
Change "Anaconda Prompt" to "Miniforge Prompt" for 1 occurrence
|
2024-06-04 10:14:02 +08:00 |
|
ZehuaCao
|
56cb992497
|
LLM: Modify CPU Installation Command for most examples (#11049)
* init
* refine
* refine
* refine
* modify hf-agent example
* modify all CPU model example
* remove readthedoc modify
* replace powershell with cmd
* fix repo
* fix repo
* update
* remove comment on windows code block
* update
* update
* update
* update
---------
Co-authored-by: xiangyuT <xiangyu.tian@intel.com>
|
2024-05-17 15:52:20 +08:00 |
|
Shaojun Liu
|
f37a1f2a81
|
Upgrade to python 3.11 (#10711)
* create conda env with python 3.11
* recommend to use Python 3.11
* update
|
2024-04-09 17:41:17 +08:00 |
|
Jin Qiao
|
10ee786920
|
Replace with IPEX-LLM in example comments (#10671)
* Replace with IPEX-LLM in example comments
* More replacement
* revert some changes
|
2024-04-07 13:29:51 +08:00 |
|
Wang, Jian4
|
16b2ef49c6
|
Update_document by heyang (#30)
|
2024-03-25 10:06:02 +08:00 |
|
Wang, Jian4
|
9df70d95eb
|
Refactor bigdl.llm to ipex_llm (#24)
* Rename bigdl/llm to ipex_llm
* rm python/llm/src/bigdl
* from bigdl.llm to from ipex_llm
|
2024-03-22 15:41:21 +08:00 |
|
Wang, Jian4
|
fe8976a00f
|
LLM: Support gguf models use low_bit and fix no json(#10408)
* support others model use low_bit
* update readme
* update to add *.json
|
2024-03-15 09:34:18 +08:00 |
|
Wang, Jian4
|
0193f29411
|
LLM : Enable gguf float16 and Yuan2 model (#10372)
* enable float16
* add yun files
* enable yun
* enable set low_bit on yuan2
* update
* update license
* update generate
* update readme
* update python style
* update
|
2024-03-13 10:19:18 +08:00 |
|
Heyang Sun
|
5184f400f9
|
Fix Mixtral GGUF Wrong Output Issue (#9930)
* Fix Mixtral GGUF Wrong Output Issue
* fix style
* fix style
|
2024-01-18 14:11:27 +08:00 |
|
ZehuaCao
|
146076bdb5
|
Support llm-awq backend (#9856)
* Support for LLM-AWQ Backend
* fix
* Update README.md
* Add awqconfig
* modify init
* update
* support llm-awq
* fix style
* fix style
* update
* fix AwqBackendPackingMethod not found error
* fix style
* update README
* fix style
---------
Co-authored-by: Uxito-Ada <414416158@qq.com>
Co-authored-by: Heyang Sun <60865256+Uxito-Ada@users.noreply.github.com>
Co-authored-by: cyita <yitastudy@gmail.com>
|
2024-01-09 13:07:32 +08:00 |
|
Wang, Jian4
|
a54cd767b1
|
LLM: Add gguf falcon (#9801)
* init falcon
* update convert.py
* update style
|
2024-01-03 14:49:02 +08:00 |
|
Wang, Jian4
|
7ed9538b9f
|
LLM: support gguf mpt (#9773)
* add gguf mpt
* update
|
2023-12-28 09:22:39 +08:00 |
|
Heyang Sun
|
66e286a73d
|
Support for Mixtral AWQ (#9775)
* Support for Mixtral AWQ
* Update README.md
* Update README.md
* Update awq_config.py
* Update README.md
* Update README.md
|
2023-12-25 16:08:09 +08:00 |
|
Wang, Jian4
|
984697afe2
|
LLM: Add bloom gguf support (#9734)
* init
* update bloom add merges
* update
* update readme
* update for llama error
* update
|
2023-12-21 14:06:25 +08:00 |
|
Heyang Sun
|
1fa7793fc0
|
Load Mixtral GGUF Model (#9690)
* Load Mixtral GGUF Model
* refactor
* fix empty tensor when to cpu
* update gpu and cpu readmes
* add dtype when set tensor into module
|
2023-12-19 13:54:38 +08:00 |
|
Wang, Jian4
|
b8437a1c1e
|
LLM: Add gguf mistral model support (#9691)
* add mistral support
* need to upgrade transformers version
* update
|
2023-12-15 13:37:39 +08:00 |
|
Wang, Jian4
|
496bb2e845
|
LLM: Support load BaiChuan model family gguf model (#9685)
* support baichuan model family gguf model
* update gguf generate.py
* add verify models
* add support model_family
* update
* update style
* update type
* update readme
* update
* remove support model_family
|
2023-12-15 13:34:33 +08:00 |
|
ZehuaCao
|
877229f3be
|
[LLM]Add Yi-34B-AWQ to verified AWQ model. (#9676)
* verfiy Yi-34B-AWQ
* update
|
2023-12-14 09:55:47 +08:00 |
|
ZehuaCao
|
503880809c
|
verfiy codeLlama (#9668)
|
2023-12-13 15:39:31 +08:00 |
|
ZehuaCao
|
45721f3473
|
verfiy llava (#9649)
|
2023-12-11 14:26:05 +08:00 |
|
Heyang Sun
|
9f02f96160
|
[LLM] support for Yi AWQ model (#9648)
|
2023-12-11 14:07:34 +08:00 |
|
Heyang Sun
|
3811cf43c9
|
[LLM] update AWQ documents (#9623)
* [LLM] update AWQ and verified models' documents
* refine
* refine links
* refine
|
2023-12-07 16:02:20 +08:00 |
|
Jason Dai
|
51b668f229
|
Update GGUF readme (#9611)
|
2023-12-06 18:21:54 +08:00 |
|
dingbaorong
|
a7bc89b3a1
|
remove q4_1 in gguf example (#9610)
* remove q4_1
* fixes
|
2023-12-06 16:00:05 +08:00 |
|
dingbaorong
|
89069d6173
|
Add gpu gguf example (#9603)
* add gpu gguf example
* some fixes
* address kai's comments
* address json's comments
|
2023-12-06 15:17:54 +08:00 |
|
Qiyuan Gong
|
d85a430a8c
|
Uing bigdl-llm-init instead of bigdl-nano-init (#9558)
* Replace `bigdl-nano-init` with `bigdl-llm-init`.
* Install `bigdl-llm` instead of `bigdl-nano`.
* Remove nano in README.
|
2023-11-30 10:10:29 +08:00 |
|
binbin Deng
|
6bec0faea5
|
LLM: support Mistral AWQ models (#9520)
|
2023-11-24 16:20:22 +08:00 |
|
Yina Chen
|
d5263e6681
|
Add awq load support (#9453)
* Support directly loading GPTQ models from huggingface
* fix style
* fix tests
* change example structure
* address comments
* fix style
* init
* address comments
* add examples
* fix style
* fix style
* fix style
* fix style
* update
* remove
* meet comments
* fix style
---------
Co-authored-by: Yang Wang <yang3.wang@intel.com>
|
2023-11-16 14:06:25 +08:00 |
|
Yang Wang
|
51d07a9fd8
|
Support directly loading gptq models from huggingface (#9391)
* Support directly loading GPTQ models from huggingface
* fix style
* fix tests
* change example structure
* address comments
* fix style
* address comments
|
2023-11-13 20:48:12 -08:00 |
|