LLM: update langchain api document page (#9222)
This commit is contained in:
parent
0c5055d38c
commit
db37edae8a
7 changed files with 146 additions and 123 deletions
|
|
@ -25,7 +25,7 @@ output = tokenizer.batch_decode(output_ids)
|
|||
```eval_rst
|
||||
.. seealso::
|
||||
|
||||
See the complete examples `here <https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/transformers/transformers_int4>`_
|
||||
See the complete CPU examples `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/HF-Transformers-AutoModels>`_ and GPU examples `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/GPU/HF-Transformers-AutoModels>`_.
|
||||
|
||||
.. note::
|
||||
|
||||
|
|
@ -35,7 +35,7 @@ output = tokenizer.batch_decode(output_ids)
|
|||
|
||||
model = AutoModelForCausalLM.from_pretrained('/path/to/model/', load_in_low_bit="sym_int5")
|
||||
|
||||
See the complete example `here <https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/transformers/transformers_low_bit>`_.
|
||||
See the CPU example `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/HF-Transformers-AutoModels/More-Data-Types>`_ and GPU example `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/GPU/HF-Transformers-AutoModels/More-Data-Types>`_.
|
||||
```
|
||||
|
||||
## Save & Load
|
||||
|
|
@ -50,5 +50,5 @@ new_model = AutoModelForCausalLM.load_low_bit(model_path)
|
|||
```eval_rst
|
||||
.. seealso::
|
||||
|
||||
See the examples `here <https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/transformers/transformers_low_bit>`_
|
||||
See the CPU example `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/HF-Transformers-AutoModels/Save-Load>`_ and GPU example `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/GPU/HF-Transformers-AutoModels/Save-Load>`_
|
||||
```
|
||||
|
|
@ -21,7 +21,7 @@ output = doc_chain.run(...)
|
|||
```eval_rst
|
||||
.. seealso::
|
||||
|
||||
See the examples `here <https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/langchain/transformers_int4>`_.
|
||||
See the examples `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/LangChain/transformers_int4>`_.
|
||||
```
|
||||
|
||||
## Using Native INT4 Format
|
||||
|
|
@ -53,5 +53,5 @@ doc_chain.run(...)
|
|||
```eval_rst
|
||||
.. seealso::
|
||||
|
||||
See the examples `here <https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/langchain/native_int4>`_.
|
||||
See the examples `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/LangChain/native_int4>`_.
|
||||
```
|
||||
|
|
|
|||
|
|
@ -28,5 +28,5 @@ output = llm.batch_decode(output_ids)
|
|||
```eval_rst
|
||||
.. seealso::
|
||||
|
||||
See the complete example `here <https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/transformers/native_int4/native_int4_pipeline.py>`_
|
||||
See the complete example `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/Native-Models>`_
|
||||
```
|
||||
|
|
@ -1,144 +1,164 @@
|
|||
BigDL-LLM LangChain API
|
||||
=====================
|
||||
|
||||
llm.langchain.embeddings.bigdlllm
|
||||
LLM Wrapper of LangChain
|
||||
----------------------------------------
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.LlamaEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
Hugging Face ``transformers`` Format
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
BigDL-LLM provides ``TransformersLLM`` and ``TransformersPipelineLLM``, which implement the standard interface of LLM wrapper of LangChain.
|
||||
|
||||
.. tabs::
|
||||
|
||||
.. tab:: AutoModel
|
||||
|
||||
.. automodule:: bigdl.llm.langchain.llms.transformersllm
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: model_id, model_kwargs, model, tokenizer, streaming, Config
|
||||
|
||||
.. tab:: pipeline
|
||||
|
||||
.. automodule:: bigdl.llm.langchain.llms.transformerspipelinellm
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: pipeline, model_id, model_kwargs, pipeline_kwargs, Config
|
||||
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.BloomEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
Native Model
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
For ``llama``/``chatglm``/``bloom``/``gptneox``/``starcoder`` model families, you could also use the following LLM wrappers with the native (cpp) implementation for maximum performance.
|
||||
|
||||
.. tabs::
|
||||
|
||||
.. tab:: Llama
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.LlamaLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
|
||||
.. tab:: ChatGLM
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.ChatGLMLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
|
||||
.. tab:: Bloom
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.BloomLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
|
||||
.. tab:: Gptneox
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.GptneoxLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
|
||||
.. tab:: Starcoder
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.StarcoderLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.GptneoxEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
Embeddings Wrapper of LangChain
|
||||
----------------------------------------
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.ChatGLMEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.StarcoderEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
|
||||
|
||||
llm.langchain.embeddings.transformersembeddings
|
||||
--------------------------------------------------
|
||||
Hugging Face ``transformers`` AutoModel
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
.. automodule:: bigdl.llm.langchain.embeddings.transformersembeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: model, tokenizer, model_id, model_kwargs, encode_kwargs, Config
|
||||
|
||||
Native Model
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
llm.langchain.llms.bigdlllm
|
||||
----------------------------------------
|
||||
For ``llama``/``bloom``/``gptneox``/``starcoder`` model families, you could also use the following wrappers.
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.LlamaLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
.. tabs::
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
.. tab:: Llama
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.LlamaEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.BloomLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
.. tab:: Bloom
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.BloomEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.GptneoxLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
.. tab:: Gptneox
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.GptneoxEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.ChatGLMLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
.. tab:: Starcoder
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.embeddings.bigdlllm.StarcoderEmbeddings
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model, ggml_module, client, model_path, kwargs
|
||||
|
||||
.. autoclass:: bigdl.llm.langchain.llms.bigdlllm.StarcoderLLM
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
:exclude-members: ggml_model , ggml_module
|
||||
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: stream
|
||||
.. automethod:: get_num_tokens
|
||||
|
||||
|
||||
llm.langchain.llms.transformersllm
|
||||
----------------------------------------
|
||||
|
||||
.. automodule:: bigdl.llm.langchain.llms.transformersllm
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
|
||||
|
||||
llm.langchain.llms.transformerspipelinellm
|
||||
---------------------------------------------
|
||||
|
||||
.. automodule:: bigdl.llm.langchain.llms.transformerspipelinellm
|
||||
:members:
|
||||
:undoc-members:
|
||||
:show-inheritance:
|
||||
.. automethod:: validate_environment
|
||||
.. automethod:: embed_documents
|
||||
.. automethod:: embed_query
|
||||
|
|
|
|||
|
|
@ -24,8 +24,8 @@ BigDL-LLM: low-Bit LLM library
|
|||
============================================
|
||||
Latest update
|
||||
============================================
|
||||
- **[New]** ``bigdl-llm`` now supports QLoRA fintuning on Intel GPU; see the the example `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/gpu/qlora_finetuning>`_.
|
||||
- ``bigdl-llm`` now supports Intel GPU (including Arc, Flex and MAX); see the the latest GPU examples `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/gpu>`_.
|
||||
- **[New]** ``bigdl-llm`` now supports QLoRA fintuning on Intel GPU; see the the example `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/GPU/QLoRA-FineTuning>`_.
|
||||
- ``bigdl-llm`` now supports Intel GPU (including Arc, Flex and MAX); see the the latest GPU examples `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/GPU>`_.
|
||||
- ``bigdl-llm`` tutorial is released `here <https://github.com/intel-analytics/bigdl-llm-tutorial>`_.
|
||||
- Over 20 models have been verified on ``bigdl-llm``, including *LLaMA/LLaMA2, ChatGLM/ChatGLM2, MPT, Falcon, Dolly, StarCoder, Whisper, InternLM, QWen, Baichuan, Aquila, MOSS* and more; see the complete list `here <https://github.com/intel-analytics/bigdl#verified-models>`_.
|
||||
|
||||
|
|
|
|||
|
|
@ -95,7 +95,8 @@ class TransformersEmbeddings(BaseModel, Embeddings):
|
|||
model_kwargs: Keyword arguments that will be passed to the model and tokenizer.
|
||||
kwargs: Extra arguments that will be passed to the model and tokenizer.
|
||||
|
||||
Returns: An object of TransformersEmbeddings.
|
||||
Returns:
|
||||
An object of TransformersEmbeddings.
|
||||
"""
|
||||
try:
|
||||
from bigdl.llm.transformers import AutoModel
|
||||
|
|
|
|||
|
|
@ -101,7 +101,8 @@ class TransformersLLM(LLM):
|
|||
model_kwargs: Keyword arguments that will be passed to the model and tokenizer.
|
||||
kwargs: Extra arguments that will be passed to the model and tokenizer.
|
||||
|
||||
Returns: An object of TransformersLLM.
|
||||
Returns:
|
||||
An object of TransformersLLM.
|
||||
"""
|
||||
try:
|
||||
from bigdl.llm.transformers import (
|
||||
|
|
@ -159,7 +160,8 @@ class TransformersLLM(LLM):
|
|||
model_kwargs: Keyword arguments that will be passed to the model and tokenizer.
|
||||
kwargs: Extra arguments that will be passed to the model and tokenizer.
|
||||
|
||||
Returns: An object of TransformersLLM.
|
||||
Returns:
|
||||
An object of TransformersLLM.
|
||||
"""
|
||||
try:
|
||||
from bigdl.llm.transformers import (
|
||||
|
|
|
|||
Loading…
Reference in a new issue