LLM: improve transformers style API doc (#9113)

This commit is contained in:
binbin Deng 2023-10-10 09:31:00 +08:00 committed by GitHub
parent 65212451cc
commit e4d1457a70
3 changed files with 66 additions and 43 deletions

View file

@ -1,7 +1,7 @@
BigDL-LLM PyTorch API BigDL-LLM PyTorch API
===================== =====================
llm.optimize optimize model
---------------------------------------- ----------------------------------------
.. automodule:: bigdl.llm.optimize .. automodule:: bigdl.llm.optimize

View file

@ -1,10 +1,16 @@
BigDL-LLM `transformers`-style API BigDL-LLM ``transformers``-style API
===================== ====================================
llm.transformers.model Hugging Face ``transformers`` AutoModel
--------------------------- ------------------------------------
.. autoclass:: bigdl.llm.transformers.model.AutoModelForCausalLM You can apply BigDL-LLM optimizations on any Hugging Face Transformers models by using the standard AutoModel APIs.
AutoModelForCausalLM
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. autoclass:: bigdl.llm.transformers.AutoModelForCausalLM
:members: :members:
:undoc-members: :undoc-members:
:show-inheritance: :show-inheritance:
@ -13,8 +19,10 @@ llm.transformers.model
.. automethod:: load_convert .. automethod:: load_convert
.. automethod:: load_low_bit .. automethod:: load_low_bit
AutoModel
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. autoclass:: bigdl.llm.transformers.model.AutoModel .. autoclass:: bigdl.llm.transformers.AutoModel
:members: :members:
:undoc-members: :undoc-members:
:show-inheritance: :show-inheritance:
@ -23,8 +31,10 @@ llm.transformers.model
.. automethod:: load_convert .. automethod:: load_convert
.. automethod:: load_low_bit .. automethod:: load_low_bit
AutoModelForSpeechSeq2Seq
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. autoclass:: bigdl.llm.transformers.model.AutoModelForSpeechSeq2Seq .. autoclass:: bigdl.llm.transformers.AutoModelForSpeechSeq2Seq
:members: :members:
:undoc-members: :undoc-members:
:show-inheritance: :show-inheritance:
@ -33,8 +43,10 @@ llm.transformers.model
.. automethod:: load_convert .. automethod:: load_convert
.. automethod:: load_low_bit .. automethod:: load_low_bit
AutoModelForSeq2SeqLM
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
.. autoclass:: bigdl.llm.transformers.model.AutoModelForSeq2SeqLM .. autoclass:: bigdl.llm.transformers.AutoModelForSeq2SeqLM
:members: :members:
:undoc-members: :undoc-members:
:show-inheritance: :show-inheritance:
@ -45,48 +57,59 @@ llm.transformers.model
llm.transformers.modelling_bigdl Native Model
---------------------------------------- ----------------------------------------
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.LlamaForCausalLM For ``llama``/``chatglm``/``bloom``/``gptneox``/``starcoder`` model families, you may also convert and run LLM using the native (cpp) implementation for maximum performance.
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. automethod:: from_pretrained
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.ChatGLMForCausalLM .. tabs::
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. automethod:: from_pretrained .. tab:: Llama
.. autoclass:: bigdl.llm.transformers.LlamaForCausalLM
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.GptneoxForCausalLM .. automethod:: from_pretrained
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. automethod:: from_pretrained .. tab:: ChatGLM
.. autoclass:: bigdl.llm.transformers.ChatGLMForCausalLM
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.BloomForCausalLM .. automethod:: from_pretrained
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. automethod:: from_pretrained .. tab:: Gptneox
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.StarcoderForCausalLM .. autoclass:: bigdl.llm.transformers.GptneoxForCausalLM
:members: :members:
:undoc-members: :undoc-members:
:show-inheritance: :show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class :exclude-members: GGML_Model, GGML_Module, HF_Class
.. automethod:: from_pretrained .. automethod:: from_pretrained
.. tab:: Bloom
.. autoclass:: bigdl.llm.transformers.BloomForCausalLM
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. automethod:: from_pretrained
.. tab:: Starcoder
.. autoclass:: bigdl.llm.transformers.StarcoderForCausalLM
:members:
:undoc-members:
:show-inheritance:
:exclude-members: GGML_Model, GGML_Module, HF_Class
.. automethod:: from_pretrained

View file

@ -46,7 +46,7 @@ from torch.serialization import StorageType
import pickle import pickle
import zipfile import zipfile
import io import io
from typing import Dict, IO, Any, Callable from typing import Dict, IO, Any, Callable, List
from dataclasses import dataclass from dataclasses import dataclass
from .common import invalidInputError from .common import invalidInputError
@ -69,7 +69,7 @@ class LazyStorage:
@dataclass @dataclass
class LazyTensor: class LazyTensor:
_load: Callable[[], torch.Tensor] _load: Callable[[], torch.Tensor]
shape: list[int] shape: List[int]
data_type: torch.dtype data_type: torch.dtype
description: str description: str