LLM: improve transformers style API doc (#9113)
This commit is contained in:
parent
65212451cc
commit
e4d1457a70
3 changed files with 66 additions and 43 deletions
|
|
@ -1,7 +1,7 @@
|
||||||
BigDL-LLM PyTorch API
|
BigDL-LLM PyTorch API
|
||||||
=====================
|
=====================
|
||||||
|
|
||||||
llm.optimize
|
optimize model
|
||||||
----------------------------------------
|
----------------------------------------
|
||||||
|
|
||||||
.. automodule:: bigdl.llm.optimize
|
.. automodule:: bigdl.llm.optimize
|
||||||
|
|
|
||||||
|
|
@ -1,10 +1,16 @@
|
||||||
BigDL-LLM `transformers`-style API
|
BigDL-LLM ``transformers``-style API
|
||||||
=====================
|
====================================
|
||||||
|
|
||||||
llm.transformers.model
|
Hugging Face ``transformers`` AutoModel
|
||||||
---------------------------
|
------------------------------------
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.model.AutoModelForCausalLM
|
You can apply BigDL-LLM optimizations on any Hugging Face Transformers models by using the standard AutoModel APIs.
|
||||||
|
|
||||||
|
|
||||||
|
AutoModelForCausalLM
|
||||||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
.. autoclass:: bigdl.llm.transformers.AutoModelForCausalLM
|
||||||
:members:
|
:members:
|
||||||
:undoc-members:
|
:undoc-members:
|
||||||
:show-inheritance:
|
:show-inheritance:
|
||||||
|
|
@ -13,8 +19,10 @@ llm.transformers.model
|
||||||
.. automethod:: load_convert
|
.. automethod:: load_convert
|
||||||
.. automethod:: load_low_bit
|
.. automethod:: load_low_bit
|
||||||
|
|
||||||
|
AutoModel
|
||||||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.model.AutoModel
|
.. autoclass:: bigdl.llm.transformers.AutoModel
|
||||||
:members:
|
:members:
|
||||||
:undoc-members:
|
:undoc-members:
|
||||||
:show-inheritance:
|
:show-inheritance:
|
||||||
|
|
@ -23,8 +31,10 @@ llm.transformers.model
|
||||||
.. automethod:: load_convert
|
.. automethod:: load_convert
|
||||||
.. automethod:: load_low_bit
|
.. automethod:: load_low_bit
|
||||||
|
|
||||||
|
AutoModelForSpeechSeq2Seq
|
||||||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.model.AutoModelForSpeechSeq2Seq
|
.. autoclass:: bigdl.llm.transformers.AutoModelForSpeechSeq2Seq
|
||||||
:members:
|
:members:
|
||||||
:undoc-members:
|
:undoc-members:
|
||||||
:show-inheritance:
|
:show-inheritance:
|
||||||
|
|
@ -33,8 +43,10 @@ llm.transformers.model
|
||||||
.. automethod:: load_convert
|
.. automethod:: load_convert
|
||||||
.. automethod:: load_low_bit
|
.. automethod:: load_low_bit
|
||||||
|
|
||||||
|
AutoModelForSeq2SeqLM
|
||||||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.model.AutoModelForSeq2SeqLM
|
.. autoclass:: bigdl.llm.transformers.AutoModelForSeq2SeqLM
|
||||||
:members:
|
:members:
|
||||||
:undoc-members:
|
:undoc-members:
|
||||||
:show-inheritance:
|
:show-inheritance:
|
||||||
|
|
@ -45,48 +57,59 @@ llm.transformers.model
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
llm.transformers.modelling_bigdl
|
Native Model
|
||||||
----------------------------------------
|
----------------------------------------
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.LlamaForCausalLM
|
For ``llama``/``chatglm``/``bloom``/``gptneox``/``starcoder`` model families, you may also convert and run LLM using the native (cpp) implementation for maximum performance.
|
||||||
:members:
|
|
||||||
:undoc-members:
|
|
||||||
:show-inheritance:
|
|
||||||
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
|
||||||
|
|
||||||
.. automethod:: from_pretrained
|
|
||||||
|
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.ChatGLMForCausalLM
|
.. tabs::
|
||||||
:members:
|
|
||||||
:undoc-members:
|
|
||||||
:show-inheritance:
|
|
||||||
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
|
||||||
|
|
||||||
.. automethod:: from_pretrained
|
.. tab:: Llama
|
||||||
|
|
||||||
|
.. autoclass:: bigdl.llm.transformers.LlamaForCausalLM
|
||||||
|
:members:
|
||||||
|
:undoc-members:
|
||||||
|
:show-inheritance:
|
||||||
|
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.GptneoxForCausalLM
|
.. automethod:: from_pretrained
|
||||||
:members:
|
|
||||||
:undoc-members:
|
|
||||||
:show-inheritance:
|
|
||||||
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
|
||||||
|
|
||||||
.. automethod:: from_pretrained
|
.. tab:: ChatGLM
|
||||||
|
|
||||||
|
.. autoclass:: bigdl.llm.transformers.ChatGLMForCausalLM
|
||||||
|
:members:
|
||||||
|
:undoc-members:
|
||||||
|
:show-inheritance:
|
||||||
|
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.BloomForCausalLM
|
.. automethod:: from_pretrained
|
||||||
:members:
|
|
||||||
:undoc-members:
|
|
||||||
:show-inheritance:
|
|
||||||
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
|
||||||
|
|
||||||
.. automethod:: from_pretrained
|
.. tab:: Gptneox
|
||||||
|
|
||||||
.. autoclass:: bigdl.llm.transformers.modelling_bigdl.StarcoderForCausalLM
|
.. autoclass:: bigdl.llm.transformers.GptneoxForCausalLM
|
||||||
:members:
|
:members:
|
||||||
:undoc-members:
|
:undoc-members:
|
||||||
:show-inheritance:
|
:show-inheritance:
|
||||||
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
||||||
|
|
||||||
.. automethod:: from_pretrained
|
.. automethod:: from_pretrained
|
||||||
|
|
||||||
|
.. tab:: Bloom
|
||||||
|
.. autoclass:: bigdl.llm.transformers.BloomForCausalLM
|
||||||
|
:members:
|
||||||
|
:undoc-members:
|
||||||
|
:show-inheritance:
|
||||||
|
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
||||||
|
|
||||||
|
.. automethod:: from_pretrained
|
||||||
|
|
||||||
|
.. tab:: Starcoder
|
||||||
|
|
||||||
|
.. autoclass:: bigdl.llm.transformers.StarcoderForCausalLM
|
||||||
|
:members:
|
||||||
|
:undoc-members:
|
||||||
|
:show-inheritance:
|
||||||
|
:exclude-members: GGML_Model, GGML_Module, HF_Class
|
||||||
|
|
||||||
|
.. automethod:: from_pretrained
|
||||||
|
|
|
||||||
|
|
@ -46,7 +46,7 @@ from torch.serialization import StorageType
|
||||||
import pickle
|
import pickle
|
||||||
import zipfile
|
import zipfile
|
||||||
import io
|
import io
|
||||||
from typing import Dict, IO, Any, Callable
|
from typing import Dict, IO, Any, Callable, List
|
||||||
from dataclasses import dataclass
|
from dataclasses import dataclass
|
||||||
from .common import invalidInputError
|
from .common import invalidInputError
|
||||||
|
|
||||||
|
|
@ -69,7 +69,7 @@ class LazyStorage:
|
||||||
@dataclass
|
@dataclass
|
||||||
class LazyTensor:
|
class LazyTensor:
|
||||||
_load: Callable[[], torch.Tensor]
|
_load: Callable[[], torch.Tensor]
|
||||||
shape: list[int]
|
shape: List[int]
|
||||||
data_type: torch.dtype
|
data_type: torch.dtype
|
||||||
description: str
|
description: str
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue