[LLM] Small fix to MPT Example (#8513)
This commit is contained in:
parent
ba0da17b40
commit
bcde8ec83e
2 changed files with 4 additions and 4 deletions
|
|
@ -16,7 +16,7 @@ conda create -n llm python=3.9
|
|||
conda activate llm
|
||||
|
||||
pip install bigdl-llm[all] # install bigdl-llm with 'all' option
|
||||
pip instll einops # additional package required for mpt-7b-chat to conduct generation
|
||||
pip install einops # additional package required for mpt-7b-chat to conduct generation
|
||||
```
|
||||
|
||||
### 2. Config
|
||||
|
|
|
|||
|
|
@ -27,9 +27,9 @@ MPT_PROMPT_FORMAT="<human>{prompt} <bot>"
|
|||
if __name__ == '__main__':
|
||||
parser = argparse.ArgumentParser(description='Transformer INT4 example for MPT model')
|
||||
parser.add_argument('--repo-id-or-model-path', type=str, default="mosaicml/mpt-7b-chat",
|
||||
help='The huggingface repo id for the large language model to be downloaded'
|
||||
help='The huggingface repo id for the MPT to be downloaded'
|
||||
', or the path to the huggingface checkpoint folder')
|
||||
parser.add_argument('--prompt', type=str, default="<human>What is AI?<bot>",
|
||||
parser.add_argument('--prompt', type=str, default="What is AI?",
|
||||
help='Prompt to infer')
|
||||
parser.add_argument('--n-predict', type=int, default=32,
|
||||
help='Max tokens to predict')
|
||||
|
|
|
|||
Loading…
Reference in a new issue