From bcde8ec83e11c6c20435556bec8dd92376c97232 Mon Sep 17 00:00:00 2001 From: Yuwen Hu <54161268+Oscilloscope98@users.noreply.github.com> Date: Thu, 13 Jul 2023 14:33:21 +0800 Subject: [PATCH] [LLM] Small fix to MPT Example (#8513) --- .../example/transformers/transformers_int4/mpt/README.md | 2 +- .../example/transformers/transformers_int4/mpt/generate.py | 6 +++--- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/python/llm/example/transformers/transformers_int4/mpt/README.md b/python/llm/example/transformers/transformers_int4/mpt/README.md index 01389f09..e8a3f917 100644 --- a/python/llm/example/transformers/transformers_int4/mpt/README.md +++ b/python/llm/example/transformers/transformers_int4/mpt/README.md @@ -16,7 +16,7 @@ conda create -n llm python=3.9 conda activate llm pip install bigdl-llm[all] # install bigdl-llm with 'all' option -pip instll einops # additional package required for mpt-7b-chat to conduct generation +pip install einops # additional package required for mpt-7b-chat to conduct generation ``` ### 2. Config diff --git a/python/llm/example/transformers/transformers_int4/mpt/generate.py b/python/llm/example/transformers/transformers_int4/mpt/generate.py index f8c62dd2..7e543219 100644 --- a/python/llm/example/transformers/transformers_int4/mpt/generate.py +++ b/python/llm/example/transformers/transformers_int4/mpt/generate.py @@ -22,14 +22,14 @@ from bigdl.llm.transformers import AutoModelForCausalLM from transformers import AutoTokenizer # you could revise it based on the MPT model you choose to use -MPT_PROMPT_FORMAT="{prompt} " +MPT_PROMPT_FORMAT = "{prompt} " if __name__ == '__main__': parser = argparse.ArgumentParser(description='Transformer INT4 example for MPT model') parser.add_argument('--repo-id-or-model-path', type=str, default="mosaicml/mpt-7b-chat", - help='The huggingface repo id for the large language model to be downloaded' + help='The huggingface repo id for the MPT to be downloaded' ', or the path to the huggingface checkpoint folder') - parser.add_argument('--prompt', type=str, default="What is AI?", + parser.add_argument('--prompt', type=str, default="What is AI?", help='Prompt to infer') parser.add_argument('--n-predict', type=int, default=32, help='Max tokens to predict')