[LLM] Correct prompt format of Qwen in generate.py (#9678)
* Change qwen prompt format to chatml
This commit is contained in:
parent
223c9622f7
commit
21c7503a42
1 changed files with 9 additions and 1 deletions
|
|
@ -23,7 +23,15 @@ from bigdl.llm.transformers import AutoModelForCausalLM
|
||||||
from transformers import AutoTokenizer
|
from transformers import AutoTokenizer
|
||||||
|
|
||||||
# you could tune the prompt based on your own model
|
# you could tune the prompt based on your own model
|
||||||
QWEN_PROMPT_FORMAT = "<human>{prompt} <bot>"
|
QWEN_PROMPT_FORMAT = """
|
||||||
|
<|im_start|>system
|
||||||
|
You are a helpful assistant.
|
||||||
|
<|im_end|>
|
||||||
|
<|im_start|>user
|
||||||
|
{prompt}
|
||||||
|
<|im_end|>
|
||||||
|
<|im_start|>assistant
|
||||||
|
"""
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
parser = argparse.ArgumentParser(description='Predict Tokens using `generate()` API for Qwen model')
|
parser = argparse.ArgumentParser(description='Predict Tokens using `generate()` API for Qwen model')
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue