WebUI quickstart: add instruct chat mode and tested models (#10436)
* Add instruct chat mode and tested models * Fix table * Remove falcon from 'tested models' * Fixes * Open image in new window
This commit is contained in:
		
							parent
							
								
									92ee2077b3
								
							
						
					
					
						commit
						a7d38bee94
					
				
					 1 changed files with 42 additions and 6 deletions
				
			
		| 
						 | 
				
			
			@ -6,7 +6,7 @@ This quickstart guide walks you through setting up and using the [Text Generatio
 | 
			
		|||
 | 
			
		||||
A preview of the WebUI in action is shown below:
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat.png">
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -71,7 +71,7 @@ In **Anaconda Prompt** with the conda environment `llm` activated, navigate to t
 | 
			
		|||
### Access the WebUI
 | 
			
		||||
Upon successful launch, URLs to access the WebUI will be displayed in the terminal as shown below. Open the provided local URL in your browser to interact with the WebUI. 
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_launch_server.png">
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_launch_server.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_launch_server.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -81,13 +81,13 @@ Upon successful launch, URLs to access the WebUI will be displayed in the termin
 | 
			
		|||
 | 
			
		||||
Place Huggingface models in `C:\text-generation-webui\models` by either copying locally or downloading via the WebUI. To download, navigate to the **Model** tab, enter the model's huggingface id (for instance, `microsoft/phi-1_5`) in the **Download model or LoRA** section, and click **Download**, as illustrated below. 
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_download_model.png">
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_download_model.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_download_model.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
After copying or downloading the models, click on the blue **refresh** button to update the **Model** drop-down menu. Then, choose your desired model from the newly updated list.  
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_select_model.png">
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_select_model.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_select_model.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -97,7 +97,7 @@ Default settings are recommended for most users. Click **Load** to activate the
 | 
			
		|||
 | 
			
		||||
If everything goes well, you will get a message as shown below.
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_load_model_success.png">
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_load_model_success.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_load_model_success.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -107,7 +107,7 @@ In the **Chat** tab, start new conversations with **New chat**.
 | 
			
		|||
 | 
			
		||||
Enter prompts into the textbox at the bottom and press the **Generate** button to receive responses.
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat.png">
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -123,6 +123,42 @@ Enter prompts into the textbox at the bottom and press the **Generate** button t
 | 
			
		|||
To shut down the WebUI server, use **Ctrl+C** in the **Anaconda Prompt** terminal where the WebUI Server is runing, then close your browser tab.
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
## 5. Advanced Usage
 | 
			
		||||
### Using Instruct mode
 | 
			
		||||
Instruction-following models are models that are fine-tuned with specific prompt formats. 
 | 
			
		||||
For these models, you should ideally use the `instruct` chat mode.
 | 
			
		||||
Under this mode, the model receives user prompts that are formatted according to prompt formats it was trained with.
 | 
			
		||||
 | 
			
		||||
To use `instruct` chat mode, select `chat` tab, scroll down the page, and then select `instruct` under `Mode`.
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat_mode_instruct.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_chat_mode_instruct.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
When a model is loaded, its corresponding instruction template, which contains prompt formatting, is automatically loaded.
 | 
			
		||||
If chat responses are poor, the loaded instruction template might be incorrect.
 | 
			
		||||
In this case, go to `Parameters` tab and then `Instruction template` tab.
 | 
			
		||||
 | 
			
		||||
<a href="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_instruction_template.png" target="_blank">
 | 
			
		||||
  <img src="https://llm-assets.readthedocs.io/en/latest/_images/webui_quickstart_instruction_template.png" width=100%; />
 | 
			
		||||
</a>
 | 
			
		||||
 | 
			
		||||
You can verify and edit the loaded instruction template in the `Instruction template` field.
 | 
			
		||||
You can also manually select an instruction template from `Saved instruction templates` and click `load` to load it into `Instruction template`.
 | 
			
		||||
You can add custom template files to this list in `/instruction-templates/` [folder](https://github.com/intel-analytics/text-generation-webui/tree/bigdl-llm/instruction-templates).
 | 
			
		||||
<!-- For instance, the automatically loaded instruction template for `chatGLM3` model is incorrect, and you should manually select the `chatGLM3` instruction template. -->
 | 
			
		||||
 | 
			
		||||
### Tested models
 | 
			
		||||
We have tested the following models with `bigdl-llm` using Text Generation WebUI.
 | 
			
		||||
 | 
			
		||||
| Model | Notes |
 | 
			
		||||
|-------|-------|
 | 
			
		||||
| llama-2-7b-chat-hf |          |
 | 
			
		||||
| chatglm3-6b        | Manually load ChatGLM3 template for Instruct chat mode |
 | 
			
		||||
| Mistral-7B-v0.1    |          |
 | 
			
		||||
| qwen-7B-Chat       |          |
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
## Troubleshooting
 | 
			
		||||
 | 
			
		||||
### Potentially slower first response
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue