Nano: fix onnx quantization document issue (#6662)
This commit is contained in:
		
							parent
							
								
									8d6c43dd09
								
							
						
					
					
						commit
						62694b420e
					
				
					 1 changed files with 1 additions and 1 deletions
				
			
		| 
						 | 
					@ -16,7 +16,7 @@ source bigdl-nano-init
 | 
				
			||||||
 | 
					
 | 
				
			||||||
To quantize model using ONNXRuntime as backend, it is required to install Intel Neural Compressor, onnxruntime-extensions as a dependency of INC and some onnx packages as below
 | 
					To quantize model using ONNXRuntime as backend, it is required to install Intel Neural Compressor, onnxruntime-extensions as a dependency of INC and some onnx packages as below
 | 
				
			||||||
```python
 | 
					```python
 | 
				
			||||||
pip install neural-compress==1.11
 | 
					pip install neural-compressor==1.11
 | 
				
			||||||
pip install onnx onnxruntime onnxruntime-extensions
 | 
					pip install onnx onnxruntime onnxruntime-extensions
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
### Step 1: Load the data
 | 
					### Step 1: Load the data
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue