Nano: quick fix typo (#7064)
This commit is contained in:
		
							parent
							
								
									dea7b8f9f2
								
							
						
					
					
						commit
						0a62df22aa
					
				
					 1 changed files with 4 additions and 4 deletions
				
			
		| 
						 | 
					@ -35,7 +35,7 @@ For example,
 | 
				
			||||||
from bigdl.nano.pytorch import Trainer
 | 
					from bigdl.nano.pytorch import Trainer
 | 
				
			||||||
 | 
					
 | 
				
			||||||
lightning_module = Trainer.compile(pytorch_module, loss, optimizer)
 | 
					lightning_module = Trainer.compile(pytorch_module, loss, optimizer)
 | 
				
			||||||
trainer = Trainer(max_epoch=10)
 | 
					trainer = Trainer(max_epochs=10)
 | 
				
			||||||
trainer.fit(lightning_module, train_loader)
 | 
					trainer.fit(lightning_module, train_loader)
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -69,7 +69,7 @@ class MyModule(LightningModule):
 | 
				
			||||||
    #  LightningModule definition
 | 
					    #  LightningModule definition
 | 
				
			||||||
 | 
					
 | 
				
			||||||
lightning_module = MyModule()
 | 
					lightning_module = MyModule()
 | 
				
			||||||
trainer = Trainer(max_epoch=10)
 | 
					trainer = Trainer(max_epochs=10)
 | 
				
			||||||
trainer.fit(lightning_module, train_loader)
 | 
					trainer.fit(lightning_module, train_loader)
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -90,7 +90,7 @@ BigDL-Nano integrates IPEX in `Trainer` and `TorchNano`. Users can turn on IPEX
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            from bigdl.nano.pytorch import Trainer
 | 
					            from bigdl.nano.pytorch import Trainer
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            trainer = Trainer(max_epoch=10, use_ipex=True)
 | 
					            trainer = Trainer(max_epochs=10, use_ipex=True)
 | 
				
			||||||
            trainer.fit(...)
 | 
					            trainer.fit(...)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    .. tab:: TorchNano
 | 
					    .. tab:: TorchNano
 | 
				
			||||||
| 
						 | 
					@ -123,7 +123,7 @@ You can just set the `num_processes` parameter in the `Trainer` or `TorchNano` c
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            from bigdl.nano.pytorch import Trainer
 | 
					            from bigdl.nano.pytorch import Trainer
 | 
				
			||||||
 | 
					
 | 
				
			||||||
            trainer = Trainer(max_epoch=10, num_processes=4)
 | 
					            trainer = Trainer(max_epochs=10, num_processes=4)
 | 
				
			||||||
            trainer.fit(...)
 | 
					            trainer.fit(...)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    .. tab:: TorchNano
 | 
					    .. tab:: TorchNano
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue