model.eval() is a kind of switch for some specific layers/parts of the model that behave differently during training and inference (evaluating) time. For example, Dropouts Layers, BatchNorm Layers etc. You need to turn off them during model evaluation, and .eval() will do it for you. In addition, the common practice for evaluating/validation is using torch.no_grad() in pair with model.eval() to turn off gradients computation:
This has any effect only on certain modules. See documentations of particular modules for details of their behaviors in training/evaluation mode, if they are affected, e.g. Dropout, BatchNorm, etc.
I recently started working with Pytorch-lightning, which wraps much of the boilerplate in the training-validation-testing pipelines.
Among other things, it makes model.eval() and model.train() near redundant by allowing the train_step and validation_step callbacks which wrap the eval and train so you never forget to.
• normalisation layers1 use per-batch statistics • activates Dropout layers2
Sets model in evaluation (inference) mode:
• normalisation layers use running statistics • de-activates Dropout layers
Equivalent to model.train(False).
You can turn off evaluation mode by running model.train(). You should use it when running your model as an inference engine - i.e. when testing, validating, and predicting (though practically it will make no difference if your model does not include any of the differently behaving layers).