Layer implementing Batch Normalization.
The input from each batch are normalized during training to have zero mean and unit variance and they are then scaled by two parameter, different for each input variable:
In addition a running batch mean and variance is computed and stored in the class During inference the inputs are not normalized using the batch mean but the previously computed at running mean and variance If momentum is in [0,1) the running mean and variances are the exponential averages using the momentum value running_mean = momentum * running_mean + (1-momentum) * batch_mean If instead momentum<1 the cumulative average is computed running_mean = (nb/(nb+1) * running_mean + 1/(nb+1) * batch_mean
See more at [https://arxiv.org/pdf/1502.03167v3.pdf]
Definition at line 64 of file BatchNormLayer.h.
Public Types | |
| using | BNormDescriptors_t = typename Architecture_t::BNormDescriptors_t |
| using | HelperDescriptor_t = typename Architecture_t::TensorDescriptor_t |
| using | Matrix_t = typename Architecture_t::Matrix_t |
| using | Scalar_t = typename Architecture_t::Scalar_t |
| using | Tensor_t = typename Architecture_t::Tensor_t |
Public Member Functions | |
| TBatchNormLayer (const TBatchNormLayer &) | |
| Copy Constructor. | |
| TBatchNormLayer (size_t batchSize, size_t inputDepth, size_t inputHeight, size_t inputWidth, const std::vector< size_t > &shape, int axis=-1, Scalar_t momentum=-1., Scalar_t epsilon=0.0001) | |
| Constructor. | |
| TBatchNormLayer (TBatchNormLayer< Architecture_t > *layer) | |
| Copy the dense layer provided as a pointer. | |
| ~TBatchNormLayer () | |
| Destructor. | |
| void | AddWeightsXMLTo (void *parent) override |
| Writes the information and the weights about the layer in an XML node. | |
| void | Backward (Tensor_t &gradients_backward, const Tensor_t &activations_backward) override |
| Compute weight, bias and activation gradients. | |
| void | CopyBiases (const std::vector< Matrix_t > &otherBiases) |
| Copies the biases provided as an input. | |
| template<typename Arch> | |
| void | CopyParameters (const VGeneralLayer< Arch > &layer) |
| Copy all trainable weight and biases from another equivalent layer but with different architecture The function can copy also extra parameters in addition to weights and biases if they are return by the function GetExtraLayerParameters. | |
| void | CopyWeights (const std::vector< Matrix_t > &otherWeights) |
| Copies the weights provided as an input. | |
| void | Forward (Tensor_t &input, bool inTraining=true) override |
| Compute activation of the layer for the given input. | |
| Tensor_t & | GetActivationGradients () |
| const Tensor_t & | GetActivationGradients () const |
| Matrix_t | GetActivationGradientsAt (size_t i) |
| const Matrix_t & | GetActivationGradientsAt (size_t i) const |
| Matrix_t & | GetBatchMean () |
| const Matrix_t & | GetBatchMean () const |
| size_t | GetBatchSize () const |
| Getters. | |
| std::vector< Matrix_t > & | GetBiases () |
| const std::vector< Matrix_t > & | GetBiases () const |
| Matrix_t & | GetBiasesAt (size_t i) |
| const Matrix_t & | GetBiasesAt (size_t i) const |
| std::vector< Matrix_t > & | GetBiasGradients () |
| const std::vector< Matrix_t > & | GetBiasGradients () const |
| Matrix_t & | GetBiasGradientsAt (size_t i) |
| const Matrix_t & | GetBiasGradientsAt (size_t i) const |
| size_t | GetDepth () const |
| Scalar_t | GetEpsilon () const |
| std::vector< Matrix_t > | GetExtraLayerParameters () const override |
| size_t | GetHeight () const |
| EInitialization | GetInitialization () const |
| size_t | GetInputDepth () const |
| size_t | GetInputHeight () const |
| size_t | GetInputWidth () const |
| Matrix_t & | GetIVariance () |
| const Matrix_t & | GetIVariance () const |
| Scalar_t | GetMomentum () const |
| Matrix_t & | GetMuVector () |
| const Matrix_t & | GetMuVector () const |
| Scalar_t | GetNormAxis () const |
| int & | GetNTrainedBatches () |
| const int & | GetNTrainedBatches () const |
| Tensor_t & | GetOutput () |
| const Tensor_t & | GetOutput () const |
| Matrix_t | GetOutputAt (size_t i) |
| const Matrix_t & | GetOutputAt (size_t i) const |
| Matrix_t & | GetReshapedData () |
| const Matrix_t & | GetReshapedData () const |
| Matrix_t & | GetVariance () |
| const Matrix_t & | GetVariance () const |
| Matrix_t & | GetVarVector () |
| const Matrix_t & | GetVarVector () const |
| std::vector< Matrix_t > & | GetWeightGradients () |
| const std::vector< Matrix_t > & | GetWeightGradients () const |
| Matrix_t & | GetWeightGradientsAt (size_t i) |
| const Matrix_t & | GetWeightGradientsAt (size_t i) const |
| std::vector< Matrix_t > & | GetWeights () |
| const std::vector< Matrix_t > & | GetWeights () const |
| Matrix_t & | GetWeightsAt (size_t i) |
| const Matrix_t & | GetWeightsAt (size_t i) const |
| size_t | GetWidth () const |
| void | Initialize () override |
| Initialize the weights and biases according to the given initialization method. | |
| bool | IsTraining () const |
| void | Print () const override |
| Printing the layer info. | |
| void | ReadMatrixXML (void *node, const char *name, Matrix_t &matrix) |
| void | ReadWeightsFromXML (void *parent) override |
| Read the information and the weights about the layer from XML node. | |
| void | ResetTraining () override |
| Reset some training flags after a loop on all batches Some layer (e.g. | |
| void | SetBatchSize (size_t batchSize) |
| Setters. | |
| void | SetDepth (size_t depth) |
| virtual void | SetDropoutProbability (Scalar_t) |
| Set Dropout probability. | |
| void | SetExtraLayerParameters (const std::vector< Matrix_t > ¶ms) override |
| void | SetHeight (size_t height) |
| void | SetInputDepth (size_t inputDepth) |
| void | SetInputHeight (size_t inputHeight) |
| void | SetInputWidth (size_t inputWidth) |
| void | SetIsTraining (bool isTraining) |
| void | SetWidth (size_t width) |
| void | Update (const Scalar_t learningRate) |
| Updates the weights and biases, given the learning rate. | |
| void | UpdateBiases (const std::vector< Matrix_t > &biasGradients, const Scalar_t learningRate) |
| Updates the biases, given the gradients and the learning rate. | |
| void | UpdateBiasGradients (const std::vector< Matrix_t > &biasGradients, const Scalar_t learningRate) |
| Updates the bias gradients, given some other weight gradients and learning rate. | |
| void | UpdateWeightGradients (const std::vector< Matrix_t > &weightGradients, const Scalar_t learningRate) |
| Updates the weight gradients, given some other weight gradients and learning rate. | |
| void | UpdateWeights (const std::vector< Matrix_t > &weightGradients, const Scalar_t learningRate) |
| Updates the weights, given the gradients and the learning rate,. | |
| void | WriteMatrixToXML (void *node, const char *name, const Matrix_t &matrix) |
| void | WriteTensorToXML (void *node, const char *name, const std::vector< Matrix_t > &tensor) |
| helper functions for XML | |
Static Protected Member Functions | |
| static size_t | CalculateNormDim (int axis, size_t c, size_t h, size_t w) |
Protected Attributes | |
| Tensor_t | fActivationGradients |
| Gradients w.r.t. the activations of this layer. | |
| size_t | fBatchSize |
| Batch size used for training and evaluation. | |
| std::vector< Matrix_t > | fBiases |
| The biases associated to the layer. | |
| std::vector< Matrix_t > | fBiasGradients |
| Gradients w.r.t. the bias values of the layer. | |
| size_t | fDepth |
| The depth of the layer. | |
| size_t | fHeight |
| The height of the layer. | |
| EInitialization | fInit |
| The initialization method. | |
| size_t | fInputDepth |
| The depth of the previous layer or input. | |
| size_t | fInputHeight |
| The height of the previous layer or input. | |
| size_t | fInputWidth |
| The width of the previous layer or input. | |
| bool | fIsTraining |
| Flag indicating the mode. | |
| Tensor_t | fOutput |
| Activations of this layer. | |
| std::vector< Matrix_t > | fWeightGradients |
| Gradients w.r.t. the weights of the layer. | |
| std::vector< Matrix_t > | fWeights |
| The weights associated to the layer. | |
| size_t | fWidth |
| The width of this layer. | |
Private Attributes | |
| Tensor_t | fDerivatives |
| First fDerivatives of the activations of this layer. | |
| TDescriptors * | fDescriptors = nullptr |
| Scalar_t | fEpsilon |
| Matrix_t | fIVar |
| Scalar_t | fMomentum |
| The weight decay. | |
| Matrix_t | fMu |
| Matrix_t | fMu_Training |
| int | fNormAxis |
| Normalization axis. For each element of this axis we will compute mean and stddev. | |
| Tensor_t | fReshapedData |
| int | fTrainedBatches = 0 |
| Matrix_t | fVar |
| Matrix_t | fVar_Training |
#include <TMVA/DNN/BatchNormLayer.h>
| using TMVA::DNN::TBatchNormLayer< Architecture_t >::BNormDescriptors_t = typename Architecture_t::BNormDescriptors_t |
Definition at line 72 of file BatchNormLayer.h.
| using TMVA::DNN::TBatchNormLayer< Architecture_t >::HelperDescriptor_t = typename Architecture_t::TensorDescriptor_t |
Definition at line 71 of file BatchNormLayer.h.
| using TMVA::DNN::TBatchNormLayer< Architecture_t >::Matrix_t = typename Architecture_t::Matrix_t |
Definition at line 68 of file BatchNormLayer.h.
| using TMVA::DNN::TBatchNormLayer< Architecture_t >::Scalar_t = typename Architecture_t::Scalar_t |
Definition at line 67 of file BatchNormLayer.h.
| using TMVA::DNN::TBatchNormLayer< Architecture_t >::Tensor_t = typename Architecture_t::Tensor_t |
Definition at line 69 of file BatchNormLayer.h.
| TMVA::DNN::TBatchNormLayer< Architecture_t >::TBatchNormLayer | ( | size_t | batchSize, |
| size_t | inputDepth, | ||
| size_t | inputHeight, | ||
| size_t | inputWidth, | ||
| const std::vector< size_t > & | shape, | ||
| int | axis = -1, | ||
| Scalar_t | momentum = -1., | ||
| Scalar_t | epsilon = 0.0001 ) |
Constructor.
Definition at line 219 of file BatchNormLayer.h.
| TMVA::DNN::TBatchNormLayer< Architecture_t >::TBatchNormLayer | ( | TBatchNormLayer< Architecture_t > * | layer | ) |
Copy the dense layer provided as a pointer.
Definition at line 242 of file BatchNormLayer.h.
| TMVA::DNN::TBatchNormLayer< Architecture_t >::TBatchNormLayer | ( | const TBatchNormLayer< Architecture_t > & | layer | ) |
Copy Constructor.
Definition at line 251 of file BatchNormLayer.h.
| TMVA::DNN::TBatchNormLayer< Architecture_t >::~TBatchNormLayer | ( | ) |
Destructor.
Definition at line 259 of file BatchNormLayer.h.
|
overridevirtual |
Writes the information and the weights about the layer in an XML node.
Implements TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 387 of file BatchNormLayer.h.
|
overridevirtual |
Compute weight, bias and activation gradients.
Uses the precomputed first partial derivatives of the activation function computed during forward propagation and modifies them. Must only be called directly a the corresponding call to Forward(...).
Implements TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 337 of file BatchNormLayer.h.
|
inlinestaticprotected |
Definition at line 199 of file BatchNormLayer.h.
|
inherited |
Copies the biases provided as an input.
Definition at line 468 of file GeneralLayer.h.
|
inherited |
Copy all trainable weight and biases from another equivalent layer but with different architecture The function can copy also extra parameters in addition to weights and biases if they are return by the function GetExtraLayerParameters.
Definition at line 478 of file GeneralLayer.h.
|
inherited |
Copies the weights provided as an input.
Definition at line 458 of file GeneralLayer.h.
|
overridevirtual |
Compute activation of the layer for the given input.
The input must be in 3D tensor form with the different matrices corresponding to different events in the batch. Computes activations as well as the first partial derivative of the activation function at those activations.
Implements TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 295 of file BatchNormLayer.h.
|
inlineinherited |
Definition at line 200 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 199 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 205 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 206 of file GeneralLayer.h.
|
inline |
Definition at line 149 of file BatchNormLayer.h.
|
inline |
Definition at line 148 of file BatchNormLayer.h.
|
inlineinherited |
Getters.
Definition at line 163 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 179 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 178 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 182 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 181 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 191 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 190 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 194 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 193 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 167 of file GeneralLayer.h.
|
inline |
Definition at line 177 of file BatchNormLayer.h.
|
inlineoverridevirtual |
Reimplemented from TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 185 of file BatchNormLayer.h.
|
inlineinherited |
Definition at line 168 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 214 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 164 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 165 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 166 of file GeneralLayer.h.
|
inline |
Definition at line 161 of file BatchNormLayer.h.
|
inline |
Definition at line 160 of file BatchNormLayer.h.
|
inline |
Definition at line 174 of file BatchNormLayer.h.
|
inline |
Definition at line 165 of file BatchNormLayer.h.
|
inline |
Definition at line 164 of file BatchNormLayer.h.
|
inline |
Definition at line 180 of file BatchNormLayer.h.
|
inline |
Definition at line 145 of file BatchNormLayer.h.
|
inline |
Definition at line 144 of file BatchNormLayer.h.
|
inlineinherited |
Definition at line 197 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 196 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 202 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 203 of file GeneralLayer.h.
|
inline |
Definition at line 183 of file BatchNormLayer.h.
|
inline |
Definition at line 182 of file BatchNormLayer.h.
|
inline |
Definition at line 157 of file BatchNormLayer.h.
|
inline |
Definition at line 156 of file BatchNormLayer.h.
|
inline |
Definition at line 169 of file BatchNormLayer.h.
|
inline |
Definition at line 168 of file BatchNormLayer.h.
|
inlineinherited |
Definition at line 185 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 184 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 188 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 187 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 173 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 172 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 176 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 175 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 169 of file GeneralLayer.h.
|
overridevirtual |
Initialize the weights and biases according to the given initialization method.
Reimplemented from TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 269 of file BatchNormLayer.h.
|
inlineinherited |
Definition at line 170 of file GeneralLayer.h.
|
overridevirtual |
Printing the layer info.
Implements TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 369 of file BatchNormLayer.h.
|
inherited |
Definition at line 544 of file GeneralLayer.h.
|
overridevirtual |
Read the information and the weights about the layer from XML node.
Implements TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 412 of file BatchNormLayer.h.
|
inlineoverridevirtual |
Reset some training flags after a loop on all batches Some layer (e.g.
batchnormalization) might need to implement the function in case some operations are needed after looping an all batches
Reimplemented from TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 129 of file BatchNormLayer.h.
|
inlineinherited |
Setters.
Definition at line 217 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 221 of file GeneralLayer.h.
|
inlinevirtualinherited |
Set Dropout probability.
Reimplemented for layers supporting droput
Reimplemented in TMVA::DNN::TDenseLayer< Architecture_t >.
Definition at line 160 of file GeneralLayer.h.
|
inlineoverridevirtual |
Reimplemented from TMVA::DNN::VGeneralLayer< Architecture_t >.
Definition at line 192 of file BatchNormLayer.h.
|
inlineinherited |
Definition at line 222 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 218 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 219 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 220 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 224 of file GeneralLayer.h.
|
inlineinherited |
Definition at line 223 of file GeneralLayer.h.
|
inherited |
Updates the weights and biases, given the learning rate.
Definition at line 410 of file GeneralLayer.h.
|
inherited |
Updates the biases, given the gradients and the learning rate.
Definition at line 428 of file GeneralLayer.h.
|
inherited |
Updates the bias gradients, given some other weight gradients and learning rate.
Definition at line 448 of file GeneralLayer.h.
|
inherited |
Updates the weight gradients, given some other weight gradients and learning rate.
Definition at line 438 of file GeneralLayer.h.
|
inherited |
Updates the weights, given the gradients and the learning rate,.
Definition at line 418 of file GeneralLayer.h.
|
inherited |
Definition at line 521 of file GeneralLayer.h.
|
inherited |
helper functions for XML
Definition at line 496 of file GeneralLayer.h.
|
protectedinherited |
Gradients w.r.t. the activations of this layer.
Definition at line 78 of file GeneralLayer.h.
|
protectedinherited |
Batch size used for training and evaluation.
Definition at line 59 of file GeneralLayer.h.
|
protectedinherited |
The biases associated to the layer.
Definition at line 72 of file GeneralLayer.h.
|
protectedinherited |
Gradients w.r.t. the bias values of the layer.
Definition at line 75 of file GeneralLayer.h.
|
protectedinherited |
The depth of the layer.
Definition at line 65 of file GeneralLayer.h.
|
private |
First fDerivatives of the activations of this layer.
Definition at line 77 of file BatchNormLayer.h.
|
private |
Definition at line 97 of file BatchNormLayer.h.
|
private |
Definition at line 82 of file BatchNormLayer.h.
|
protectedinherited |
The height of the layer.
Definition at line 66 of file GeneralLayer.h.
|
protectedinherited |
The initialization method.
Definition at line 80 of file GeneralLayer.h.
|
protectedinherited |
The depth of the previous layer or input.
Definition at line 61 of file GeneralLayer.h.
|
protectedinherited |
The height of the previous layer or input.
Definition at line 62 of file GeneralLayer.h.
|
protectedinherited |
The width of the previous layer or input.
Definition at line 63 of file GeneralLayer.h.
|
protectedinherited |
Flag indicating the mode.
Definition at line 69 of file GeneralLayer.h.
|
private |
Definition at line 86 of file BatchNormLayer.h.
|
private |
The weight decay.
Definition at line 81 of file BatchNormLayer.h.
|
private |
Definition at line 84 of file BatchNormLayer.h.
|
private |
Definition at line 88 of file BatchNormLayer.h.
|
private |
Normalization axis. For each element of this axis we will compute mean and stddev.
Definition at line 79 of file BatchNormLayer.h.
|
protectedinherited |
Activations of this layer.
Definition at line 77 of file GeneralLayer.h.
|
private |
Definition at line 92 of file BatchNormLayer.h.
|
private |
Definition at line 95 of file BatchNormLayer.h.
|
private |
Definition at line 85 of file BatchNormLayer.h.
|
private |
Definition at line 89 of file BatchNormLayer.h.
|
protectedinherited |
Gradients w.r.t. the weights of the layer.
Definition at line 74 of file GeneralLayer.h.
|
protectedinherited |
The weights associated to the layer.
Definition at line 71 of file GeneralLayer.h.
|
protectedinherited |
The width of this layer.
Definition at line 67 of file GeneralLayer.h.