Tutorial by Examples

The following is an example definition for training a BatchNorm layer with channel-wise scale and bias. Typically a BatchNorm layer is inserted between convolution and rectification layers. In this example, the convolution would output the blob layerx and the rectification would receive the layerx-b...
The main change needed is to switch use_global_stats to true. This switches to using the moving average. layer { bottom: 'layerx' top: 'layerx-bn' name: 'layerx-bn' type: 'BatchNorm' batch_norm_param { use_global_stats: true # use pre-calculated average and variance } param { lr_mult...

Page 1 of 1