I use a ImageNet pretrained ResNet18 to classify something else. The activation function throughout the network is ReLU. Ideally the mean and std should be 0.5 and 1 right? However, the mean and std of activation are around 0.2 for all layers. They tend to increase or decrease together. Is this normal?