BatchNormalization layer vs hasLayerNorm in the Dense Layer?

What difference between the BatchNormalization layer and hasLayerNorm in the Dense Layer? Should I create new Batch Layer with the same ins-outs from previous Dense Layer or it’s the same as true flag at hasLayerNorm?


Batch Normalization und Layer Normalization are two entirely different things.

Unlike batch normalization, Layer Normalization directly estimates the normalization statistics from the summed inputs to the neurons within a hidden layer so the normalization does not introduce any new dependencies between training cases.
– from Layer Normalization on Papers With Code

Oh, thank you a lot, Treo, the link clearified my question.