[DNN] Batch Normalization

yozzum·2025년 2월 3일
0

Machine Learning

목록 보기
30/30

In logistic regression, normalizing INPUT FEATURES speeds up learning.

In deep learning, for any hidden layer, can we normalize the activation functions so as to train W, b faster?

  • There are some debates whether you should apply normalizing before or after activation function.

  • But in practice it is known that normalizing the values before activation is better.

  • You let the model learn gamma and beta to reshape the distribution of Z.
  • Note that you don't want your values in hidden layers to have a mean of 0 and variance of 1 since you want the advantage of non-linearity.

  • Batch Norm is carried out between z and a.

※ This beta is different from the beta of momentum.

profile
yozzum

1개의 댓글

comment-user-thumbnail
2025년 2월 24일

Software licenses can add up quickly, especially for businesses requiring multiple tools. Opting for subscription-based software models or hosted desktops open-source alternatives helps reduce costs while ensuring businesses have access to necessary applications without large initial investments.

답글 달기