#KLDivergence Articles


KL Divergence Regularization for RL

Glen B

Intro

Regularization is a very common practice in machine learning. In supervized learning it is used to reduce model complexity. It also helps prevent the model from over fitting the test data. In RL because of the availablility of limitless data regularization does not need to be used to reduce …