Skip to content

KL divergence : size of update #3

@astariul

Description

@astariul

In the section Policy Gradients diagnostics, it is mentioned :

If KL is .01 then very small.
If 10 then too much.

I couldn't find the given values in the slides of the original talk. Where these values are from ?


Also, in the previous part (about entropy), you mentioned how to fix the problem with fast-dropping entropy, and it is really helpful.

Does anyone know how to fix the problem of KL divergence being too low ?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions