kldivergence
	
function defined in module 
	Flux.Losses
			kldivergence(ŷ, y; agg = mean, ϵ = eps(ŷ))
Return the Kullback-Leibler divergence between the given probability distributions.
The KL divergence is a measure of how much one probability distribution is different from the other. It is always non-negative, and zero only when both the distributions are equal.
			julia> p1 = [1 0; 0 1]
2×2 Matrix{Int64}:
 1  0
 0  1
julia> p2 = fill(0.5, 2, 2)
2×2 Matrix{Float64}:
 0.5  0.5
 0.5  0.5
julia> Flux.kldivergence(p2, p1) ≈ log(2)
true
julia> Flux.kldivergence(p2, p1; agg = sum) ≈ 2log(2)
true
julia> Flux.kldivergence(p2, p2; ϵ = 0)  # about -2e-16 with the regulator
0.0
julia> Flux.kldivergence(p1, p2; ϵ = 0)  # about 17.3 with the regulator
Inf
There is
			1
			method for Flux.Losses.kldivergence:
		
The following pages link back here: