WeightDecay
struct defined in module Flux.Optimise
struct
Flux.Optimise
WeightDecay(λ = 0)
Decay weights by λ. Typically composed with other optimizers as the first transformation to the gradient, making it equivalent to adding L_2 regularization with coefficient λ to the loss.
opt = Optimiser ( WeightDecay ( 1f-4 ) , Adam ( ) )
There are 2 methods for Flux.Optimise.WeightDecay:
Flux.Optimise.WeightDecay
The following pages link back here:
FastAI.jl , training/onecycle.jl , Flux.jl , optimise/Optimise.jl , optimise/optimisers.jl