AdaGrad

struct defined in module Flux.Optimise


			AdaGrad(η = 0.1, ϵ = 1.0e-8)

AdaGrad optimiser. It has parameter specific learning rates based on how frequently it is updated. Parameters don't need tuning.

Parameters

  • Learning rate ( η): Amount by which gradients are discounted before updating the weights.

Examples


			
			
			
			opt
			 
			=
			 
			

			AdaGrad
			(
			)
			

			

			
			opt
			 
			=
			 
			

			AdaGrad
			(
			0.001
			)
Methods