public gated_mlp_block — function
gated_mlp(gate_layer, inplanes::Integer, hidden_planes::Integer, 
          outplanes::Integer = inplanes; dropout = 0., activation = gelu)
Feedforward block based on the implementation in the paper “Pay Attention to MLPs”. (reference)
Arguments
gate_layer: Layer to use for the gating.inplanes: Number of dimensions in the input.hidden_planes: Number of dimensions in the intermediate layer.outplanes: Number of dimensions in the output - by default it is the same asinplanes.dropout: Dropout rate.activation: Activation function to use.