nntoolbox.components.activation module¶
-
class
nntoolbox.components.activation.
LWTA
(block_size)[source]¶ Bases:
torch.nn.modules.module.Module
Local Winner-Take-All Layer
For every k consecutive units, keep only the one with highest activations and zero-out the rest.
- References:
Rupesh Kumar Srivastava, Jonathan Masci, Sohrob Kazerounian, Faustino Gomez, Jürgen Schmidhuber. “Compete to Compute.” https://papers.nips.cc/paper/5059-compete-to-compute.pdf
-
forward
(input: torch.Tensor) → torch.Tensor[source]¶ Defines the computation performed at every call.
Should be overridden by all subclasses.
Note
Although the recipe for forward pass needs to be defined within this function, one should call the
Module
instance afterwards instead of this since the former takes care of running the registered hooks while the latter silently ignores them.
-
training
: bool¶
-
class
nntoolbox.components.activation.
ZeroCenterRelu
(inplace: bool = False)[source]¶ Bases:
torch.nn.modules.activation.ReLU
As described by Jeremy of FastAI
-
forward
(input: torch.Tensor) → torch.Tensor[source]¶ Defines the computation performed at every call.
Should be overridden by all subclasses.
Note
Although the recipe for forward pass needs to be defined within this function, one should call the
Module
instance afterwards instead of this since the former takes care of running the registered hooks while the latter silently ignores them.
-
inplace
: bool¶
-