This is the repository of the paper ERA: Enhanced Rational Activations. The code uses TensorFlow 2.5. (Code for PyTorch can be found here.) For other pip recommended pip packages, please check the requirements.txt file. To run on GPUs, we use CUDA 11.2.2 and CuDNN 8.1.0.77.
Enhanced Rational Activations (ERAs) are a new type of activation functions. They are rational functions with a safe polynomial denominator based on quadratic factorisation. ERAs of degree
The parameters of
ERAs work best when the pre-activation values are normalized by standard techniques like instance normalization for CNNs and layer normalization for MLPs. Normalization makes the pre-activation distribution evolve slowly during training, which benefits the learning process.
Plot of Swish-initialised ERA with denominator of degree 4:
Feel free to open an issue if you have any questions. The corresponding author is Martin Trimmel from Lund University.
@inproceedings{ERA2022,
title={ERA: Enhanced Rational Activations},
author={Trimmel, Martin and Zanfir, Mihai and Hartley, Richard and Sminchisescu, Cristian},
booktitle={European Conference on Computer Vision},
year={2022}
}