P-TELU: Parametric Tan Hyperbolic Linear Unit Activation for Deep Neural Networks

Rahul Duggal, Anubha Gupta; Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 974-978

Abstract


This paper proposes a new activation function, namely, Parametric Tan Hyperbolic Linear Unit (P-TELU) for deep neural networks. The work is inspired from two recently proposed functions: Parametric RELU (P-RELU) and Exponential Linear Unit (ELU). The specific design of P-TELU allows it to leverage two advantages: (1) the flexibility of tuning parameters from the data distribution similar to P-RELU and (2) better noise robustness similar to ELU. Owing to larger gradient and early saturation of tan hyperbolic compared to exponential function, the proposed activation allows a neuron to reach/exit from the noise robust deactivation state earlier and faster. The performance of the proposed function is evaluated on CIFAR10 and CIFAR100 image dataset using two convolutional neural network (CNN) architectures : KerasNet, a small 6 layer CNN model, and on 76 layer deep ResNet architecture. Results demonstrate enhanced performance of the proposed activation function.

Related Material


[pdf]
[bibtex]
@InProceedings{Duggal_2017_ICCV,
author = {Duggal, Rahul and Gupta, Anubha},
title = {P-TELU: Parametric Tan Hyperbolic Linear Unit Activation for Deep Neural Networks},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV) Workshops},
month = {Oct},
year = {2017}
}