Research Article

Mathematical Analysis and Performance Evaluation of the GELU Activation Function in Deep Learning

Table 1

Test loss and test accuracy for different activation functions on CIFAR-10 dataset.

ActivationsTest lossTest accuracy (%)

ELU0.423286.22
Hardshrink1.126660.81
Hardsigmoid1.429654.00
Hardtanh0.557382.01
Hardswish0.392188.77
LeakyReLU0.403687.93
LogSigmoid0.575581.42
PReLU0.555286.33
ReLU0.447887.19
ReLU60.414588.70
RReLU0.430885.91
SELU0.498383.37
CELU0.426086.21
Sigmoid3.210233.90
Softplus0.576280.82
Softshrink0.562681.93
Softsign0.681978.33
Tanh0.531882.91
Tanhshrink0.577680.78
GELU0.368589.52

Bold indicates the best performance; italic indicates the second-best.