Research Article

Mathematical Analysis and Performance Evaluation of the GELU Activation Function in Deep Learning

Table 2

Test loss and test accuracy for selected activation functions on CIFAR-100 and STL-10 datasets.

DatasetsActivationTest lossTest accuracy (%)

CIFAR-100ELU1.560957.26
Hardswish1.312264.12
LeakyReLU1.424861.71
ReLU1.422361.84
ReLU61.418561.58
RReLU1.450959.81
SELU1.831551.09
GELU1.335164.71

STL-10ELU1.553341.78
Hardswish1.245754.40
LeakyReLU1.165056.26
ReLU1.210554.86
ReLU61.504447.01
RReLU1.281451.25
SELU1.522141.18
GELU1.185358.48

Bold indicates the best performance; italic indicates the second-best.