Table 6 Strengths and gaps in traditional CNN models.

From: Improved facial emotion recognition model based on a novel deep convolutional structure

 

Accuracy

Strength

Gap

DenseNet121

74%

Reduce overfitting, increase feature reuse, reduce the number of parameters

The massive increase in parameters caused by an increase in the number of feature maps in every layer

InceptionV3

77%

End up replacing large - sized filters with (1 × 7), and (1 × 5) filters

Complicated structure.

Lack of consistency.

VGG16

83%

Max Pooling has been applied

The use relatively demanding fully - connected layers

VGG19

88%

More powerful than VGG16

More complex, more computationally intensive potentially more prone to overfitting.

ResNet50

87%

The concept of residual learning was introduced. Error rate for deeper networks has been lessened

Little-complex architecture

Xception

91%

More efficient in terms of processing time than standard convolutions.

Depth-wise separable convolution layer replaces the typical Inception modules

EfficientNetB0

93%

Employs a compounded coefficient to scale all depth/width/resolution elements similarly

Slow training at the large image resolution