The Mixture-of-Experts (MoE) model has succeeded in Deep Learning (DL). However, its complex architecture and advantages over dense models in image classification remain unclear. In previous studies, MoE performance has often been affected by noise and outliers in the input space. Some