Recent advances in neural network architectures reveal the importance of diverse representations. However, simply integrating more branches or increasing the width for the diversity would inevitably increase model complexity, leading to prohibitive inference costs. In this paper, we revisit the learnable parameters in neural networks and showcase that it is feasible to disentangle learnable parameters to latent sub-parameters, which focus on different patterns and representations. This important finding leads us to study further the aggregation of diverse representations in a network structure. To this end, we propose Parameter Disentanglement for Diverse Representations (PDDR), which considers diverse patterns in parallel during training, and aggregates them into one for efficient inference. To further enhance the diverse representations, we develop a lightweight refinement module in PDDR, which adaptively refines the combination of diverse representations according to the input. PDDR can be seamlessly integrated into modern networks, significantly improving the learning capacity of a network while maintaining the same complexity for inference. Experimental results show great improvements on various tasks, with an improvement of 1.47% over Residual Network 50 (ResNet50) on ImageNet, and we improve the detection results of Retina Residual Network 50 (Retina-ResNet50) by 1.7% Mean Average Precision (mAP). Integrating PDDR into recent lightweight vision transformer models, the resulting model outperforms related works by a clear margin.
S. Y. Wang, Z. Qu, and C. J. Li, A dense-aware cross-splitNet for object detection and recognition, IEEE Trans. Circuits Syst. Video Technol., vol. 33, no. 5, pp. 2290–2301, 2023.
Y. Xie, X. Hou, Y. Guo, X. Wang, and J. Zheng, Joint-guided distillation binary neural network via dynamic channel-wise diversity enhancement for object detection, IEEE Trans. Circuits Syst. Video Technol., vol. 34, no. 1, pp. 448–460, 2024.
L. Zhang, M. Wang, M. Liu, and D. Zhang, A survey on deep learning for neuroimaging-based brain disorder analysis, Front. Neurosci., vol. 14, pp. 779, 2020.
G. Zhu, J. Cao, L. Chen, Y. Wang, Z. Bu, S. Yang, J. Wu, and Z. Wang, A multi-task graph neural network with variational graph auto-encoders for session-based travel packages recommendation, ACM Trans. Web, vol. 17, no. 3, pp. 1–30, 2023.
J. P. S. Schuler, S. Romani, M. Abdel-Nasser, H. Rashwan, and D. Puig, Grouped pointwise convolutions reduce parameters in convolutional neural networks, Mendel, vol. 28, no. 1, pp. 23–31, 2022.
M. Tan, W. Gao, H. Li, J. Xie, and M. Gong, Universal binary neural networks design by improved differentiable neural architecture search, IEEE Trans. Circuits Syst. Video Technol., vol. 34, no. 10, pp. 9153–9165, 2024.
Y. Han, G. Huang, S. Song, L. Yang, H. Wang, and Y. Wang, Dynamic neural networks: A survey, IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 11, pp. 7436–7456, 2022.
M. H. Guo, C. Z. Lu, Z. N. Liu, M. M. Cheng, and S. M. Hu, Visual attention network, Comput. Vis. Medium., vol. 9, no. 4, pp. 733–752, 2023.
H. Touvron, P. Bojanowski, M. Caron, M. Cord, A. El-Nouby, E. Grave, G. Izacard, A. Joulin, G. Synnaeve, J. Verbeek, et al., ResMLP: Feedforward networks for image classification with data-efficient training, IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 4, pp. 5314–5321, 2023.
X. Liu, C. Hu, and P. Li, Automatic segmentation of overlapped poplar seedling leaves combining mask R-CNN and DBSCAN, Comput. Electron. Agric., vol. 178, pp. 105753, 2020.
Z. Yang, W. He, X. Fan, and T. Tjahjadi, PlantNet: Transfer learning-based fine-grained network for high-throughput plants recognition, Soft Comput. A Fusion Found. Methodol. Appl., vol. 26, no. 20, pp. 10581–10590, 2022.
P. Helber, B. Bischke, A. Dengel, and D. Borth, EuroSAT: A novel dataset and deep learning benchmark for land use and land cover classification, IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens., vol. 12, no. 7, pp. 2217–2226, 2019.