AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (1.8 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Open Access

Federated Learning on Heterogeneous Images in Internet of Biomedical Things

College of Information Science and Technology, Jinan University, Guangzhou 510632, China
Division of Natural and Applied Sciences, Duke Kunshan University, Suzhou 215316, China
College of Computer, Qinghai Normal University, Qinghai 810000, China
Show Author Information

Abstract

Federated Learning (FL) allows the Internet of bioMedical Things (IoMT) devices to collaboratively train a global model without centralizing data, yet the heterogeneity of biomedical images poses challenges in achieving optimal performance for each IoMT device. Although many personalized FL methods have been proposed, their adaptability is often limited to data with certain levels of heterogeneity. In this paper, we propose a novel FL method to handle heterogeneous images in IoMT. First, we propose a gradient-fusion method that integrates both shared and local gradients into the locally adapted model for improved personalization. The shared gradients, aggregated from all IoMT devices through a learnable matrix, capture collective intelligence, while the local gradients, originating from each device’s data, reflect individual data distributions. This dynamic integration of collective and device-specific insights effectively mitigates data heterogeneity. Second, we propose a privacy-enhanced approach that delegates part of the gradient computation to devices, thereby protecting data privacy without compromising the efficacy of the gradient-fusion process. Finally, for enhanced performance, we introduce a layer-wise aggregation method to precisely measure the contribution of different layers in local models. Extensive evaluations on imaging datasets, featuring various types and degrees of data heterogeneity, demonstrate the superior performance of our methods over existing baselines.

References

[1]

Y. A. Qadri, A. Nauman, Y. B. Zikria, A. V. Vasilakos, and S. W. Kim, The future of healthcare Internet of Things: A survey of emerging technologies, IEEE Commun. Surv. Tutor., vol. 22, no. 2, pp. 1121–1167, 2020.

[2]

Z. Ning, P. Dong, X. Wang, X. Hu, L. Guo, B. Hu, Y. Guo, T. Qiu, and R. Y. K. Kwok, Mobile edge computing enabled 5G health monitoring for Internet of medical things: A decentralized game theoretic approach, IEEE J. Sel. Areas Commun., vol. 39, no. 2, pp. 463–478, 2021.

[3]

G. Shamai, A. Livne, A. Polónia, E. Sabo, A. Cretu, G. Bar-Sela, and R. Kimmel, Deep learning-based image analysis predicts PD-L1 status from H & E-stained histopathology images in breast cancer, Nat. Commun., vol. 13, no. 1, p. 6753, 2022.

[4]

J. De Fauw, J. R. Ledsam, B. Romera-Paredes, S. Nikolov, N. Tomasev, S. Blackwell, H. Askham, X. Glorot, B. O’Donoghue, D. Visentin, et al., Clinically applicable deep learning for diagnosis and referral in retinal disease, Nat. Med., vol. 24, pp. 1342–1350, 2018.

[5]

A. Y. Hannun, P. Rajpurkar, M. Haghpanahi, G. H. Tison, C. Bourn, M. P. Turakhia, and A. Y. Ng, Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network, Nat. Med., vol. 25, no. 1, pp. 65–69, 2019.

[6]

M. Pandey, M. Fernandez, F. Gentile, O. Isayev, A. Tropsha, A. C. Stern, and A. Cherkasov, The transformational role of GPU computing and deep learning in drug discovery, Nat. Mach. Intell., vol. 4, no. 3, pp. 211–221, 2022.

[7]
B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, Communication-efficient learning of deep networks from decentralized data, in Proc. 20th Int. Conf. on Artificial Intelligence and Statistics, doi:10.48550/arXiv.1602.05629.
[8]
S. P. Karimireddy, S. Kale, M. Mohri, S. Reddi, S. Stich, and A. T. Suresh, SCAFFLD: Stochastic controlled averaging for federated learning, in Proc. 37 th Int. Conf. on Machine Learning, Virtual Event, 2020, pp. 5132–5143.
[9]
T. Li, S. Hu, A. Beirami, and V. Smith, Ditto: Fair and robust federated learning through personalization, arXiv preprint arXiv:2012.04221, 2020.
[10]

T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, Federated optimization in heterogeneous networks, Proceedings of Machine Learning and Systems, vol. 2, pp. 429–450, 2020.

[11]
J. Liu, J. Wu, J. Chen, M. Hu, Y. Zhou, and D. Wu, FedDWA: Personalized federated learning with dynamic weight adjustment, in Proc. Thirty-Second Int. Joint Conf. Artificial Intelligence, Macau, China, 2023, pp. 3993–4001.
[12]
F. Chen, G. Long, Z. Wu, T. Zhou, and J. Jiang, Personalized federated learning with a graph, in Proc. Thirty-First Int. Joint Conf. Artificial Intelligence, Vienna, Austria, 2022, pp. 2575–2582.
[13]
L. Collins, H. Hassani, A. Mokhtari, and S. Shakkottai, Exploiting shared representations for personalized federated learning, in Proc. Int. Conf. on Machine Learning, Virtual Event, 2021, pp. 2089–2099.
[14]
H.-Y. Chen and W.-L. Chao, On bridging generic and personalized federated learning for image classification, in Proc. Int. Conf. on Learning Representations, Virtual Event, doi: 10.48550/arXiv.2107.00778.
[15]
P. Tian, W. Liao, W. Yu, and E. Blasch, WSCC: A weight-similarity-based client clustering approach for non-IID federated learning, IEEE Internet Things J., vol. 9, no. 20, pp. 20243–20256, 2022.
[16]

H. Jamali-Rad, M. Abdizadeh, and A. Singh, Federated learning with taskonomy for non-IID data, IEEE Trans. Neural Netw. Learn. Syst., vol. 34, no. 11, pp. 8719–8730, 2023.

[17]
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, Attention is all you need, in Proc. 31st Int. Conf. on Neural Information Processing Systems, Long Beach, CA, USA, pp. 6000 – 6010.
[18]
S. Ji, S. Pan, G. Long, X. Li, J. Jiang, and Z. Huang, Learning private neural language modeling with attentive aggregation, in Proc. Int. Joint Conf. Neural Networks (IJCNN ), Budapest, Hungary, 2019, pp. 1–8.
[19]
Q. Li, B. He, and D. Song, Model-contrastive federated learning, in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, Nashville, TN, USA, 2021, pp. 10708–10717.
[20]
Y. Zhao, M. Li, L. Lai, N. Suda, D. Civin, and V. Chandra, Federated learning with non-IID data, arXiv preprint arXiv:1806.00582, 2018.
[21]

Q. Wu, X. Chen, Z. Zhou, and J. Zhang, FedHome: Cloud-edge based personalized federated learning for in-home health monitoring, IEEE Trans. Mob. Comput., vol. 21, no. 8, pp. 2818–2832, 2022.

[22]

Z. He, Y. Li, D. Seo, and Z. Cai, FedCPD: Addressing label distribution skew in federated learning with class proxy decoupling and proxy regularization, Information Fusion, vol. 110, p. 102481, 2024.

[23]

M. Duan, D. Liu, X. Chen, R. Liu, Y. Tan, and L. Liang, Self-balancing federated learning with global imbalanced data in mobile systems, IEEE Trans. Parallel Distrib. Syst., vol. 32, no. 1, pp. 59–71, 2021.

[24]
L. Wang, Z. He, and Z. Cai, Local differentially private clustered federated learning on heterogeneous data, in Proc. IEEE 20th Int. Conf. Mobile Ad Hoc and Smart Systems, Toronto, Canada, 2023, pp. 288–294.
[25]

Z. He, L. Wang, and Z. Cai, Clustered federated learning with adaptive local differential privacy on heterogeneous IoT data, IEEE Internet Things J., vol. 11, no. 1, pp. 137–146, 2024.

[26]
E. Jeong, S. Oh, H. Kim, J. Park, M. Bennis, and S.-L. Kim, Communication-efficient on-device machine learning: Federated distillation and augmentation under non-IID private data, arXiv preprint arXiv:1811.11479, 2018.
[27]
A. Fallah, A. Mokhtari, and A. Ozdaglar, Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach, in Proc. 33rd Int. Conf. on Neural Information Processing Systems, Vancouver, Canada, 2020, pp. 3557–3568.
[28]
Y. Jiang, J. Konečný, K. Rush, and S. Kannan, Improving federated learning personalization via model agnostic meta learning, arXiv preprint arXiv:1909.12488, 2019.
[29]
C. T. Dinh, N. Tran, and J. Nguyen, Personalized federated learning with moreau envelopes, in Proc. 33rd Int. Conf. on Neural Information Processing Systems, Vancouver, Canada, 2019, pp. 21394–21405.
[30]
M. Khodak, M.-F. F. Balcan, and A. S. Talwalkar, Adaptive gradient-based meta-learning methods, in Proc. 33rd Int. Conf. on Neural Information Processing Systems, Vancouver, Canada, 2019, pp. 5917 – 5928.
[31]
C. Finn, P. Abbeel, and S. Levine, Model agnosticmeta-learning for fast adaptation of deep networks, in Proc. 34th Int. Conf. on Machine Learning, Sydney, Australia, 2017, pp. 1126 – 1135.
[32]
P. P. Liang, T. Liu, Z. Liu, N. B. Allen, R. P. Auerbach, D. Brent, R. Salakhutdinov, and L.-P. Morency, Think locally, act globally: Federated learning with local and global representations, arXiv preprint arXiv:2001.01523, 2020.
[33]
F. Chen, G. Long, Z. Wu, T. Zhou, and J. Jiang, Personalized federated learning with graph, arXiv preprint arXiv:2203.00829, 2022.
[34]
K. Wang, R. Mathews, C. Kiddon, H. Eichner, F. Beaufays, and D. Ramage, Federated evaluation of on-device personalization, arXiv preprint arXiv:1910.10252, 2019.
[35]
M. G. Arivazhagan, V. Aggarwal, A. K. Singh, and S. Choudhary, Federated learning with personalization layers, arXiv preprint arXiv:1912.00818, 2019.
[36]
J. Schneider and M. Vlachos, Personalization of deep learning, in Data Science—Analytics and Applications : Proceedings of the 3 rd International Data Science Conference, https://doi.org/10.1007/978-3-658-32182-6_14, 2021.
[37]

Z. He, F. Zhang, Y. Li, Y. Cao, and Z. Cai, Privacy-enhanced personalized federated learning with layer-wise gradient shielding on heterogeneous IoT data, IEEE Internet Things J., vol. 11, no. 15, pp. 25771–25781, 2024.

[38]
G. Long, M. Xie, T. Shen, T. Zhou, X. Wang, J. Jiang, C. Zhang, G. Long, M. Xie, T. Shen, et al., Multi-center federated learning: Clients clustering for better personalization, arXiv preprint arXiv:2108.08647, 2021.
[39]
Y. Mansour, M. Mohri, J. Ro, and A. T. Suresh, Three approaches for personalization with applications to federated learning, arXiv preprint arXiv:2002.10619, 2020.
[40]

J. Xu, J. Lin, Y. Li, and Z. Xu, MultiFed: A fast converging federated learning framework for services QoS prediction via cloud-edge collaboration mechanism, Knowl. Based Syst., vol. 268, p. 110463, 2023.

[41]
A. Ghosh, J. Chung, D. Yin, and K. Ramchandran, An efficient framework for clustered federated learning, in Proc. 34th Int. Conf. on Neural Information Processing Systems, Vancouver, Canada, 2020, pp. 19586 – 19597.
[42]

L. Huang, A. L. Shea, H. Qian, A. Masurkar, H. Deng, and D. Liu, Patient clustering improves efficiency of federated machine learning to predict mortality and hospital stay time using distributed electronic medical records, J. Biomed. Inform., vol. 99, p. 103291, 2019.

[43]
X. Ma, J. Zhang, S. Guo, and W. Xu, Layer-wised model aggregation for personalized federated learning, in Proc. IEEE/CVF Conf. Computer Vision and Pattern Recognition, New Orleans, LA, USA, 2022, pp. 10082–10091.
[44]

Y. Tan, G. Long, L. Liu, T. Zhou, Q. Lu, J. Jiang, and C. Zhang, FedProto: Federated prototype learning across heterogeneous clients, Proc. AAAI Conf. Artif. Intell., vol. 36, no. 8, pp. 8432–8440, 2022.

[45]
M. Luo, F. Chen, D. Hu, Y. Zhang, J. Liang, and J. Feng, No fear of heterogeneity: Classifier calibration for federated learning with non-iid data, in Proc. 35th Int. Conf. on Neural Information Processing Systems, Virtual Event, 2021, pp. 5972 – 5984.
[46]

A. Krizhevsky, I. Sutskever, and G. E. Hinton, Learning multiple layers of features from tiny images, Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2009.

[47]
L. N. Darlow, E. J. Crowley, A. Antoniou, and A. J. Storkey, CINIC-10 is not ImageNet or CIFAR-10, arXiv preprint arXiv:1810.03505, 2018.
[48]
S. Zheng, Q. Meng, T. Wang, W. Chen, N. Yu, Z.-M. Ma, and T.-Y. Liu, Asynchronous stochastic gradient descent with delay compensation, in Proc. 34th Int. Conf. on Machine Learning, Sydney, Australia, 2017, pp. 4120 – 4129.
[49]
Y. Tay, D. Bahri, D. Metzler, D.-C. Juan, Z. Zhao, and C. Zheng, Synthesizer: Rethinking self-attention for transformer models, in Proc. Int. Conf. on Machine Learning, doi: 10.48550/arXiv.2005.00743.
Big Data Mining and Analytics
Pages 1237-1250
Cite this article:
Li Y, He Z, He P, et al. Federated Learning on Heterogeneous Images in Internet of Biomedical Things. Big Data Mining and Analytics, 2024, 7(4): 1237-1250. https://doi.org/10.26599/BDMA.2024.9020067

54

Views

0

Downloads

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Altmetrics

Received: 19 February 2024
Revised: 27 June 2024
Accepted: 30 September 2024
Published: 04 December 2024
© The author(s) 2024.

The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).

Return