The proliferation of Large Language Models (LLMs) has catalyzed the growth of various industries. It is therefore imperative to ensure the controlled and beneficial application of LLMs across specific domains for downstream tasks through transfer learning, while preserving their general capabilities. We propose a novel and on-device efficient fine-tuning optimization algorithm for LLMs, utilizing federated transfer learning. Specifically, we introduce the Fusion of low Rank Adaptation (FoRA) optimization algorithm from a micro perspective, which enhances multi-dimensional feature aggregation through the addition of efficient parameters. From a meso perspective, we extend the application of the FoRA algorithm across all linear layers within the Transformer architecture to facilitate downstream task performance. Finally, from a macro perspective and with a focus on the medical domain, we incorporate quantization techniques into the federated learning framework to achieve on-device efficient fine-tuning optimization, thereby offering dual protection for data and model integrity. Our results indicate that, compared to existing state-of-the-art methods, our algorithm significantly improves LLM performance while ensuring dual privacy protection of both data and models.
Y. Liu, Y. Qu, C. Xu, Z. Hao, and B. Gu, Blockchainenabled asynchronous federated learning in edge computing, Sensors, vol. 21, no. 10, p. 3335, 2021.
K. I. K. Wang, X. Zhou, W. Liang, Z. Yan, and J. She, Federated transfer learning based cross-domain prediction for smart manufacturing, IEEE Trans. Ind. Inf., vol. 18, no. 6, pp. 4088–4096, 2022.
X. D. Wang, J. Hu, H. Lin, W. X. Liu, H. Moon, and M. J. Piran, Federated learning-empowered disease diagnosis mechanism in the internet of medical things: From the privacy-preservation perspective, IEEE Trans. Ind. Inf., vol. 19, no. 7, pp. 7905–7913, 2023.
K. Sultana, K. Ahmed, B. Gu, and H. Wang, Elastic optimization for stragglers in edge federated learning, Big Data Mining and Analytics, vol. 6, no. 4, pp. 404–420, 2023.
T. Liang, J. Glossner, L. Wang, S. Shi, and X. Zhang, Pruning and quantization for deep neural network acceleration: A survey, Neurocomputing, vol. 461, pp. 370–403, 2021.
P. Kaywan, K. Ahmed, A. Ibaida, Y. Miao, and B. Gu, Early detection of depression using a conversational AI bot: A non-clinical trial, PLoS One, vol. 18, no. 2, p. e0279743, 2023.
J. Xu, B. S. Glicksberg, C. Su, P. Walker, J. Bian, and F. Wang, Federated learning for healthcare informatics, J. Healthc. Inform. Res., vol. 5, no. 1, pp. 1–19, 2021.