消息
×
loading..
Federated Fine-Tuning for Pre-Trained Foundation Models Over Wireless Networks
2025
发表期刊IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS (IF:8.9[JCR-2023],8.6[5-Year])
ISSN1536-1276
EISSN1558-2248
卷号PP期号:99
发表状态已发表
DOI10.1109/TWC.2025.3531128
摘要

Pre-trained foundation models (FMs), with extensive number of neurons, are key to advancing next-generation intelligence services, where personalizing these models requires massive amount of task-specific data and computational resources. The prevalent solution involves centralized processing at the edge server, which, however, raises privacy concerns due to the transmission of raw data. Instead, federated fine-tuning (FedFT) is an emerging privacy-preserving fine-tuning (FT) paradigm for personalized pre-trained foundation models. In particular, by integrating low-rank adaptation (LoRA) with federated learning (FL), federated LoRA enables the collaborative FT of a global model with edge devices, achieving comparable learning performance to full FT while training fewer parameters over distributed data and preserving raw data privacy. However, the limited radio resources and computation capabilities of edge devices pose significant challenges for deploying 3 LoRA over wireless networks. To this paper, we propose a split federated LoRA framework, which deploys the computationally-intensive encoder of a pre-trained model at the edge server, while keeping the embedding and task modules at the edge devices. The information exchanges between these modules occur over wireless networks. Building on this split framework, the paper provides a rigorous analysis of the upper bound of the convergence gap for the wireless federated LoRA system. This analysis reveals the weighted impact of the number of edge devices participating in FedFT over all rounds, motivating the formulation of a long-term upper bound minimization problem. To address the long-term constraint, we decompose the formulated long-term mixed-integer programming (MIP) problem into sequential sub-problems using the Lyapunov technique. We then develop an online algorithm for effective device scheduling and bandwidth allocation. Simulation results demonstrate the effectiveness of the proposed online algorithm in enhancing learning performance. © 2025 IEEE.

关键词Adversarial machine learning Data privacy Differential privacy Edge computing Frequency allocation Integer programming Learning to rank Network embeddings Resource allocation Edge server Fine tuning Foundation models Intelligence services Learning performance On-line algorithms Parameter-efficient fine-tuning Pre-trained foundation model Resources allocation Upper Bound
URL查看原文
收录类别EI
语种英语
出版者Institute of Electrical and Electronics Engineers Inc.
EI入藏号20250617817146
EI主题词Federated learning
EI分类号1101.2 Machine Learning ; 1105 Computer Networks ; 1106.2 Data Handling and Data Processing ; 1108 Security and Privacy ; 1108.1 Cybersecurity ; 1201.7 Optimization Techniques ; 716.3 Radio Systems and Equipment ; 912.2 Management
原始文献类型Article in Press
来源库IEEE
文献类型期刊论文
条目标识符https://kms.shanghaitech.edu.cn/handle/2MSLDSTB/490309
专题信息科学与技术学院
信息科学与技术学院_PI研究组_石远明组
信息科学与技术学院_PI研究组_周勇组
通讯作者Zhou, Yong; Shi, Yuanming
作者单位
1.The Hong Kong University of Science and Technology, Department of Electronic and Computer Engineering, Hong Kong;
2.ShanghaiTech University, School of Information Science and Technology, Shanghai; 201210, China
通讯作者单位信息科学与技术学院
推荐引用方式
GB/T 7714
Wang, Zixin,Zhou, Yong,Shi, Yuanming,et al. Federated Fine-Tuning for Pre-Trained Foundation Models Over Wireless Networks[J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS,2025,PP(99).
APA Wang, Zixin,Zhou, Yong,Shi, Yuanming,&Letaief, Khaled B..(2025).Federated Fine-Tuning for Pre-Trained Foundation Models Over Wireless Networks.IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS,PP(99).
MLA Wang, Zixin,et al."Federated Fine-Tuning for Pre-Trained Foundation Models Over Wireless Networks".IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS PP.99(2025).
条目包含的文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
个性服务
查看访问统计
谷歌学术
谷歌学术中相似的文章
[Wang, Zixin]的文章
[Zhou, Yong]的文章
[Shi, Yuanming]的文章
百度学术
百度学术中相似的文章
[Wang, Zixin]的文章
[Zhou, Yong]的文章
[Shi, Yuanming]的文章
必应学术
必应学术中相似的文章
[Wang, Zixin]的文章
[Zhou, Yong]的文章
[Shi, Yuanming]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。