消息
×
loading..
Federated Low-Rank Adaptation for Large Language Model Fine-Tuning Over Wireless Networks
2024-12-12
会议录名称GLOBECOM 2024 - 2024 IEEE GLOBAL COMMUNICATIONS CONFERENCE
ISSN1930-529X
页码3063-3068
发表状态已发表
DOI10.1109/GLOBECOM52923.2024.10901572
摘要Low-rank adaptation (LoRA) is an emerging fine-tuning method for personalized large language models (LLMs) due to its capability of achieving comparable learning performance to full fine-tuning by training a much smaller number of parameters. Federated fine-tuning (FedFT) combines LoRA with federated learning (FL) to enable collaborative fine-tuning of a global model with edge devices, leveraging distributed data while ensuring privacy. However, limited radio resources and computation capabilities of edge devices pose critical challenges on deploying FedFT over wireless networks. In this paper, we propose a split FedFT framework to separately deploy the computationally-intensive encoder of a pre-trained model at the edge server while reserving the embedding and the task modules at the edge devices, where the information exchanges between these modules are carried out over wireless networks. By exploiting the low-rank property of LoRA, the proposed FedFT framework reduces communication overhead by aggregating the gradient of the task module with respect to the output of a low-rank matrix. To enhance learning performance under stringent resource constraints, we formulate a joint device scheduling and bandwidth allocation problem while considering average transmission delay. By applying the Lyapunov technique, we decompose the formulated long-term mixed-integer programming (MIP) problem into sequential subproblems, followed by developing an online algorithm for effective device scheduling and bandwidth allocation. Simulation results demonstrate the effectiveness of our proposed online algorithm in enhancing learning performance.
关键词Edge computing Frequency allocation Integer programming Learning to rank Mixed-integer linear programming Problem oriented languages Resource allocation Critical challenges Device scheduling Distributed data Fine tuning Fine-tuning methods Global models Language model Learning performance On-line algorithms Radio resources
会议名称2024 IEEE Global Communications Conference, GLOBECOM 2024
会议地点Cape Town, South Africa
会议日期8-12 Dec. 2024
URL查看原文
收录类别EI
语种英语
出版者Institute of Electrical and Electronics Engineers Inc.
EI入藏号20251318125097
EI主题词Bandwidth
EISSN2576-6813
EI分类号716.1 Information Theory and Signal Processing ; 716.3 Radio Systems and Equipment ; 912.2 Management ; 1101.2 Machine Learning ; 1105 Computer Networks ; 1106.1 Computer Programming ; 1106.1.1 Computer Programming Languages ; 1201.7 Optimization Techniques
原始文献类型Conference article (CA)
来源库IEEE
文献类型会议论文
条目标识符https://kms.shanghaitech.edu.cn/handle/2MSLDSTB/500280
专题信息科学与技术学院
信息科学与技术学院_PI研究组_石远明组
信息科学与技术学院_PI研究组_周勇组
作者单位
1.Department of ECE, Hong Kong University of Science and Technology, Hong Kong, China
2.School of Information Science and Technology, ShanghaiTech University, Shanghai, China
推荐引用方式
GB/T 7714
Zixin Wang,Yong Zhou,Yuanming Shi,et al. Federated Low-Rank Adaptation for Large Language Model Fine-Tuning Over Wireless Networks[C]:Institute of Electrical and Electronics Engineers Inc.,2024:3063-3068.
条目包含的文件
条目无相关文件。
个性服务
查看访问统计
谷歌学术
谷歌学术中相似的文章
[Zixin Wang]的文章
[Yong Zhou]的文章
[Yuanming Shi]的文章
百度学术
百度学术中相似的文章
[Zixin Wang]的文章
[Yong Zhou]的文章
[Yuanming Shi]的文章
必应学术
必应学术中相似的文章
[Zixin Wang]的文章
[Yong Zhou]的文章
[Yuanming Shi]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。