Progressive self-supervised learning: A pre-training method for crowd counting
2025-02
发表期刊PATTERN RECOGNITION LETTERS (IF:3.9[JCR-2023],4.2[5-Year])
ISSN0167-8655
卷号188页码:148-154
发表状态已发表
DOI10.1016/j.patrec.2024.12.007
摘要

Crowd counting technologies possess substantial social significance, and deep learning methods are increasingly seen as potent tools for advancing this field. Traditionally, many approaches have sought to enhance model performance by transferring knowledge from ImageNet, utilizing its classification weights to initialize models. However, the application of these pre-training weights is suboptimal for crowd counting, which involves dense prediction significantly different from image classification. To address these limitations, we introduce a progressive self-supervised learning approach, designed to generate more suitable pre-training weights from a large collection of density-related images. We gathered 173k images using custom-designed prompts and implemented a two-stage learning process to refine the feature representations of image patches with similar densities. In the first stage, mutual information between overlapping patches within the same image is maximized. Subsequently, a combination of global and local losses is evaluated to enhance feature similarity, with the latter assessing patches from different images of comparable densities. Our innovative pre-training approach demonstrated substantial improvements, reducing the Mean Absolute Error (MAE) by 7.5%, 17.6%, and 28.7% on the ShanghaiTech Part A & Part B and UCF_QNRF datasets respectively. Furthermore, when these pre-training weights were used to initialize existing models, such as CSRNet for density map regression and DM-Count for point supervision, a significant enhancement in performance was observed. © 2024 Elsevier B.V.

关键词Adversarial machine learning Contrastive Learning Federated learning Crowd counting Dataset construction Feature representation Images classification Learning methods Learning process Modeling performance Pre-training Supervised learning approaches Training methods
收录类别EI
语种英语
出版者Elsevier B.V.
EI入藏号20250217652941
EI主题词Self-supervised learning
EI分类号1101.2 ; 1101.2.1
原始文献类型Journal article (JA)
文献类型期刊论文
条目标识符https://kms.shanghaitech.edu.cn/handle/2MSLDSTB/483841
专题创意与艺术学院_PI研究组(P)_谢广平组
信息科学与技术学院_硕士生
创意与艺术学院_PI研究组(P)_武颖娜组
通讯作者Ni, Na
作者单位
Center for Adaptive System Engineering, ShanghaiTech University, Shanghai; 201210, China
第一作者单位上海科技大学
通讯作者单位上海科技大学
第一作者的第一单位上海科技大学
推荐引用方式
GB/T 7714
Gu, Yao,Zheng, Zhe,Wu, Yingna,et al. Progressive self-supervised learning: A pre-training method for crowd counting[J]. PATTERN RECOGNITION LETTERS,2025,188:148-154.
APA Gu, Yao,Zheng, Zhe,Wu, Yingna,Xie, Guangping,&Ni, Na.(2025).Progressive self-supervised learning: A pre-training method for crowd counting.PATTERN RECOGNITION LETTERS,188,148-154.
MLA Gu, Yao,et al."Progressive self-supervised learning: A pre-training method for crowd counting".PATTERN RECOGNITION LETTERS 188(2025):148-154.
条目包含的文件
文件名称/大小 文献类型 版本类型 开放类型 使用许可
个性服务
查看访问统计
谷歌学术
谷歌学术中相似的文章
[Gu, Yao]的文章
[Zheng, Zhe]的文章
[Wu, Yingna]的文章
百度学术
百度学术中相似的文章
[Gu, Yao]的文章
[Zheng, Zhe]的文章
[Wu, Yingna]的文章
必应学术
必应学术中相似的文章
[Gu, Yao]的文章
[Zheng, Zhe]的文章
[Wu, Yingna]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。