Offset-decoupled deformable convolution for efficient crowd counting
- PMID: 35851829
- PMCID: PMC9293988
- DOI: 10.1038/s41598-022-16415-9
Offset-decoupled deformable convolution for efficient crowd counting
Abstract
Crowd counting is considered a challenging issue in computer vision. One of the most critical challenges in crowd counting is considering the impact of scale variations. Compared with other methods, better performance is achieved with CNN-based methods. However, given the limit of fixed geometric structures, the head-scale features are not completely obtained. Deformable convolution with additional offsets is widely used in the fields of image classification and pattern recognition, as it can successfully exploit the potential of spatial information. However, owing to the randomly generated parameters of offsets in network initialization, the sampling points of the deformable convolution are disorderly stacked, weakening the effectiveness of feature extraction. To handle the invalid learning of offsets and the inefficient utilization of deformable convolution, an offset-decoupled deformable convolution (ODConv) is proposed in this paper. It can completely obtain information within the effective region of sampling points, leading to better performance. In extensive experiments, average MAE of 62.3, 8.3, 91.9, and 159.3 are achieved using our method on the ShanghaiTech A, ShanghaiTech B, UCF-QNRF, and UCF_CC_50 datasets, respectively, outperforming the state-of-the-art methods and validating the effectiveness of the proposed ODConv.
© 2022. The Author(s).
Conflict of interest statement
The authors declare no competing interests.
Figures









Similar articles
-
An Adaptive Multi-Scale Network Based on Depth Information for Crowd Counting.Sensors (Basel). 2023 Sep 11;23(18):7805. doi: 10.3390/s23187805. Sensors (Basel). 2023. PMID: 37765861 Free PMC article.
-
Crowd Counting Based on Multiscale Spatial Guided Perception Aggregation Network.IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):17465-17478. doi: 10.1109/TNNLS.2023.3304348. Epub 2024 Dec 2. IEEE Trans Neural Netw Learn Syst. 2024. PMID: 37610898
-
HADF-Crowd: A Hierarchical Attention-Based Dense Feature Extraction Network for Single-Image Crowd Counting.Sensors (Basel). 2021 May 17;21(10):3483. doi: 10.3390/s21103483. Sensors (Basel). 2021. PMID: 34067707 Free PMC article.
-
COMAL: compositional multi-scale feature enhanced learning for crowd counting.Multimed Tools Appl. 2022;81(15):20541-20560. doi: 10.1007/s11042-022-12249-9. Epub 2022 Mar 11. Multimed Tools Appl. 2022. PMID: 35291715 Free PMC article.
-
SPCANet: congested crowd counting via strip pooling combined attention network.PeerJ Comput Sci. 2024 Sep 18;10:e2273. doi: 10.7717/peerj-cs.2273. eCollection 2024. PeerJ Comput Sci. 2024. PMID: 39314741 Free PMC article.
Cited by
-
Multiscale regional calibration network for crowd counting.Sci Rep. 2025 Jan 22;15(1):2866. doi: 10.1038/s41598-025-86247-w. Sci Rep. 2025. PMID: 39843571 Free PMC article.
-
Apple Grading Based on Multi-Dimensional View Processing and Deep Learning.Foods. 2023 May 24;12(11):2117. doi: 10.3390/foods12112117. Foods. 2023. PMID: 37297365 Free PMC article.
References
-
- Q. Wang, J. Gao & W. Lin. NWPU-crowd: A large-scale benchmark for crowd counting and localization. in IEEE Transactions on Pattern Analysis and Machine Intelligence, 3013269 (2020). - PubMed
-
- V. A. Sindagi & V. M. Patel. Generating high-quality crowd density maps using contextual pyramid cnns. in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 1879–1888 (2017).
-
- Feris RS, Siddiquie B, Petterson J. Large-scale vehicle detection, indexing & search in urban surveillance videos. IEEE Trans. Multimed. 2012;14(1):28–42. doi: 10.1109/TMM.2011.2170666. - DOI
-
- Wang G, Li B, Zhang Y, Yang J. Background modeling and referencing for moving cameras-captured surveillance video coding in hevc. IEEE Trans. Multimed. 2018;20(11):2921–2934. doi: 10.1109/TMM.2018.2829163. - DOI
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources