TY - GEN
T1 - Weakly-Supervised Crowdsourcing Localization Based on Self-Learning Spatial-Temporal Attention Network
AU - Xia, Huang
AU - Boateng, Gordon Owusu
AU - Si, Haonan
AU - Guo, Xiansheng
AU - Qian, Bocheng
AU - Liu, Xinhao
AU - Cao, Yu
AU - Liu, Yinong
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Crowdsourcing offers an effective solution to alleviate the heavy burden of data collection and annotation in fingerprint-based localization, particularly for Sixth-Generation (6G) communications. However, extracting robust and reliable features from sparsely labeled crowdsourced data remains highly challenging, significantly limiting localization accuracy. To address this issue, this paper proposes a Weakly-Supervised Crowd-sourcing Localization (WSCL) framework, which is based on a Self-learning Spatial-Temporal Attention Network (SSTAN). The framework consists of two phases: self-learning phase and fine-tuning phase. By applying spatial-temporal supervisory signals, a linear embedding attention encoder, and a linear projection layer, SSTAN extracts features rich in spatial-temporal information via self-learning. In the fine-tuning phase, a small amount of labeled data is used to learn the mapping relationship from the feature space to the location space, achieving satisfactory localization results. Experiments conducted in a real-world underground parking lot environment reveal that our proposed WSCL framework achieves a localization accuracy of 1.41m within an area of 1223m2 using only 10% labeled fingerprint samples.
AB - Crowdsourcing offers an effective solution to alleviate the heavy burden of data collection and annotation in fingerprint-based localization, particularly for Sixth-Generation (6G) communications. However, extracting robust and reliable features from sparsely labeled crowdsourced data remains highly challenging, significantly limiting localization accuracy. To address this issue, this paper proposes a Weakly-Supervised Crowd-sourcing Localization (WSCL) framework, which is based on a Self-learning Spatial-Temporal Attention Network (SSTAN). The framework consists of two phases: self-learning phase and fine-tuning phase. By applying spatial-temporal supervisory signals, a linear embedding attention encoder, and a linear projection layer, SSTAN extracts features rich in spatial-temporal information via self-learning. In the fine-tuning phase, a small amount of labeled data is used to learn the mapping relationship from the feature space to the location space, achieving satisfactory localization results. Experiments conducted in a real-world underground parking lot environment reveal that our proposed WSCL framework achieves a localization accuracy of 1.41m within an area of 1223m2 using only 10% labeled fingerprint samples.
KW - 6G
KW - crowdsourcing
KW - Indoor localization
KW - self-learning
KW - weakly-supervised learning
UR - https://www.scopus.com/pages/publications/105017675836
U2 - 10.1109/ICCCWorkshops67136.2025.11148152
DO - 10.1109/ICCCWorkshops67136.2025.11148152
M3 - Conference Proceeding
AN - SCOPUS:105017675836
T3 - 2025 IEEE/CIC International Conference on Communications in China, ICCC Workshops 2025
BT - 2025 IEEE/CIC International Conference on Communications in China, ICCC Workshops 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2025 IEEE/CIC International Conference on Communications in China, ICCC Workshops 2025
Y2 - 10 August 2025 through 13 August 2025
ER -