Skip to main navigation Skip to search Skip to main content

Self-supervised Pre-training for Robust and Generic Spatial-Temporal Representations

  • Mingzhi Hu*
  • , Zhuoyun Zhong
  • , Xin Zhang
  • , Yanhua Li
  • , Yiqun Xie
  • , Xiaowei Jia
  • , Xun Zhou
  • , Jun Luo
  • *Corresponding author for this work
  • Worcester Polytechnic Institute
  • San Diego State University
  • University of Maryland, College Park
  • University of Pittsburgh
  • University of Iowa
  • Logistics and Supply Chain MultiTech R&D Centre

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Advancements in mobile sensing, data mining, and artificial intelligence have revolutionized the collection and analysis of Human-generated Spatial-Temporal Data (HSTD), paving the way for diverse applications across multiple domains. However, previous works have primarily focused on designing task-specific models for different problems, which lack transferability and generalizability when confronted with diverse HSTD. Additionally, these models often require a large amount of labeled data for optimal performance. While pre-trained models in Natural Language Processing (NLP) and Computer Vision (CV) domains have showcased impressive transferability and generalizability, similar efforts in the spatial-temporal data domain have been limited. In this paper, we take the lead and introduce the Spatial-Temporal Pre-Training model, i.e., STPT, which is connected with a self-supervised learning task, to address these limitations. STPT enables the creation of robust and versatile representations of HSTD. We validate our framework using real-world data and demonstrate its efficacy through two downstream tasks, i.e., trajectory classification and driving activity identification (e.g., identifying seeking vs. serving behaviors in taxi trajectories). Our results achieve an accuracy of 83.125% (16.2% higher than the average baseline) for human mobility identification and an accuracy of 77.88% (13.0% higher than the average baseline) for the human activity identification task. These outcomes underscore the potential of our pre-trained model for diverse downstream applications within the spatial-temporal data domain.

Original languageEnglish
Title of host publicationProceedings - 23rd IEEE International Conference on Data Mining, ICDM 2023
EditorsGuihai Chen, Latifur Khan, Xiaofeng Gao, Meikang Qiu, Witold Pedrycz, Xindong Wu
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages150-159
Number of pages10
ISBN (Electronic)9798350307887
DOIs
StatePublished - 2023
Externally publishedYes
Event23rd IEEE International Conference on Data Mining, ICDM 2023 - Shanghai, China
Duration: 1 Dec 20234 Dec 2023

Publication series

NameProceedings - IEEE International Conference on Data Mining, ICDM
ISSN (Print)1550-4786

Conference

Conference23rd IEEE International Conference on Data Mining, ICDM 2023
Country/TerritoryChina
CityShanghai
Period1/12/234/12/23

Keywords

  • driver identification
  • human decision analysis
  • pre-training
  • self-supervised learning
  • spatial-temporal data mining

Fingerprint

Dive into the research topics of 'Self-supervised Pre-training for Robust and Generic Spatial-Temporal Representations'. Together they form a unique fingerprint.

Cite this