Publications
2025
- TKDEExploring Progress in Multivariate Time Series Forecasting: Comprehensive Benchmarking and Heterogeneity AnalysisZezhi Shao, Fei Wang*, Yongjun Xu*, Wei Wei, Chengqing Yu, Zhao Zhang, Di Yao, Tao Sun, Guangyin Jin, Xin Cao, and 3 more authorsIEEE Transactions on Knowledge and Data Engineering, Jan 2025
BasicTS+ has entered ESI Highly Cited Papers and acquired 1.3k+ Stars 2025 for the First, one of the Most Popular, fair and scalable benchmark of timeseries forecasting
@article{10726722, author = {Shao, Zezhi and Wang, Fei and Xu, Yongjun and Wei, Wei and Yu, Chengqing and Zhang, Zhao and Yao, Di and Sun, Tao and Jin, Guangyin and Cao, Xin and Cong, Gao and Jensen, Christian S. and Cheng, Xueqi}, journal = {IEEE Transactions on Knowledge and Data Engineering}, title = {Exploring Progress in Multivariate Time Series Forecasting: Comprehensive Benchmarking and Heterogeneity Analysis}, year = {2025}, month = jan, volume = {37}, number = {1}, pages = {291-305}, keywords = {Forecasting;Time series analysis;Benchmark testing;Transformers;Predictive models;Data models;Computer science;Reliability;Proposals;Electricity;Benchmarking;multivariate time series;spatial-temporal forecasting;long-term time series forecasting}, doi = {10.1109/TKDE.2024.3484454}, }
- Information FusionMGSFformer: A Multi-Granularity Spatiotemporal Fusion Transformer for air quality predictionChengqing Yu, Fei Wang*, Yilun Wang, Zezhi Shao, Tao Sun, Di Yao, and Yongjun Xu*Information Fusion, Jan 2025
Air quality spatiotemporal prediction can provide technical support for environmental governance and sustainable city development. As a classic multi-source spatiotemporal data, effective multi-source information fusion is key to achieving accurate air quality predictions. However, due to not fully fusing two pieces of information, classical deep learning models struggle to achieve satisfactory prediction results: (1) Multi-granularity: each air monitoring station collects air quality data at different sampling intervals, which show distinct time series patterns. (2) Spatiotemporal correlation: due to human activities and atmospheric diffusion, there exist correlations between air quality data from different air monitoring stations, necessitating the consideration of other air monitoring stations’ influences when modeling each air quality time series. In this study, to achieve satisfactory prediction results, we propose the Multi-Granularity Spatiotemporal Fusion Transformer, comprised of the residual de-redundant block, spatiotemporal attention block, and dynamic fusion block. Specifically, the residual de-redundant block eliminates information redundancy between data with different granularities and prevents the model from being misled by redundant information. The spatiotemporal attention block captures the spatiotemporal correlation of air quality data and facilitates prediction modeling. The dynamic fusion block evaluates the importance of data with different granularities and integrates the prediction results. Experimental results demonstrate that the proposed model surpasses 11 baselines by 5% in performance on three real-world datasets.
@article{YU2025102607, title = {MGSFformer: A Multi-Granularity Spatiotemporal Fusion Transformer for air quality prediction}, journal = {Information Fusion}, volume = {113}, pages = {102607}, year = {2025}, month = jan, issn = {1566-2535}, doi = {https://doi.org/10.1016/j.inffus.2024.102607}, url = {https://www.sciencedirect.com/science/article/pii/S1566253524003853}, author = {Yu, Chengqing and Wang, Fei and Wang, Yilun and Shao, Zezhi and Sun, Tao and Yao, Di and Xu, Yongjun}, keywords = {Air quality prediction, Multi-Granularity Spatiotemporal Fusion Transformer, Spatiotemporal correlation, Multi-source information fusion}, }
- PRTrajectory-User Linking via Multi-Scale Graph Attention NetworkYujie Li, Tao Sun, Zezhi Shao, Yiqiang Zhen, Yongjun Xu, and Fei Wang*Pattern Recognition, Feb 2025
Trajectory-User Linking (TUL) aims to link anonymous trajectories to their owners, which is considered an essential task in discovering human mobility patterns. Although existing TUL studies have shown promising results, they still have specific defects in the perception of spatio-temporal properties of trajectories, which manifested in the following three problems: missing context of the original trajectory, ignorance of spatial information, and high computational complexity. To address those issues, we revisit the characteristics of the trajectory and propose a novel model called TULMGAT (TUL via Multi-Scale Graph Attention Network) based on masked self-attention graph neural networks. Specifically, TULMGAT consists of four components: construction of check-in oriented graphs, node embedding, trajectory embedding, and trajectory user linking. Sufficient experiments on two publicly available datasets have shown that TULMGAT is the state-of-the-art model in task TUL compared to the baselines with an improvement of about 8% in accuracy and only a quarter of the fastest baseline in runtime. Furthermore, model validity experiments have verified the role of each module.
@article{LI2025110978, title = {Trajectory-User Linking via Multi-Scale Graph Attention Network}, journal = {Pattern Recognition}, volume = {158}, pages = {110978}, year = {2025}, month = feb, issn = {0031-3203}, doi = {https://doi.org/10.1016/j.patcog.2024.110978}, url = {https://www.sciencedirect.com/science/article/pii/S0031320324007295}, author = {Li, Yujie and Sun, Tao and Shao, Zezhi and Zhen, Yiqiang and Xu, Yongjun and Wang, Fei}, keywords = {Trajectory-user linking, Graph neural network, Trajectory classification, Spatio-temporal data mining, Check-in data}, }
- InnovationMetaCity: Data-driven sustainable development of complex citiesYunke Zhang, Yuming Lin, Guanjie Zheng, Yu Liu, Nicholas Sukiennik, Fengli Xu, Yongjun Xu, Feng Lu, Qi Wang, Yuan Lai, and 9 more authorsThe Innovation, Feb 2025
Cities are complex systems that develop under complicated interactions among their human and environmental components. Urbanization generates substantial outcomes and opportunities while raising challenges including congestion, air pollution, inequality, etc., calling for efficient and reasonable solutions to sustainable developments. Fortunately, booming technologies generate large-scale data of complex cities, providing a chance to propose data-driven solutions for sustainable urban developments. This paper provides a comprehensive overview of data-driven urban sustainability practice. In this review article, we conceptualize MetaCity, a general framework for optimizing resource usage and allocation problems in complex cities with data-driven approaches. Under this framework, we decompose specific urban sustainable goals, e.g., efficiency and resilience, review practical urban problems under these goals, and explore the probability of using data-driven technologies as potential solutions to the challenge of complexity. On the basis of extensive urban data, we integrate urban problem discovery, operation of urban systems simulation, and complex decision-making problem solving into an entire cohesive framework to achieve sustainable development goals by optimizing resource allocation problems in complex cities.
@article{ZHANG2025100775, title = {MetaCity: Data-driven sustainable development of complex cities}, journal = {The Innovation}, volume = {6}, number = {2}, pages = {100775}, year = {2025}, month = feb, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2024.100775}, url = {https://www.sciencedirect.com/science/article/pii/S2666675824002133}, author = {Zhang, Yunke and Lin, Yuming and Zheng, Guanjie and Liu, Yu and Sukiennik, Nicholas and Xu, Fengli and Xu, Yongjun and Lu, Feng and Wang, Qi and Lai, Yuan and Tian, Li and Li, Nan and Fang, Dongping and Wang, Fei and Zhou, Tao and Li, Yong and Zheng, Yu and Wu, Zhiqiang and Guo, Huadong}, keywords = {urban complex systems, sustainable development, data-driven methods, artificial intelligence}, }
- InnovationSpatial-temporal large models: A super hub linking multiple scientific areas with artificial intelligenceZezhi Shao, Tangwen Qian, Tao Sun, Fei Wang*, and Yongjun Xu*The Innovation, Feb 2025
@article{SHAO2025100763, title = {Spatial-temporal large models: A super hub linking multiple scientific areas with artificial intelligence}, journal = {The Innovation}, volume = {6}, number = {2}, pages = {100763}, year = {2025}, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2024.100763}, url = {https://www.sciencedirect.com/science/article/pii/S2666675824002017}, author = {Shao, Zezhi and Qian, Tangwen and Sun, Tao and Wang, Fei and Xu, Yongjun}, }
- InnovationToward the robustness of autonomous vehicles in the AI eraSiheng Chen*, Yiyi Liao*, Fei Wang*, Gang Wang*, Liang Wang*, Yafei Wang*, and Xichan Zhu*The Innovation, Mar 2025
@article{CHEN2025100780, title = {Toward the robustness of autonomous vehicles in the AI era}, journal = {The Innovation}, volume = {6}, number = {3}, pages = {100780}, year = {2025}, month = mar, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2024.100780}, url = {https://www.sciencedirect.com/science/article/pii/S2666675824002182}, author = {Chen, Siheng and Liao, Yiyi and Wang, Fei and Wang, Gang and Wang, Liang and Wang, Yafei and Zhu, Xichan}, }
- InnovationToward more economical large-scale foundation models: No longer a game for the fewYiqing Wu, Zhao Zhang*, Fei Wang, Yongjun Xu*, and Jincai HuangThe Innovation, Apr 2025
@article{WU2025100832, title = {Toward more economical large-scale foundation models: No longer a game for the few}, journal = {The Innovation}, volume = {6}, number = {4}, pages = {100832}, year = {2025}, month = apr, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2025.100832}, url = {https://www.sciencedirect.com/science/article/pii/S2666675825000359}, author = {Wu, Yiqing and Zhang, Zhao and Wang, Fei and Xu, Yongjun and Huang, Jincai}, }
- AAAIEditing Memories Through Few Targeted NeuronsWei Zhou, Wei Wei, Guibang Cao, and Fei WangProceedings of the AAAI Conference on Artificial Intelligence, Apr 2025
@article{Zhou_Wei_Cao_Wang_2025, title = {Editing Memories Through Few Targeted Neurons}, volume = {39}, url = {https://ojs.aaai.org/index.php/AAAI/article/view/34807}, doi = {10.1609/aaai.v39i24.34807}, abstractn = {Model editing is a novel research topic in large language models (LLMs), aimed at efficiently handling various knowledge editing tasks. Since irrelevant knowledge is difficult to measure,existing editing methods often lack explicit ways to preserve it, especially for editing methods based on the fine-tuning paradigm. They generally control the locality performance of model editing by constraining the range of changes in model parameters. However, their performance improvements are not always ideal, and may even lead to a decrease in the editing reliability. In this paper, we try to explore effective editing locality control methods based on the relationship between the stored knowledge and the strongly associated model components. Based on the discovery of ``knowledge neurons’’ and enough experimental results, we further explore the potential characteristics between knowledge and model components, confirm and point out: (1) only 1% neurons have significant contributions to specific knowledge storage, and (2) these targeted neurons often have a high overlap for knowledge with similar relational descriptions, which means that knowledge with similar relationships may be severely affected when these targeted neurons are modified. Based on these findings, we propose Targeted Neurons Fine-tuning with Data Augmentation (TNF-DA), which performs data augmentation based on the relational representation of edited knowledge to improve editing locality. By freezing most of the model parameters and only fine-tuning the highly contributing neurons corresponding to the edited knowledge, we obtain desirable results in terms of generalization and specificity compared with previous fine-tuning-based methods. Extensive experiments have demonstrated the superior editing performance achieved by our proposed method.}, number = {24}, journal = {Proceedings of the AAAI Conference on Artificial Intelligence}, author = {Zhou, Wei and Wei, Wei and Cao, Guibang and Wang, Fei}, year = {2025}, month = apr, pages = {26111-26119}, }
- AAAIMPQ-DM: Mixed Precision Quantization for Extremely Low Bit Diffusion ModelsWeilun Feng, Haotong Qin, Chuanguang Yang, Zhulin An, Libo Huang, Boyu Diao, Fei Wang, Renshuai Tao, Yongjun Xu, and Michele MagnoProceedings of the AAAI Conference on Artificial Intelligence, Apr 2025
Diffusion models have received wide attention in generation tasks. However, the expensive computation cost prevents the application of diffusion models in resource-constrained scenarios. Quantization emerges as a practical solution that significantly saves storage and computation by reducing the bit-width of parameters. However, the existing quantization methods for diffusion models still cause severe degradation in performance, especially under extremely low bit-widths (2-4 bit). The primary decrease in performance comes from the significant discretization of activation values at low bit quantization. Too few activation candidates are unfriendly for outlier significant weight channel quantization, and the discretized features prevent stable learning over different time steps of the diffusion model. This paper presents MPQ-DM, a Mixed-Precision Quantization method for Diffusion Models. The proposed MPQ-DM mainly relies on two techniques: (1) To mitigate the quantization error caused by outlier severe weight channels, we propose an Outlier-Driven Mixed Quantization (OMQ) technique that uses Kurtosis to quantify outlier salient channels and apply optimized intra-layer mixed-precision bit-width allocation to recover accuracy performance within target efficiency. (2) To robustly learn representations crossing time steps, we construct a Time-Smoothed Relation Distillation (TRD) scheme between the quantized diffusion model and its full-precision counterpart, transferring discrete and continuous latent to a unified relation space to reduce the representation inconsistency. Comprehensive experiments demonstrate that MPQ-DM achieves significant accuracy gains under extremely low bit-widths compared with SOTA quantization methods. MPQ-DM achieves a 58% FID decrease under W2A4 setting compared with baseline, while all other methods even collapse.
@article{Feng_Qin_Yang_An_Huang_Diao_Wang_Tao_Xu_Magno_2025, title = {MPQ-DM: Mixed Precision Quantization for Extremely Low Bit Diffusion Models}, volume = {39}, url = {https://ojs.aaai.org/index.php/AAAI/article/view/33823}, doi = {10.1609/aaai.v39i16.33823}, number = {16}, journal = {Proceedings of the AAAI Conference on Artificial Intelligence}, author = {Feng, Weilun and Qin, Haotong and Yang, Chuanguang and An, Zhulin and Huang, Libo and Diao, Boyu and Wang, Fei and Tao, Renshuai and Xu, Yongjun and Magno, Michele}, year = {2025}, month = apr, pages = {16595-16603}, }
- JCSTA Model-Agnostic Hierarchical Framework Towards Trajectory PredictionTang-Wen Qian, Yuan Wang, Yong-Jun Xu, Zhao Zhang, Lin Wu, Qiang Qiu, and Fei Wang*Journal of Computer Science and Technology, May 2025
Predicting the future trajectories of multiple agents is essential for various applications in real life, such as surveillance systems, autonomous driving, and social robots. The trajectory prediction task is influenced by many factors, including the individual historical trajectory, interactions between agents, and the fuzzy nature of the observed agents’ motion. While existing methods have made great progress on the topic of trajectory prediction, they treat all the information uniformly, which limits the effectiveness of information utilization. To this end, in this paper, we propose and utilize a model-agnostic framework to regard all the information in a two-level hierarchical view. Particularly, the first-level view is the inter-trajectory view. In this level, we observe that the difficulty in predicting different trajectory samples varies. We define trajectory difficulty and train the proposed framework in an “easy-to-hard” schema. The second-level view is the intra-trajectory level. We find the influencing factors for a particular trajectory can be divided into two parts. The first part is global features, which keep stable within a trajectory, i.e., the expected destination. The second part is local features, which change over time, i.e., the current position. We believe that the two types of information should be handled in different ways. The hierarchical view is beneficial to take full advantage of the information in a fine-grained way. Experimental results validate the effectiveness of the proposed model-agnostic framework.
@article{JCST-2212-13013, title = {A Model-Agnostic Hierarchical Framework Towards Trajectory Prediction}, journal = {Journal of Computer Science and Technology}, volume = {40}, number = {2}, pages = {322-339}, year = {2025}, month = may, issn = {1000-9000(Print) /1860-4749(Online)}, doi = {10.1007/s11390-023-3013-4}, url = {https://jcst.ict.ac.cn/en/article/doi/10.1007/s11390-023-3013-4}, author = {Qian, Tang-Wen and Wang, Yuan and Xu, Yong-Jun and Zhang, Zhao and Wu, Lin and Qiu, Qiang and Wang, Fei}, }
- TKDEAdaE: Knowledge Graph Embedding with Adaptive Embedding SizesZhanpeng Guan, Fuwei Zhang, Zhao Zhang, Fuzhen Zhuang, Fei Wang, Zhulin An, and Yongjun XuIEEE Transactions on Knowledge and Data Engineering, May 2025
@article{10981648, author = {Guan, Zhanpeng and Zhang, Fuwei and Zhang, Zhao and Zhuang, Fuzhen and Wang, Fei and An, Zhulin and Xu, Yongjun}, journal = {IEEE Transactions on Knowledge and Data Engineering}, title = {AdaE: Knowledge Graph Embedding with Adaptive Embedding Sizes}, year = {2025}, volume = {}, number = {}, pages = {1-14}, keywords = {Knowledge graphs;Adaptation models;Training;Data models;Search problems;Vectors;Overfitting;Tail;Optimization;Tensors;Knowledge graph embedding;Data imbalance issue;Dimension selection}, doi = {10.1109/TKDE.2025.3566270}, }
- TKDEGinAR+: A Robust End-To-End Framework for Multivariate Time Series Forecasting with Missing ValuesChengqing Yu, Fei Wang*, Zezhi Shao, Tangwen Qian, Zhao Zhang, Wei Wei, Zhulin An, Qi Wang, and Yongjun XuIEEE Transactions on Knowledge and Data Engineering, May 2025
@article{11002729, author = {Yu, Chengqing and Wang, Fei and Shao, Zezhi and Qian, Tangwen and Zhang, Zhao and Wei, Wei and An, Zhulin and Wang, Qi and Xu, Yongjun}, journal = {IEEE Transactions on Knowledge and Data Engineering}, title = {GinAR+: A Robust End-To-End Framework for Multivariate Time Series Forecasting with Missing Values}, year = {2025}, month = may, volume = {}, number = {}, pages = {1-14}, keywords = {Correlation;Predictive models;Forecasting;Time series analysis;Data models;Robustness;Adaptation models;Imputation;Contrastive learning;Training;Contrastive learning;Graph interpolation attention recursive network;Multivariate Time Series Forecasting with Missing Values}, doi = {10.1109/TKDE.2025.3569649}, }
- InnovationFoundation models and intelligent decision-making: Progress, challenges, and perspectivesJincai Huang†, Yongjun Xu†, Qi Wang†, Qi (Cheems) Wang†, Xingxing Liang†, Fei Wang†, Zhao Zhang†, Wei Wei†, Boxuan Zhang†, Libo Huang†, and 61 more authorsThe Innovation, Jun 2025
Intelligent decision-making (IDM) is a cornerstone of artificial intelligence (AI) designed to automate or augment decision processes. Modern IDM paradigms integrate advanced frameworks to enable intelligent agents to make effective and adaptive choices and decompose complex tasks into manageable steps, such as AI agents and high-level reinforcement learning. Recent advances in multimodal foundation-based approaches unify diverse input modalities—such as vision, language, and sensory data—into a cohesive decision-making process. Foundation models (FMs) have become pivotal in science and industry, transforming decision-making and research capabilities. Their large-scale, multimodal data-processing abilities foster adaptability and interdisciplinary breakthroughs across fields such as healthcare, life sciences, and education. This survey examines IDM’s evolution, advanced paradigms with FMs and their transformative impact on decision-making across diverse scientific and industrial domains, highlighting the challenges and opportunities in building efficient, adaptive, and ethical decision systems.
@article{HUANG2025100948, title = {Foundation models and intelligent decision-making: Progress, challenges, and perspectives}, journal = {The Innovation}, volume = {6}, number = {6}, pages = {100948}, year = {2025}, month = jun, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2025.100948}, url = {https://www.sciencedirect.com/science/article/pii/S2666675825001511}, author = {Huang, Jincai and Xu, Yongjun and Wang, Qi and Wang, Qi (Cheems) and Liang, Xingxing and Wang, Fei and Zhang, Zhao and Wei, Wei and Zhang, Boxuan and Huang, Libo and Chang, Jingru and Ma, Liantao and Ma, Ting and Liang, Yuxuan and Zhang, Jie and Guo, Jian and Jiang, Xuhui and Fan, Xinxin and An, Zhulin and Li, Tingting and Li, Xuefei and Shao, Zezhi and Qian, Tangwen and Sun, Tao and Diao, Boyu and Yang, Chuanguang and Yu, Chenqing and Wu, Yiqing and Li, Mengxian and Zhang, Haifeng and Zeng, Yongcheng and Zhang, Zhicheng and Zhu, Zhengqiu and Lv, Yiqin and Li, Aming and Chen, Xu and An, Bo and Xiao, Wei and Bai, Chenguang and Mao, Yuxing and Yin, Zhigang and Gui, Sheng and Su, Wentao and Zhu, Yinghao and Gao, Junyi and He, Xinyu and Li, Yizhou and Jin, Guangyin and Ao, Xiang and Zhai, Xuehao and Tan, Haoran and Yun, Lijun and Shi, Hongquan and Li, Jun and Fan, Changjun and Huang, Kuihua and Harrison, Ewen and Leung, Victor C.M. and Qiu, Sihang and Dong, Yanjie and Zheng, Xiaolong and Wang, Gang and Zheng, Yu and Wang, Yuanzhuo and Guo, Jiafeng and Wang, Lizhe and Cheng, Xueqi and Wang, Yaonan and Yang, Shanlin and Fu, Mengyin and Fei, Aiguo}, keywords = {artificial intelligence, intelligent decision-making, foundation models, agent, large language model}, }
- KDDBLAST: Balanced Sampling Time Series Corpus for Universal Forecasting ModelsZezhi Shao, Yujie Li, Fei Wang*, Chengqing Yu, Yisong Fu, Tangwen Qian, Bin Xu, Boyu Diao, Yongjun Xu, and Xueqi ChengIn Proceedings of the 31th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Toronton, ON, Canada, Aug 2025
The advent of universal time series forecasting models has revolutionized zero-shot forecasting across diverse domains, yet the critical role of data diversity in training these models remains underexplored. Existing large-scale time series datasets often suffer from inherent biases and imbalanced distributions, leading to suboptimal model performance and generalization. To address this gap, we introduce BLAST, a novel pre-training corpus designed to enhance data diversity through a balanced sampling strategy. First, BLAST incorporates 321 billion observations from publicly available datasets and employs a comprehensive suite of statistical metrics to characterize time series patterns. Then, to facilitate pattern-oriented sampling, the data is implicitly clustered using grid-based partitioning. Furthermore, by integrating grid sampling and grid mixup techniques, BLAST ensures a balanced and representative coverage of diverse patterns. Experimental results demonstrate that models pre-trained on BLAST achieve state-of-the-art performance with a fraction of the computational resources and training tokens required by existing methods. Our findings highlight the pivotal role of data diversity in improving both training efficiency and model performance for the universal forecasting task.
@inproceedings{blast, author = {Shao, Zezhi and Li, Yujie and Wang, Fei and Yu, Chengqing and Fu, Yisong and Qian, Tangwen and Xu, Bin and Diao, Boyu and Xu, Yongjun and Cheng, Xueqi}, title = {BLAST: Balanced Sampling Time Series Corpus for Universal Forecasting Models}, year = {2025}, month = aug, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, booktitle = {Proceedings of the 31th ACM SIGKDD Conference on Knowledge Discovery and Data Mining}, keywords = {large-scale time series dataset, balanced sampling, universal time series forecasting}, location = {Toronton, ON, Canada}, series = {KDD '25}, }
2024
- ICASSPDynamic Frequency Domain Graph Convolutional Network for Traffic ForecastingYujie Li, Zezhi Shao, Yongjun Xu, Qiang Qiu, Zhaogang Cao, and Fei WangIn ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Apr 2024
@inproceedings{10446144, author = {Li, Yujie and Shao, Zezhi and Xu, Yongjun and Qiu, Qiang and Cao, Zhaogang and Wang, Fei}, booktitle = {ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)}, title = {Dynamic Frequency Domain Graph Convolutional Network for Traffic Forecasting}, year = {2024}, month = apr, volume = {}, number = {}, pages = {5245-5249}, keywords = {Convolution;Frequency-domain analysis;Time series analysis;Transportation;Traffic control;Spatial databases;Sensors;Traffic prediction;frequency domain signal processing;multivariate time series analysis;dynamic graph learning;graph convolution}, doi = {10.1109/ICASSP48485.2024.10446144}, }
- COLINGSelf-Improvement Programming for Temporal Knowledge Graph Question AnsweringZhuo Chen, Zhao Zhang, Zixuan Li, Fei Wang, Yutao Zeng, Xiaolong Jin, and Yongjun XuIn Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), May 2024
Temporal Knowledge Graph Question Answering (TKGQA) aims to answer questions with temporal intent over Temporal Knowledge Graphs (TKGs). The core challenge of this task lies in understanding the complex semantic information regarding multiple types of time constraints (e.g., before, first) in questions. Existing end-to-end methods implicitly model the time constraints by learning time-aware embeddings of questions and candidate answers, which is far from understanding the question comprehensively. Motivated by semantic-parsing-based approaches that explicitly model constraints in questions by generating logical forms with symbolic operators, we design fundamental temporal operators for time constraints and introduce a novel self-improvement Programming method for TKGQA (Prog-TQA). Specifically, Prog-TQA leverages the in-context learning ability of Large Language Models (LLMs) to understand the combinatory time constraints in the questions and generate corresponding program drafts with a few examples given. Then, it aligns these drafts to TKGs with the linking module and subsequently executes them to generate the answers. To enhance the ability to understand questions, Prog-TQA is further equipped with a self-improvement strategy to effectively bootstrap LLMs using high-quality self-generated drafts. Extensive experiments demonstrate the superiority of the proposed Prog-TQA on MultiTQ and CronQuestions datasets, especially in the Hits@1 metric.
@inproceedings{chen-etal-2024-self, title = {Self-Improvement Programming for Temporal Knowledge Graph Question Answering}, author = {Chen, Zhuo and Zhang, Zhao and Li, Zixuan and Wang, Fei and Zeng, Yutao and Jin, Xiaolong and Xu, Yongjun}, editor = {Calzolari, Nicoletta and Kan, Min-Yen and Hoste, Veronique and Lenci, Alessandro and Sakti, Sakriani and Xue, Nianwen}, booktitle = {Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)}, month = may, year = {2024}, address = {Torino, Italia}, publisher = {ELRA and ICCL}, url = {https://aclanthology.org/2024.lrec-main.1270/}, pages = {14579--14594}, }
- ICDEAdapTraj: A Multi-Source Domain Generalization Framework for Multi-Agent Trajectory PredictionTangwen Qian, Yile Chen, Gao Cong, Yongjun Xu, and Fei Wang*In 2024 IEEE 40th International Conference on Data Engineering (ICDE), May 2024
@inproceedings{10598115, author = {Qian, Tangwen and Chen, Yile and Cong, Gao and Xu, Yongjun and Wang, Fei}, booktitle = {2024 IEEE 40th International Conference on Data Engineering (ICDE)}, title = {AdapTraj: A Multi-Source Domain Generalization Framework for Multi-Agent Trajectory Prediction}, year = {2024}, month = may, volume = {}, number = {}, pages = {5048-5060}, keywords = {Degradation;Adaptation models;Buildings;Predictive models;Data engineering;Data models;Trajectory;multi-agent trajectory prediction;multi-source domain generalization;distribution shift}, doi = {10.1109/ICDE60146.2024.00113}, }
- KDDGinAR: An End-To-End Multivariate Time Series Forecasting Model Suitable for Variable MissingChengqing Yu, Fei Wang*, Zezhi Shao, Tangwen Qian, Zhao Zhang, Wei Wei, and Yongjun Xu*In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Barcelona, Spain, Aug 2024
Multivariate time series forecasting (MTSF) is crucial for decision-making to precisely forecast the future values/trends, based on the complex relationships identified from historical observations of multiple sequences. Recently, Spatial-Temporal Graph Neural Networks (STGNNs) have gradually become the theme of MTSF model as their powerful capability in mining spatial-temporal dependencies, but almost of them heavily rely on the assumption of historical data integrity. In reality, due to factors such as data collector failures and time-consuming repairment, it is extremely challenging to collect the whole historical observations without missing any variable. In this case, STGNNs can only utilize a subset of normal variables and easily suffer from the incorrect spatial-temporal dependency modeling issue, resulting in the degradation of their forecasting performance. To address the problem, in this paper, we propose a novel Graph Interpolation Attention Recursive Network (named GinAR) to precisely model the spatial-temporal dependencies over the limited collected data for forecasting. In GinAR, it consists of two key components, that is, interpolation attention and adaptive graph convolution to take place of the fully connected layer of simple recursive units, and thus are capable of recovering all missing variables and reconstructing the correct spatial-temporal dependencies for recursively modeling of multivariate time series data, respectively. Extensive experiments conducted on five real-world datasets demonstrate that GinAR outperforms 11 SOTA baselines, and even when 90% of variables are missing, it can still accurately predict the future values of all variables.
@inproceedings{10.1145/3637528.3672055, author = {Yu, Chengqing and Wang, Fei and Shao, Zezhi and Qian, Tangwen and Zhang, Zhao and Wei, Wei and Xu, Yongjun}, title = {GinAR: An End-To-End Multivariate Time Series Forecasting Model Suitable for Variable Missing}, year = {2024}, month = aug, isbn = {9798400704901}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3637528.3672055}, doi = {10.1145/3637528.3672055}, booktitle = {Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining}, pages = {3989-4000}, numpages = {12}, keywords = {adaptive graph convolution, graph interpolation attention recursive network, interpolation attention, multivariate time series forecasting, variable missing}, location = {Barcelona, Spain}, series = {KDD '24}, }
- InnovationArtificial intelligence for geoscience: Progress, challenges, and perspectivesTianjie Zhao†, Sheng Wang†, Chaojun Ouyang†, Min Chen†, Chenying Liu†, Jin Zhang†, Long Yu†, Fei Wang†, Yong Xie†, Jun Li†, and 41 more authorsThe Innovation, Sep 2024
This paper explores the evolution of geoscientific inquiry, tracing the progression from traditional physics-based models to modern data-driven approaches facilitated by significant advancements in artificial intelligence (AI) and data collection techniques. Traditional models, which are grounded in physical and numerical frameworks, provide robust explanations by explicitly reconstructing underlying physical processes. However, their limitations in comprehensively capturing Earth’s complexities and uncertainties pose challenges in optimization and real-world applicability. In contrast, contemporary data-driven models, particularly those utilizing machine learning (ML) and deep learning (DL), leverage extensive geoscience data to glean insights without requiring exhaustive theoretical knowledge. ML techniques have shown promise in addressing Earth science-related questions. Nevertheless, challenges such as data scarcity, computational demands, data privacy concerns, and the “black-box” nature of AI models hinder their seamless integration into geoscience. The integration of physics-based and data-driven methodologies into hybrid models presents an alternative paradigm. These models, which incorporate domain knowledge to guide AI methodologies, demonstrate enhanced efficiency and performance with reduced training data requirements. This review provides a comprehensive overview of geoscientific research paradigms, emphasizing untapped opportunities at the intersection of advanced AI techniques and geoscience. It examines major methodologies, showcases advances in large-scale models, and discusses the challenges and prospects that will shape the future landscape of AI in geoscience. The paper outlines a dynamic field ripe with possibilities, poised to unlock new understandings of Earth’s complexities and further advance geoscience exploration.
@article{ZHAO2024100691, title = {Artificial intelligence for geoscience: Progress, challenges, and perspectives}, journal = {The Innovation}, volume = {5}, number = {5}, pages = {100691}, year = {2024}, month = sep, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2024.100691}, url = {https://www.sciencedirect.com/science/article/pii/S2666675824001292}, author = {Zhao, Tianjie and Wang, Sheng and Ouyang, Chaojun and Chen, Min and Liu, Chenying and Zhang, Jin and Yu, Long and Wang, Fei and Xie, Yong and Li, Jun and Wang, Fang and Grunwald, Sabine and Wong, Bryan M. and Zhang, Fan and Qian, Zhen and Xu, Yongjun and Yu, Chengqing and Han, Wei and Sun, Tao and Shao, Zezhi and Qian, Tangwen and Chen, Zhao and Zeng, Jiangyuan and Zhang, Huai and Letu, Husi and Zhang, Bing and Wang, Li and Luo, Lei and Shi, Chong and Su, Hongjun and Zhang, Hongsheng and Yin, Shuai and Huang, Ni and Zhao, Wei and Li, Nan and Zheng, Chaolei and Zhou, Yang and Huang, Changping and Feng, Defeng and Xu, Qingsong and Wu, Yan and Hong, Danfeng and Wang, Zhenyu and Lin, Yinyi and Zhang, Tangtang and Kumar, Prashant and Plaza, Antonio and Chanussot, Jocelyn and Zhang, Jiabao and Shi, Jiancheng and Wang, Lizhe}, keywords = {artificial intelligence, machine learning, deep learning, geoscience}, }
- InnovationArtificial intelligence is restructuring a new worldYongjun Xu*, Fei Wang*, and Tangtang Zhang*The Innovation, Nov 2024
@article{XU2024100725, title = {Artificial intelligence is restructuring a new world}, journal = {The Innovation}, volume = {5}, number = {6}, pages = {100725}, year = {2024}, month = nov, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2024.100725}, url = {https://www.sciencedirect.com/science/article/pii/S2666675824001632}, author = {Xu, Yongjun and Wang, Fei and Zhang, Tangtang}, }
2023
- CIKMDSformer: A Double Sampling Transformer for Multivariate Time Series Long-term PredictionChengqing Yu, Fei Wang*, Zezhi Shao, Tao Sun, Lin Wu, and Yongjun XuIn Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, United Kingdom, Oct 2023
The 2nd-Most Cited paper in CIKM 2023 2023 2/676
Multivariate time series long-term prediction, which aims to predict the change of data in a long time, can provide references for decision-making. Although transformer-based models have made progress in this field, they usually do not make full use of three features of multivariate time series: global information, local information, and variables correlation. To effectively mine the above three features and establish a high-precision prediction model, we propose a double sampling transformer (DSformer), which consists of the double sampling (DS) block and the temporal variable attention (TVA) block. Firstly, the DS block employs down sampling and piecewise sampling to transform the original series into feature vectors that focus on global information and local information respectively. Then, TVA block uses temporal attention and variable attention to mine these feature vectors from different dimensions and extract key information. Finally, based on a parallel structure, DSformer uses multiple TVA blocks to mine and integrate different features obtained from DS blocks respectively. The integrated feature information is passed to the generative decoder based on a multi-layer perceptron to realize multivariate time series long-term prediction. Experimental results on nine real-world datasets show that DSformer can outperform eight existing baselines.
@inproceedings{10.1145/3583780.3614851, author = {Yu, Chengqing and Wang, Fei and Shao, Zezhi and Sun, Tao and Wu, Lin and Xu, Yongjun}, title = {DSformer: A Double Sampling Transformer for Multivariate Time Series Long-term Prediction}, year = {2023}, month = oct, isbn = {9798400701245}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3583780.3614851}, doi = {10.1145/3583780.3614851}, booktitle = {Proceedings of the 32nd ACM International Conference on Information and Knowledge Management}, pages = {3062-3072}, numpages = {11}, keywords = {double sampling transformer, multivariate time series long-term prediction, temporal variable attention block}, location = {Birmingham, United Kingdom}, series = {CIKM '23}, }
- TKDEHeterogeneous Graph Neural Network With Multi-View Representation LearningZezhi Shao, Yongjun Xu, Wei Wei, Fei Wang, Zhao Zhang, and Feida ZhuIEEE Transactions on Knowledge and Data Engineering, Nov 2023
@article{9961953, author = {Shao, Zezhi and Xu, Yongjun and Wei, Wei and Wang, Fei and Zhang, Zhao and Zhu, Feida}, journal = {IEEE Transactions on Knowledge and Data Engineering}, title = {Heterogeneous Graph Neural Network With Multi-View Representation Learning}, year = {2023}, month = nov, volume = {35}, number = {11}, pages = {11476-11488}, keywords = {Semantics;Mercury (metals);Graph neural networks;Aggregates;Task analysis;Representation learning;Adaptation models;Heterogeneous graphs;graph neural networks;graph embedding}, doi = {10.1109/TKDE.2022.3224193}, }
- CIKMClustering-property Matters: A Cluster-aware Network for Large Scale Multivariate Time Series ForecastingYuan Wang, Zezhi Shao, Tao Sun, Chengqing Yu, Yongjun Xu, and Fei Wang*In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, United Kingdom, Oct 2023
Large-scale Multivariate Time Series(MTS) widely exist in various real-world systems, imposing significant demands on model efficiency. A recent work, STID, addressed the high complexity issue of popular Spatial-Temporal Graph Neural Networks(STGNNs). Despite its success, when applied to large-scale MTS data, the number of parameters of STID for modeling spatial dependencies increases substantially, leading to over-parameterization issues and suboptimal performance. These observations motivate us to explore new approaches for modeling spatial dependencies in a parameter-friendly manner. In this paper, we argue that the spatial properties of variables are essentially the superposition of multiple cluster centers. Accordingly, we propose a Cluster-Aware Network(CANet), which effectively captures spatial dependencies by mining the implicit cluster centers of variables. CANet solely optimizes the cluster centers instead of the spatial information of all nodes, thereby significantly reducing the parameter amount. Extensive experiments on two large-scale datasets validate our motivation and demonstrate the superiority of CANet.
@inproceedings{10.1145/3583780.3615253, author = {Wang, Yuan and Shao, Zezhi and Sun, Tao and Yu, Chengqing and Xu, Yongjun and Wang, Fei}, title = {Clustering-property Matters: A Cluster-aware Network for Large Scale Multivariate Time Series Forecasting}, year = {2023}, month = oct, isbn = {9798400701245}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3583780.3615253}, doi = {10.1145/3583780.3615253}, booktitle = {Proceedings of the 32nd ACM International Conference on Information and Knowledge Management}, pages = {4340-4344}, numpages = {5}, keywords = {multivariate time series forecasting, large-scale, cluster centers}, location = {Birmingham, United Kingdom}, series = {CIKM '23}, }
- InnovationAI-enhanced spatial-temporal data-mining technology: New chance for next-generation urban computingFei Wang*, Di Yao*, Yong Li*, Tao Sun, and Zhao ZhangThe Innovation, Mar 2023
@article{WANG2023100405, title = {AI-enhanced spatial-temporal data-mining technology: New chance for next-generation urban computing}, journal = {The Innovation}, volume = {4}, number = {2}, pages = {100405}, year = {2023}, month = mar, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2023.100405}, url = {https://www.sciencedirect.com/science/article/pii/S2666675823000334}, author = {Wang, Fei and Yao, Di and Li, Yong and Sun, Tao and Zhang, Zhao}, }
2022
- DASFAAHuman Mobility Identification by Deep Behavior Relevant Location RepresentationTao Sun, Fei Wang, Zhao Zhang, Lin Wu, and Yongjun XuIn Database Systems for Advanced Applications, Apr 2022
Fei Wang receveid the Best Student Paper Award 2022 for the contribution to method of Deep Behavior Relevant Location Representation
This paper focuses on Trajectory User Link (TUL), which aims at identifying user identities through exploiting their mobility patterns. Existing TUL approaches are based on location representation, a way to learn location associations by embedding vectors that can indicate the level of semantic similarity between the locations. However, existing methods for location representation don’t consider the semantic diversity of locations, which will lead to a misunderstanding of the semantic information of trajectory when linking anonymous trajectories to candidate users. To solve this problem, in this paper, we propose Deep Behavior Relevant Location representation (DBRLr) to map the polysemous locations into distinct vectors, from the perspective of users’ behavior to reflect the semantic polysemy of locations. To learn this representation, we build a Location Prediction-based Movement Model (LP-based MM), which learns user behavior representation at each visited location from a large history trajectory corpora. LP-based MM considers both Continuity and Cyclicity characteristics of user’s movement. We employ the combination of the intermediate layer representation in LP-based MM as DBRLr. An effective recurrent neural network is used to link anonymous trajectories with candidate users. Experiments are conducted on two real-world datasets, and the result shows that our method performs beyond existing methods.
@inproceedings{10.1007/978-3-031-00126-0_33, author = {Sun, Tao and Wang, Fei and Zhang, Zhao and Wu, Lin and Xu, Yongjun}, title = {Human Mobility Identification by Deep Behavior Relevant Location Representation}, booktitle = {Database Systems for Advanced Applications}, year = {2022}, month = apr, publisher = {Springer International Publishing}, address = {Cham}, pages = {439--454}, isbn = {978-3-031-00126-0}, }
- VLDBDecoupled Dynamic Spatial-Temporal Graph Neural Network for Traffic ForecastingZezhi Shao, Zhao Zhang, Wei Wei*, Fei Wang*, Yongjun Xu, Xin Cao, and Christian S. JensenProc. VLDB Endow., Jul 2022
The 3rd-Most Cited paper in VLDB 2022 2022 3/357
We all depend on mobility, and vehicular transportation affects the daily lives of most of us. Thus, the ability to forecast the state of traffic in a road network is an important functionality and a challenging task. Traffic data is often obtained from sensors deployed in a road network. Recent proposals on spatial-temporal graph neural networks have achieved great progress at modeling complex spatial-temporal correlations in traffic data, by modeling traffic data as a diffusion process. However, intuitively, traffic data encompasses two different kinds of hidden time series signals, namely the diffusion signals and inherent signals. Unfortunately, nearly all previous works coarsely consider traffic signals entirely as the outcome of the diffusion, while neglecting the inherent signals, which impacts model performance negatively. To improve modeling performance, we propose a novel Decoupled Spatial-Temporal Framework (DSTF) that separates the diffusion and inherent traffic information in a data-driven manner, which encompasses a unique estimation gate and a residual decomposition mechanism. The separated signals can be handled subsequently by the diffusion and inherent modules separately. Further, we propose an instantiation of DSTF, Decoupled Dynamic Spatial-Temporal Graph Neural Network (D2STGNN), that captures spatial-temporal correlations and also features a dynamic graph learning module that targets the learning of the dynamic characteristics of traffic networks. Extensive experiments with four real-world traffic datasets demonstrate that the framework is capable of advancing the state-of-the-art.
@article{10.14778/3551793.3551827, author = {Shao, Zezhi and Zhang, Zhao and Wei, Wei and Wang, Fei and Xu, Yongjun and Cao, Xin and Jensen, Christian S.}, title = {Decoupled Dynamic Spatial-Temporal Graph Neural Network for Traffic Forecasting}, year = {2022}, publisher = {VLDB Endowment}, volume = {15}, number = {11}, issn = {2150-8097}, url = {https://doi.org/10.14778/3551793.3551827}, doi = {10.14778/3551793.3551827}, journal = {Proc. VLDB Endow.}, month = jul, pages = {2733-2746}, numpages = {14}, }
- KDDPre-training Enhanced Spatial-temporal Graph Neural Network for Multivariate Time Series ForecastingZezhi Shao, Zhao Zhang, Fei Wang*, and Yongjun XuIn Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington DC, USA, Aug 2022
The 3rd-Most Cited paper in KDD 2022 2022 3/254
Multivariate Time Series (MTS) forecasting plays a vital role in a wide range of applications. Recently, Spatial-Temporal Graph Neural Networks (STGNNs) have become increasingly popular MTS forecasting methods. STGNNs jointly model the spatial and temporal patterns of MTS through graph neural networks and sequential models, significantly improving the prediction accuracy. But limited by model complexity, most STGNNs only consider short-term historical MTS data, such as data over the past one hour. However, the patterns of time series and the dependencies between them (i.e., the temporal and spatial patterns) need to be analyzed based on long-term historical MTS data. To address this issue, we propose a novel framework, in which STGNN is Enhanced by a scalable time series Pre-training model (STEP). Specifically, we design a pre-training model to efficiently learn temporal patterns from very long-term history time series (e.g., the past two weeks) and generate segment-level representations. These representations provide contextual information for short-term time series input to STGNNs and facilitate modeling dependencies between time series. Experiments on three public real-world datasets demonstrate that our framework is capable of significantly enhancing downstream STGNNs, and our pre-training model aptly captures temporal patterns.
@inproceedings{10.1145/3534678.3539396, author = {Shao, Zezhi and Zhang, Zhao and Wang, Fei and Xu, Yongjun}, title = {Pre-training Enhanced Spatial-temporal Graph Neural Network for Multivariate Time Series Forecasting}, year = {2022}, month = aug, isbn = {9781450393850}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3534678.3539396}, doi = {10.1145/3534678.3539396}, booktitle = {Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining}, pages = {1567-1577}, numpages = {11}, keywords = {multivariate time series forecasting, pre-training model, spatial-temporal graph neural network}, location = {Washington DC, USA}, series = {KDD '22}, }
- CIKMSpatial-Temporal Identity: A Simple yet Effective Baseline for Multivariate Time Series ForecastingZezhi Shao, Zhao Zhang, Fei Wang*, Wei Wei, and Yongjun XuIn Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, Oct 2022
The Most Cited paper in CIKM 2022 2022 1/561
Multivariate Time Series (MTS) forecasting plays a vital role in a wide range of applications. Recently, Spatial-Temporal Graph Neural Networks (STGNNs) have become increasingly popular MTS forecasting methods due to their state-of-the-art performance. However, recent works are becoming more sophisticated with limited performance improvements. This phenomenon motivates us to explore the critical factors of MTS forecasting and design a model that is as powerful as STGNNs, but more concise and efficient. In this paper, we identify the indistinguishability of samples in both spatial and temporal dimensions as a key bottleneck, and propose a simple yet effective baseline for MTS forecasting by attaching Spatial and Temporal IDentity information (STID), which achieves the best performance and efficiency simultaneously based on simple Multi-Layer Perceptrons (MLPs). These results suggest that we can design efficient and effective models as long as they solve the indistinguishability of samples, without being limited to STGNNs.
@inproceedings{10.1145/3511808.3557702, author = {Shao, Zezhi and Zhang, Zhao and Wang, Fei and Wei, Wei and Xu, Yongjun}, title = {Spatial-Temporal Identity: A Simple yet Effective Baseline for Multivariate Time Series Forecasting}, year = {2022}, month = oct, isbn = {9781450392365}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3511808.3557702}, doi = {10.1145/3511808.3557702}, booktitle = {Proceedings of the 31st ACM International Conference on Information \& Knowledge Management}, pages = {4454-4458}, numpages = {5}, keywords = {spatial-temporal graph neural network, multivariate time series forecasting, baseline}, location = {Atlanta, GA, USA}, series = {CIKM '22}, }
- ACM MMTrajectory Prediction from Hierarchical PerspectiveTangwen Qian, Yongjun Xu, Zhao Zhang, and Fei WangIn Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, Oct 2022
Predicting the future trajectories of multiple agents is essential for various applications in real life, such as surveillance systems, autonomous driving and social robots. The trajectory prediction task is influenced by many factors, including the individual historical trajectory, interactions between agents and fuzzy nature of an agent’s motion. While existing methods have made great progress on the topic of trajectory prediction, they treat all the information uniformly, which limits the sufficiency of using information. To this end, in this paper, we propose to regard all the information in a two-level hierarchical view. Particularly, the first-level view is the inter-trajectory view. In this level, we observe that the difficulty to predict different trajectory samples is different. We define trajectory difficulty and train the proposed model in an "easy-to-hard” schema. The second-level view is the intra-trajectory level. We find the influencing factors for a particular trajectory can be divided into two parts. The first part is global features, which keep stable within a trajectory, i.e., the expected destination. The second part is local features, which change over time, i.e., the current position. We believe that the two types of information should be handled in different ways. The hierarchical view is beneficial to take full advantage of the information in a fine-grained way. Experimental results validate the effectiveness of the proposed model.
@inproceedings{10.1145/3503161.3548092, author = {Qian, Tangwen and Xu, Yongjun and Zhang, Zhao and Wang, Fei}, title = {Trajectory Prediction from Hierarchical Perspective}, year = {2022}, month = oct, isbn = {9781450392037}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3503161.3548092}, doi = {10.1145/3503161.3548092}, booktitle = {Proceedings of the 30th ACM International Conference on Multimedia}, pages = {6822-6830}, numpages = {9}, keywords = {hierarchical perspective, spatial-temporal modeling, trajectory prediction}, location = {Lisboa, Portugal}, series = {MM '22}, }
2021
- IJCNNOn Accurate Computation of Trajectory Similarity via Single Image Super-ResolutionHanlin Cao, Haina Tang*, Yulei Wu, Fei Wang, and Yongjun XuIn 2021 International Joint Conference on Neural Networks (IJCNN), Jul 2021
@inproceedings{9533802, author = {Cao, Hanlin and Tang, Haina and Wu, Yulei and Wang, Fei and Xu, Yongjun}, booktitle = {2021 International Joint Conference on Neural Networks (IJCNN)}, title = {On Accurate Computation of Trajectory Similarity via Single Image Super-Resolution}, year = {2021}, month = jul, volume = {}, number = {}, pages = {1-9}, keywords = {Recurrent neural networks;Computational modeling;Superresolution;Predictive models;Nonuniform sampling;Data models;Trajectory}, doi = {10.1109/IJCNN52387.2021.9533802}, }
- ICPRTrajectory-User Link with Attention Recurrent NetworksTao Sun, Yongjun Xu, Fei Wang, Lin Wu, Tangwen Qian, and Zezhi ShaoIn 2020 25th International Conference on Pattern Recognition (ICPR), May 2021
@inproceedings{9412453, author = {Sun, Tao and Xu, Yongjun and Wang, Fei and Wu, Lin and Qian, Tangwen and Shao, Zezhi}, booktitle = {2020 25th International Conference on Pattern Recognition (ICPR)}, title = {Trajectory-User Link with Attention Recurrent Networks}, year = {2021}, month = may, volume = {}, number = {}, pages = {4589-4596}, keywords = {Weight measurement;Training;Recurrent neural networks;Semantics;Graphics processing units;Trajectory;Pattern recognition}, doi = {10.1109/ICPR48806.2021.9412453}, }
- InnovationArtificial intelligence: A powerful paradigm for scientific researchYongjun Xu, Xin Liu, Xin Cao, Changping Huang, Enke Liu, Sen Qian, Xingchen Liu, Yanjun Wu, Fengliang Dong, Cheng-Wei Qiu, and 38 more authorsThe Innovation, Nov 2021
Fei Wang receveid the Best Paper Award 2024 for the contribution to a comprehensive review of AI for Science
Artificial intelligence (AI) coupled with promising machine learning (ML) techniques well known from computer science is broadly affecting many aspects of various fields including science and technology, industry, and even our day-to-day life. The ML techniques have been developed to analyze high-throughput data with a view to obtaining useful insights, categorizing, predicting, and making evidence-based decisions in novel ways, which will promote the growth of novel applications and fuel the sustainable booming of AI. This paper undertakes a comprehensive survey on the development and application of AI in different aspects of fundamental sciences, including information science, mathematics, medical science, materials science, geoscience, life science, physics, and chemistry. The challenges that each discipline of science meets, and the potentials of AI techniques to handle these challenges, are discussed in detail. Moreover, we shed light on new research trends entailing the integration of AI into each scientific discipline. The aim of this paper is to provide a broad research guideline on fundamental sciences with potential infusion of AI, to help motivate researchers to deeply understand the state-of-the-art applications of AI-based fundamental sciences, and thereby to help promote the continuous development of these fundamental sciences.
@article{XU2021100179, title = {Artificial intelligence: A powerful paradigm for scientific research}, journal = {The Innovation}, volume = {2}, number = {4}, pages = {100179}, year = {2021}, month = nov, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2021.100179}, url = {https://www.sciencedirect.com/science/article/pii/S2666675821001041}, author = {Xu, Yongjun and Liu, Xin and Cao, Xin and Huang, Changping and Liu, Enke and Qian, Sen and Liu, Xingchen and Wu, Yanjun and Dong, Fengliang and Qiu, Cheng-Wei and Qiu, Junjun and Hua, Keqin and Su, Wentao and Wu, Jian and Xu, Huiyu and Han, Yong and Fu, Chenguang and Yin, Zhigang and Liu, Miao and Roepman, Ronald and Dietmann, Sabine and Virta, Marko and Kengara, Fredrick and Zhang, Ze and Zhang, Lifu and Zhao, Taolan and Dai, Ji and Yang, Jialiang and Lan, Liang and Luo, Ming and Liu, Zhaofeng and An, Tao and Zhang, Bin and He, Xiao and Cong, Shan and Liu, Xiaohong and Zhang, Wei and Lewis, James P. and Tiedje, James M. and Wang, Qi and An, Zhulin and Wang, Fei and Zhang, Libo and Huang, Tao and Lu, Chuan and Cai, Zhipeng and Wang, Fang and Zhang, Jiabao}, keywords = {artificial intelligence, machine learning, deep learning, information science, mathematics, medical science, materials science, geoscience, life science, physics, chemistry}, }
2020
- InnovationModeling the COVID-19 Outbreak in China through Multi-source Information FusionLin Wu*, Lizhe Wang, Nan Li, Tao Sun, Tangwen Qian, Yu Jiang, Fei Wang*, and Yongjun Xu*The Innovation, Aug 2020
Modeling the outbreak of a novel epidemic, such as coronavirus disease 2019 (COVID-19), is crucial for estimating its dynamics, predicting future spread and evaluating the effects of different interventions. However, there are three issues that make this modeling a challenging task: uncertainty in data, roughness in models, and complexity in programming. We addressed these issues by presenting an interactive individual-based simulator, which is capable of modeling an epidemic through multi-source information fusion.
@article{WU2020100033, title = {Modeling the COVID-19 Outbreak in China through Multi-source Information Fusion}, journal = {The Innovation}, volume = {1}, number = {2}, pages = {100033}, year = {2020}, month = aug, issn = {2666-6758}, doi = {https://doi.org/10.1016/j.xinn.2020.100033}, url = {https://www.sciencedirect.com/science/article/pii/S2666675820300333}, author = {Wu, Lin and Wang, Lizhe and Li, Nan and Sun, Tao and Qian, Tangwen and Jiang, Yu and Wang, Fei and Xu, Yongjun}, }
- ICANNCABIN: A Novel Cooperative Attention Based Location Prediction Network Using Internal-External Trajectory DependenciesTangwen Qian, Fei Wang*, Yongjun Xu, Yu Jiang, Tao Sun, and Yong YuIn Artificial Neural Networks and Machine Learning – ICANN 2020, Aug 2020
Nowadays, large quantities of advanced locating sensors have been widely used, which makes it possible to deploy location-based service (LBS) enhanced by intelligent technologies. Location prediction, as one of the most fundamental technologies, aims to acquire possible location at next timestamp based on the moving pattern of current trajectories. High accuracy of location prediction could enrich and increase user experience of various LBSs and brings lots of benefits to service providers. Lots of state-of-the-art research try to model spatial-temporal trajectories based on recurrent neural networks (RNNs), yet fails to arrive at a practical usability. We observe that there exists two ways to improve through attention mechanism which performs well in computer vision and natural language processing domains. Firstly recent location prediction methods are usually equipped with single-head attention mechanism to promote accuracy, which is only able to capture limited information in a specific subspace at a specific position. Secondly, existing methods focus on external relations between spatial-temporal trajectories, but miss internal relations in each spatial-temporal trajectory. To tackle the problem of model spatial-temporal patterns of mobility, we propose a novel Cooperative Attention Based location prediction network using Internal-External trajectory dependencies correspondingly in this paper. We also design and perform experiments on two real-world check-in datasets, Foursquare data in New York and Tokyo cities. Evaluation results demonstrate that our method outperforms state-of-the-art models.
@inproceedings{10.1007/978-3-030-61616-8_42, author = {Qian, Tangwen and Wang, Fei and Xu, Yongjun and Jiang, Yu and Sun, Tao and Yu, Yong}, editor = {Farka{\v{s}}, Igor and Masulli, Paolo and Wermter, Stefan}, title = {CABIN: A Novel Cooperative Attention Based Location Prediction Network Using Internal-External Trajectory Dependencies}, booktitle = {Artificial Neural Networks and Machine Learning -- ICANN 2020}, year = {2020}, publisher = {Springer International Publishing}, address = {Cham}, pages = {521--532}, isbn = {978-3-030-61616-8}, }
- IJCNNTULSN: Siamese Network for Trajectory-user LinkingYong Yu, Haina Tang*, Fei Wang, Lin Wu, Tangwen Qian, Tao Sun, and Yongjun XuIn 2020 International Joint Conference on Neural Networks (IJCNN), Aug 2020
@inproceedings{9206609, author = {Yu, Yong and Tang, Haina and Wang, Fei and Wu, Lin and Qian, Tangwen and Sun, Tao and Xu, Yongjun}, booktitle = {2020 International Joint Conference on Neural Networks (IJCNN)}, title = {TULSN: Siamese Network for Trajectory-user Linking}, year = {2020}, volume = {}, number = {}, pages = {1-8}, keywords = {Trajectory;Semantics;Learning systems;Training;Data mining;Task analysis;Data models;Siamese Network;Spatio-temporal data;User identification}, doi = {10.1109/IJCNN48605.2020.9206609}, }
- IEEE NetworkData Security and Privacy Challenges of Computing Offloading in FINsFei Wang*, Boyu Diao, Tao Sun, and Yongjun XuIEEE Network, Apr 2020
@article{9055731, author = {Wang, Fei and Diao, Boyu and Sun, Tao and Xu, Yongjun}, journal = {IEEE Network}, title = {Data Security and Privacy Challenges of Computing Offloading in FINs}, year = {2020}, month = apr, volume = {34}, number = {2}, pages = {14-20}, keywords = {Task analysis;Cloud computing;Data privacy;Edge computing;Smart homes;Data security}, doi = {10.1109/MNET.001.1900140}, }
2017
- NavigationMapping Global Shipping Density from AIS DataLin Wu, Yongjun Xu, Qi Wang, Fei Wang, and Zhiwei XuJournal of Navigation, Apr 2017
@article{Wu_Xu_Wang_Wang_Xu_2017, title = {Mapping Global Shipping Density from AIS Data}, volume = {70}, doi = {10.1017/S0373463316000345}, number = {1}, journal = {Journal of Navigation}, author = {Wu, Lin and Xu, Yongjun and Wang, Qi and Wang, Fei and Xu, Zhiwei}, year = {2017}, pages = {67-81}, }
2016
- IEEE TVT2FLIP: A Two-Factor Lightweight Privacy-Preserving Authentication Scheme for VANETFei Wang, Yongjun Xu, Hanwen Zhang, Yujun Zhang, and Liehuang ZhuIEEE Transactions on Vehicular Technology, Feb 2016
@article{7038220, author = {Wang, Fei and Xu, Yongjun and Zhang, Hanwen and Zhang, Yujun and Zhu, Liehuang}, journal = {IEEE Transactions on Vehicular Technology}, title = {2FLIP: A Two-Factor Lightweight Privacy-Preserving Authentication Scheme for VANET}, year = {2016}, month = feb, volume = {65}, number = {2}, pages = {896-911}, keywords = {Vehicles;Privacy;Vehicular ad hoc networks;Authentication;Telematics;Biology;Privacy;vehicular ad hoc network;two factor authentication;conditional traceability;strong non-repudiation;Conditional traceability;privacy;strong nonrepudiation;two-factor authentication;vehicular ad-hoc network (VANET)}, doi = {10.1109/TVT.2015.2402166}, }