CityTransformer: A Transformer-Based Model for Contaminant Dispersion Prediction in a Realistic Urban Area

被引:2
|
作者
Asahi, Yuuichi [1 ]
Onodera, Naoyuki [1 ]
Hasegawa, Yuta [1 ]
Shimokawabe, Takashi [2 ]
Shiba, Hayato [2 ]
Idomura, Yasuhiro [1 ]
机构
[1] Japan Atom Energy Agcy, Ctr Computat Sci & E Syst, Chiba 2770827, Japan
[2] Univ Tokyo, Informat Technol Ctr, Chiba 2770882, Japan
关键词
Deep learning; Graphics-processing-unit-based computing; Lattice Boltzmann method; Urban plume dispersion; LATTICE BOLTZMANN METHOD; LARGE-EDDY SIMULATION; PLUME DISPERSION; NEURAL-NETWORKS; FLOW; PARAMETRIZATION; TURBULENCE; CANOPY; CFD;
D O I
10.1007/s10546-022-00777-8
中图分类号
P4 [大气科学(气象学)];
学科分类号
0706 ; 070601 ;
摘要
We develop a Transformer-based deep learning model to predict the plume concentrations in the urban area in statistically stationary flow conditions under a stationary and homogeneous forcing. Our model has two distinct input layers: Transformer layers for sequential data and convolutional layers in convolutional neural networks for image-like data. Our model can predict the plume concentration from realistically available data such as the time series monitoring data at a few observation stations, and the building shapes and the source location. It is shown that the model can give reasonably accurate prediction in less than a second. It is also shown that exactly the same model can be applied to predict the source location and emission rate, which also gives reasonable prediction accuracy.
引用
收藏
页码:659 / 692
页数:34
相关论文
共 50 条
  • [31] Traffic Transformer: Transformer-based framework for temporal traffic accident prediction
    Al-Thani, Mansoor G.
    Sheng, Ziyu
    Cao, Yuting
    Yang, Yin
    AIMS MATHEMATICS, 2024, 9 (05): : 12610 - 12629
  • [32] Transformer and Graph Transformer-Based Prediction of Drug-Target Interactions
    Qian, Meiling
    Lu, Weizhong
    Zhang, Yu
    Liu, Junkai
    Wu, Hongjie
    Lu, Yaoyao
    Li, Haiou
    Fu, Qiming
    Shen, Jiyun
    Xiao, Yongbiao
    CURRENT BIOINFORMATICS, 2024, 19 (05) : 470 - 481
  • [33] MM-Transformer: A Transformer-Based Knowledge Graph Link Prediction Model That Fuses Multimodal Features
    Wang, Dongsheng
    Tang, Kangjie
    Zeng, Jun
    Pan, Yue
    Dai, Yun
    Li, Huige
    Han, Bin
    SYMMETRY-BASEL, 2024, 16 (08):
  • [34] CFD simulations of flow and dust dispersion in a realistic urban area
    Luo, Kun
    Yu, HaiJun
    Dai, Zhixiang
    Fang, MingMing
    Fan, Jianren
    ENGINEERING APPLICATIONS OF COMPUTATIONAL FLUID MECHANICS, 2016, 10 (01) : 229 - 243
  • [35] A Transformer-Based Bridge Structural Response Prediction Framework
    Li, Ziqi
    Li, Dongsheng
    Sun, Tianshu
    SENSORS, 2022, 22 (08)
  • [36] Rethinking Transformer-based Set Prediction for Object Detection
    Sun, Zhiqing
    Cao, Shengcao
    Yang, Yiming
    Kitani, Kris
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 3591 - 3600
  • [37] Transformer-based attention network for stock movement prediction
    Zhang, Qiuyue
    Qin, Chao
    Zhang, Yunfeng
    Bao, Fangxun
    Zhang, Caiming
    Liu, Peide
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 202
  • [38] TransCFD: A transformer-based decoder for flow field prediction
    Jiang, Jundou
    Li, Guanxiong
    Jiang, Yi
    Zhang, Laiping
    Deng, Xiaogang
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [39] Deep Transformer-Based Asset Price and Direction Prediction
    Gezici, Abdul Haluk Batur
    Sefer, Emre
    IEEE ACCESS, 2024, 12 : 24164 - 24178
  • [40] Transformer-based Architecture for Empathy Prediction and Emotion Classification
    Vasava, Himil
    Uikey, Pramegh
    Wasnik, Gaurav
    Sharma, Raksha
    PROCEEDINGS OF THE 12TH WORKSHOP ON COMPUTATIONAL APPROACHES TO SUBJECTIVITY, SENTIMENT & SOCIAL MEDIA ANALYSIS, 2022, : 261 - 264