Spatiotemporal neural network with attention mechanism for El Nino forecasts SCIE SCOPUS

Cited 5 time in WEB OF SCIENCE Cited 11 time in Scopus
Title
Spatiotemporal neural network with attention mechanism for El Nino forecasts
Author(s)
Kim, Jinah; Kwon, Min Ho; Kim, Sung Dae; Kug, Jong-Seong; Ryu, Joon-Gyu; Kim, Jaeil
KIOST Author(s)
Kim, Jinah(김진아)Kwon, Min Ho(권민호)Kim, Sung Dae(김성대)
Alternative Author(s)
김진아; 권민호; 김성대
Publication Year
2022-05
Abstract
To learn spatiotemporal representations and anomaly predictions from geophysical data, we propose STANet, a spatiotemporal neural network with a trainable attention mechanism, and apply it to El Nino predictions for long-lead forecasts. The STANet makes two critical architectural improvements: it learns spatial features globally by expanding the network's receptive field and encodes long-term sequential features with visual attention using a stateful long-short term memory network. The STANet conducts multitask learning of Nino3.4 index prediction and calendar month classification for predicted indices. In a comparison of the proposed STANet performance with the state-of-the-art model, the accuracy of the 12-month forecast lead correlation coefficient was improved by 5.8% and 13% for Nino3.4 index prediction and corresponding temporal classification, respectively. Furthermore, the spatially attentive regions for the strong El Nino events displayed spatial relationships consistent with the revealed precursor for El Nino occurrence, indicating that the proposed STANet provides good understanding of the spatiotemporal behavior of global sea surface temperature and oceanic heat content for El Nino evolution.
ISSN
2045-2322
URI
https://sciwatch.kiost.ac.kr/handle/2020.kiost/42479
DOI
10.1038/s41598-022-10839-z
Bibliographic Citation
Scientific Reports, v.12, no.1, 2022
Publisher
Nature Publishing Group
Type
Article
Language
English
Document Type
Article
Files in This Item:
There are no files associated with this item.

qrcode

Items in ScienceWatch@KIOST are protected by copyright, with all rights reserved, unless otherwise indicated.

Browse