1
Multi-Scale Adaptive Graph Neural Network for
Multivariate Time Series Forecasting
Ling Chen, Donghui Chen, Zongjiang Shang, Binqing Wu, Cen Zheng, Bo Wen, and Wei Zhang
Abstract—Multivariate time series (MTS) forecasting plays an
important role in the automation and optimization of intelligent
applications. It is a challenging task, as we need to consider
both complex intra-variable dependencies and inter-variable
dependencies. Existing works only learn temporal patterns with
the help of single inter-variable dependencies. However, there
are multi-scale temporal patterns in many real-world MTS.
Single inter-variable dependencies make the model prefer to learn
one type of prominent and shared temporal patterns. In this
paper, we propose a multi-scale adaptive graph neural network
(MAGNN) to address the above issue. MAGNN exploits a multi-
scale pyramid network to preserve the underlying temporal
dependencies at different time scales. Since the inter-variable
dependencies may be different under distinct time scales, an
adaptive graph learning module is designed to infer the scale-
specific inter-variable dependencies without pre-defined priors.
Given the multi-scale feature representations and scale-specific
inter-variable dependencies, a multi-scale temporal graph neural
network is introduced to jointly model intra-variable dependen-
cies and inter-variable dependencies. After that, we develop a
scale-wise fusion module to effectively promote the collaboration
across different time scales, and automatically capture the im-
portance of contributed temporal patterns. Experiments on six
real-world datasets demonstrate that MAGNN outperforms the
state-of-the-art methods across various settings.
Index Terms—Multivariate time series forecasting, multi-scale
modeling, graph neural network, graph learning.
I. INTRODUCTION
Multivariate time series (MTS) are ubiquitous in various
real-world scenarios, e.g., the traffic flows in a city, the
stock prices in a stock market, and the household power
consumption in a city block [1]. MTS forecasting, which aims
at forecasting the future trends based on a group of historical
observed time series, has been widely studied in recent years.
It is of great importance in a wide range of applications, e.g.,
a better driving route can be planned in advance based on
the forecasted traffic flows, and an investment strategy can be
designed with the forecasting of the near-future stock market
[2]–[5].
Making accurate MTS forecasting is a challenging task, as
both intra-variable dependencies (i.e., the temporal dependen-
cies within one time series) and inter-variable dependencies
(i.e., the forecasting values of a single variable are affected
Ling Chen and Donghui Chen are co-first authors. Corresponding author:
Ling Chen.
Ling Chen, Donghui Chen, Zongjiang Shang, and Binqing Wu are with
the College of Computer Science and Technology, Zhejiang University,
Hangzhou 310027, China (e-mail: {lingchen, chendonghui, zongjiangshang,
binqingwu}@cs.zju.edu.cn).
Cen Zheng, Bo Wen, and Wei Zhang are with Alibaba Group, Hangzhou
311100, China (e-mail: {mingyan.zc, wenbo.wb, zwei}@alibaba-inc.com).
by other variables) need to be considered jointly. To solve
this problem, traditional methods [6]–[8], e.g., vector auto-
regression (VAR), temporal regularized matrix factorization
(TRMF), vector auto-regression moving average (VARMA),
and gaussian process (GP), often rely on the strict stationary
assumption and cannot capture the non-linear dependencies
among variables. Deep neural networks have shown superior-
ity on modeling non-stationary and non-linear dependencies.
Particularly, two variants of recurrent neural network (RNNs)
[9], namely the long-short term memory (LSTM) and the gated
recurrent unit (GRU), and temporal convolutional networks
(TCNs) [10] have significantly achieved impressive perfor-
mance in time series modeling. To capture both long-term and
short-term temporal dependencies, existing works [3], [11]–
[14] introduce several strategies, e.g., skip-connection, atten-
tion mechanism, and memory-based network. These works
focus on modeling temporal dependencies, and process the
MTS input as vectors and assume that the forecasting values
of a single variable are affected by all other variables, which
is unreasonable and hard to meet in realistic applications. For
example, the traffic flows of a street are largely affected by
its neighboring streets, while the impact from distant streets
is relatively small. Thus, it is crucial to model the pairwise
inter-variable dependencies explicitly.
Graph is an abstract data type representing relations between
nodes. Graph neural networks (GNNs) [15], [16], which can
effectively capture nodes’ high-level representations while
exploiting pairwise dependencies, have been considered as a
promising way to handle graph data. MTS forecasting can be
considered from the perspective of graph modeling. The vari-
ables in MTS can be regarded as the nodes in a graph, while
the pairwise inter-variable dependencies as edges. Recently,
several works [17]–[19] exploit GNNs to model MTS taking
advantage of the rich structural information (i.e., featured
nodes and weighted edges) of a graph. These works stack GNN
and temporal convolution modules to learn temporal patterns,
and have achieved promising results. Nevertheless, there are
still two important aspects neglected in above works.
First, existing works only consider temporal dependencies
on a single time scale, which may not properly reflect the
variations in many real-world scenarios. In fact, the temporal
patterns hidden in real-world MTS are much more compli-
cated, including daily, weekly, monthly, and other specific
periodic patterns. For example, Fig. 1 shows the power con-
sumptions of 4 households within two weeks. There exists
a mixture of short-term and long-term repeating patterns (i.e.,
daily and weekly). These multi-scale temporal patterns provide
abundant information to model MTS. Furthermore, if the
arXiv:2201.04828v2 [cs.LG] 9 Apr 2023
评论