0% found this document useful (0 votes)
55 views

Time Series Analysis

The document discusses using quantum computers for time series analysis, noting that some quantum machine learning models and algorithms like quantum PCA could be useful for analyzing time series data since time series can be viewed as continuous functions. It then reviews classical approaches to time series analysis and neural networks before discussing how quantum analogues of classical preprocessing and ARIMA forecasting models could be implemented using simple quantum gates.

Uploaded by

Rohit Pandey
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
55 views

Time Series Analysis

The document discusses using quantum computers for time series analysis, noting that some quantum machine learning models and algorithms like quantum PCA could be useful for analyzing time series data since time series can be viewed as continuous functions. It then reviews classical approaches to time series analysis and neural networks before discussing how quantum analogues of classical preprocessing and ARIMA forecasting models could be implemented using simple quantum gates.

Uploaded by

Rohit Pandey
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

1

A walk through of time series analysis on quantum


computers
Ammar Daskin
Dept. of Computer Engineering
Istanbul Medeniyet University, Istanbul, Turkey

Abstract—Because of the rotational components on quantum algorithm in an unordered list [4], and the quantum principal
arXiv:2205.00986v2 [quant-ph] 22 Jun 2022

circuits, some quantum neural networks based on variational component analysis (PCA) [5].
circuits can be considered equivalent to the classical Fourier
networks. In addition, they can be used to predict the Fourier
coefficients of continuous functions. Time series data indicates A. Classical neural networks
a state of a variable in time. Since some time series data
can be also considered as continuous functions, we can expect The representation theorem indicates any multivariate con-
quantum machine learning models to do many data analysis tasks tinuous function can be represented as a superposition of
successfully on time series data. Therefore, it is important to functions of one variable and addition [6, 7]. An artificial
investigate new quantum logics for temporal data processing and
analyze intrinsic relationships of data on quantum computers.
neural network can be considered as a composition of simple
In this paper, we go through the quantum analogues of nonlinear activation function[8]: Let I n be the n-dimensional
classical data preprocessing and forecasting with ARIMA models unit-cube [0, 1]n and C(I n ) be the space of continuous func-
by using simple quantum operators requiring a few number of tions on I n . In addition, let σ be a sigmoidal function. Given
quantum gates. Then we discuss future directions and some of any f ∈ C(I n ) and  > 0, there is a G of the following form:
the tools/algorithms that can be used for temporal data analysis
on quantum computers. n
X
Index Terms—quantum machine learning, quantum algo- G(x) = αj σ(wjT x + θj ) (1)
rithms, quantum optimization, quantum time series, quantum j=1
temporal data,
for which |G(x) − f (x)| < .
Gallant et al. [9] showed a single feed-forward neural
I. I NTRODUCTION network where the output is squashed by using a cosine
It is shown that nature itself is explained better in quantum activation function: i.e. by considering σ as cosine, sine, or
T
physics than classical one. The idea of quantum computers exponential (i.e. in the form eiwj x+θj ) functions, one can
was put forward by Feynman [1] and other pioneers [2] so approximate the Fourier transform of a continuous function.
as to simulate and understand intrinsic physical properties These networks are often called Fourier neural networks
of nature in a better way: Simulating quantum systems with (FNN). On the other hand, Fourier analysis of data also
many particles is difficult on classical computers because sheds lights on the generalization performances of deep neural
the computational complexity in general grows exponentially networks[10].
with the number of involved particles. On the other hand,
quantum computers are based on qubits implemented by B. Time series problems
quantum particles which in turn indicates that in principle, we
can increase the computational power of quantum computers Temporal data indicates a state of a variable in time,
exponentially by adding controllable new qubits. Therefore, which could be collected through a wide variety of monitor-
the difficulty of simulating nature for classical computers ing devices. Examples include financial market data; sensor
would become the main advantage of quantum computers. data which may indicate temperature, humidity and so on;
Unfortunately since the input/output of quantum computers and medical data collected by different monitoring medical
are still in classical ones and zeros, it necessitates a state devices. Depending on the collection method, the data can
preparation that involves classical processing of the data be continuous or discrete. Continuous temporal data sets are
and a measurement that requires many repetitions of the generally referred to as times series data. [11] The data mining
same computation. Therefore, the complexity boundaries in tasks on time series data include clustering, classification, and
classical computation mostly remain the same for quantum forecasting: In general it is required to specialize the known
computing. Nonetheless, it is still possible to design more algorithms in data analysis in order to use in the analysis of
efficient algorithms by employing quantum entanglement and time series data: e.g. see the classification with deep learning
superposition (quantum parallelism): Some known examples for time series data [12] and the distance based time series
are Shor’s integer factoring algorithm [3], Grover’s search classification [13]. It is shown that time series data can be also
modeled by using Fourier deep neural networks [14], which
adaskin25@gmail.com is beriefly explained above.
2

C. Quantum machine learning II. P REPARING DATA FOR QUANTUM ALGORITHMS


Since the quantum machine learning [15, 16] is a very
new field, there are only a small number of partial studies The time series data is generally given in contextual repre-
which can be found for the analysis of time series data on sentations where attributes describe the context and behavioral
quantum computers: the time series data analysis is done by values at any reference points [11]. If more than one behavioral
using a quantum approach based on the Schrödinger’s equation attribute is associated with the series, then the data is referred
[17]. In a very recent paper, the quantum machine learning as the multivariate time series: i.e. A time series data with
approaches are investigated for the problems in finance [18]. dimension d and length n. The data is formed by a set of the
It is shown that the variational quantum machine learning values received at the time stamp ti . In the ith time stamp, the
models can be used to approximate Fourier transforms of the received data is a vector of dimension d: yi = (yi1 , . . . , yid ).
continuous functions [19]. It is also shown that the continuous In classical data analyses, the time series data is taken
variable (CV) quantum neural networks can be used for curve under preprocessing operations: e.g., the missing values may
fitting and function approximations[20]. It is still important be handled,the data may be normalized, smoothed, or scaled.
to study and build generic separate models and packages for Then, a variety of methods are used to transform and reduce
temporal data analysis. the dimension of the data. After these steps, the standard data
Quantum circuits are in general based on quantum gates analysis tasks such as classification and clustering are also
described by some form of rotations in the Hilbert space. For applicable to time series data. The classification and clustering
instance, the rotation gates for one-qubit can be defined in the tasks involve determining the correlation across the series and
following forms; classifying the series based on their shapes. These can be
!
cos(θ) −i sin(θ) offline or real time. And class labels may need to be assigned
Rx (2θ) = , (2) to the individual time stamps (point labels) or the whole series.
−i sin(θ) cos(θ)
! For quantum computers, the algorithms either use the data
cos(θ) − sin(θ) after some classical preprocessing, or as done in PCA [5],
Ry (2θ) = , (3)
sin(θ) cos(θ) it assumes the data is processed and stored in quantum RAM
! [22]. In the former case, the data is prepared as a quantum state
e−iθ 0 by a state preparation circuit that converts a standard initial
and Rz (2θ) = . (4)
0 eiθ state:e.g. |0i, into the state that represents the data |xi. If the
length of |xi is n, then this kind of state preparation requires a
Because of these rotational components, the quantum circuits circuit with O(n) number of controlled-NOT quantum gates in
can in theory provide a natural equivalent to the classical general [23]. In some cases, the data is also used as parameters
Fourier networks (e.g., see our previous paper[21] which that determine the rotation angle of the quantum gates(e.g.
gives an example quantum neural net with periodic activation in [24], the data is used with parameters to define quantum
function.). In addition, they can be used to predict Fourier rotation gates. The learning task is done by reusing the data
coefficients of continuous functions as done in Ref.[20] and adjusting the value of the parameters for quantum gates.).
After the mapping, |xi is a vector defined in the Hilbert space.
D. Quantum temporal data analysis It is shown in Ref.[25] that the Hilbert space can be considered
Since some time series data can also be considered as as a kernel space used in classical machine learning to ease the
continuous functions, quantum machine learning models can training process. Therefore, the training of a learning model
successfully do many data analysis tasks on the time series can be achieved also in the Hilbert space. Although this solves
data. Therefore, it is important to investigate new quantum the representability of the data in the Hilbert space, it is still
logics for temporal data processing and analyze intrinsic necessary to design efficient data mapping methods without
relationships of data on quantum computers. The processing disrupting the representability of the data (e.g. for MNIST
data on quantum computers in general requires the followings: data, it is reduced to 8 qubits by using classical PCA [26]
• classical preprocessing and quantum state preparation, which reduces the training results to as low as 70%. ).
• quantum preprocessing, Consequently, the representation is highly dependent on
• quantum processing of data, the quantum algorithm to be used in the data processing
• quantum measurement, step. However, in most cases, since the time series data is
• and analyses of the measurement results on classical multivariate in the form yi = (yi1 , . . . , yid ), the quantum
computers superposition approach can be used to efficiently represent the
In general, all or some of these steps are repeated until a de- data on quantum computers. This in general may reduce the
sired result/model is obtained in data analysis. For continuous d × n data into an n-dimensional quantum state:
temporal data, these steps are visualized in Fig. 1.
In the following sections, we go through the quantum 1 X
n
analogues of classical data preprocessing and forecasting. |yi = √ |yi i (5)
n i
Then we discuss some of the tools/algorithms that can be used
for temporal data analyses on quantum computers and discuss
future directions. Alternatively, one can build the following quantum state from
3

Fig. 1. General picture for processing temporal data on quantum computers (the measurement step is not shown.)

the input data:   a d-dimensional quantum system with O(log(d)) qubits


y1 or a d-level qubit.
 y2  stamp (i + 1) is prepared
2) The new observed data at time
 
|yi = 
 .. 
 (6) as a quantum state yi+1 on different qubits: i.e.,
 . 

yi+1 |yi i). Or we can create the following quantum
yn state by using a control qubit and data qubit.
Here, the dimension of the quantum state is log(nd). Another
yi+1
!
alternative can be the following: (10)
|yi i)
|yi = |y1 i |y2 i . . . |yn i . (7)
Note that this state can be formed with the help of
Although the required number of qubit which is n log(d) is control qubit by applying the operators related to yi+1
much more than the first two methods, this representation and |yi i) to the data qubit in |0i state.
may allow us to apply quantum operations to individual set 3) From the above state, a quantum entangling gate (e.g. the
of values more easily. However, since the required number of Hadamard gate) can be used to generate the following
qubits depend on n, it may not be possible on current quantum recurrence:
computers to use this representation for the series with large
0
E √   √  
n. yi+1 = α yi+1 + yi0 + 1 − α yi+1 − yi0

III. Q UANTUM PREPROCESSING OF TEMPORAL DATA =a yi+1 + b yi0


Below we show how to implement quantum analogues of (11)


some of the famous classical preprocessing techniques: the √ √  √ √ 
exponential smoothing, binning and data transformation. Note where a = α + 1 − α and b = α− 1−α .
that in most cases, we go through the steps for univariate data. In the case of the Hadamard gates, α is some power of
1/2. Using the aboveE recursion, if we substitute yi0
0
A. Exponential smoothing into the above, yi+1 becomes:
In the exponential smoothing, the smoothed value yi0 is E  E
0 0
obtained through the combination of yi and the pervious values yi+1 =a yi+1 + b a |yi i + b yi−1

with the ratios determined by the decay factor α:
=a yi+1 + ab |yi i + ab2 yi−1 . . . (12)

yi0 = αyi + (1 − α)yi−1
0
. (8) i
X
=abi+1 y00 + abi−k yk+1 ,

The recursive substitution in the above equation yields:
k=1
i
X
yi0 = (1 − α)i y00 + α yj (1 − α)i−j . (9) If we choose an α which makes b < 1, we can
j=1 have the quantum equivalent of the classical exponential
smoothing. In this case, the choice of α regulates the
In the analyses, the choice of α determines the impact of the
decay of the terms through b; therefore, we can call b as
previous data points.
the decay factor that determines the importance of the
In quantum analyses, we can construct a similar exponential
recent data points in any analysis using yi0 .

smoothing technique:
1) Different from the classical analysis, |yi i, the data
observed at the ith time stamp, can be a single data point B. Binning and moving average smoothing
or multiple data points observed on d dimension. There- The classical moving average is a special case of the binning
fore, as a physical qubit, this state can be implemented as method where the date is divided into equally spaced time
4

intervals of size k: [t1 , tk ], [tk+1 , t2k ]. Then, the data points In the wavelet transform, y can be represented with the
are assigned into the corresponding bins and the mean values basis vectors that form Hn :
of the bins are used for the analyses: q
X
Pk y0 = ai wi , (17)
0 yi.k+r
yi+1 = r=1 (13) i=1
k
where wi is the ith basis vector and ai is the normalized
Here, the number of the data points are reduced by a factor coefficient. The dimension reduction is done by disregarding
of k. the lowest ai s.
Utilizing the Hadamard gates on certain qubits, one can In quantum computing, since the Haar matrix is orthogonal
obtain the average of the neighboring states on certain quantum and can be built from the recurrence that is described by the
states. After finding the average, we can also disregard the rest Kronecker product, it can be implemented as a quantum gate in
of the state by collapsing quantum state onto the interested part a similar fashion to the gate. Therefore, given |yi,
Hadamard
of the system: One example could be as follows: y0 . The elimination of lowest parts of

  one
0 can easily obtain
yi1 y can be done through either sampling or measuring a few
 .  qubits. This leads to a problem of finding the best qubits that
• Let |yi i =  .. . If we multiply this quantum state
  gives the maximum reduction with the minimum information
yid loss.
Ubin = I ⊗ log2 (d/k) ⊗H ⊗ log2 k , where H is the Hadamard 1) Quantum Fourier transform (QFT): The discrete Fourier
gate: ! transform (DFT) can be used to transform a data vector into
√ 1 1 a linear combination of sinusoidal series. The similarity of
H = 1/ 2 (14) two vectors can be measured by taking the Euclidean distance
1 −1
of the Fourier coefficients. In quantum computing, DFT-which
If we apply Ubin to the quantum state |yi i, then when has the complexity bound O(nlogn) for n-dimensional vector-
the first log2 k qubits in |0 . . . 0i state, we would have an can be implemented with O(logn) computational steps. It
0
equivalent yi+1 which could be considered reduced data provides the main speedup of many quantum algorithms such
points. as integer factoring. The quantum Fourier transform for the
The classical moving average smoothing uses overlapping data |yi can be obtained by applying the operation QF T . If
bins which can be easily done on quantum case by utilizing a we have two data vectors y1 and y2 of the same dimension
quantum operator of the form: H ⊗ log2 k ⊗ I ⊗ log2 (d/k) . n (we assume this for convenience and brevity in notations, it
can be easily generalized to any dimension.), we first construct
the following quantum state:
C. Data transformation !
In classical data analyses, often the temporal data is made y1
|yi = (18)
smaller or discreticezed through data transformation methods y2
such as discrete wavelet and Fourier transforms (DWT and Then we apply the quantum Fourier transform to each part of
DFT respectively). the vectors:
Similar to the Fourier transform, DWT represents a data ! !
(function) in terms of an orthonormal basis. One of the QF T QF T y1
|yi = (19)
simplest wavelet transform is the Haar transform which can QF T QF T y2
be described a orthogonal matrix: e.g., 4 × 4 Haar transform Here note that
is   !
1 1 1 1 QF T
= I ⊗ QF T, (20)
1 1 1 −1 −1  QF T
H4 = √ √ (15)

2  2 − 2 √0 0

√  where I is a 2×2 identity matrix. We then apply the Hadamard
0 0 2 − 2 gate to the first qubit:
Similarly to Walsh-Hadamard transform, the unnormalized
! !

⊗log(n)
 QF T y
1 1 QF T y1 + QF T y2
Haar transform matrix can be obtained by the following H ⊗I =√
QF T y2 2 QF T y1 − QF T y2
recurrence: " # (21)
HN ⊗ [1, 1]
H2N = . (16) We can also start with a superposition of the input state:
IN ⊗ [1, −1]
1 
|yi = √ |y1 i − |y2 i . (22)
The transformation on an input vector y0 is described by y0 = 2
HN x. From this product, the rows of the Haar matrix allow us Then we apply QF T :
to draw different properties of the temporal data: For instance,
1 
for H4 , the first row takes the average of the input vector, while QF T |yi = √ QF T |y1 i − QF T |y2 i . (23)
the second row measures the low frequencies. The rest of the 2
rows measures different parts of the input vector. This also generates a kind of the desired distance.
5

As mentioned before, we can also use a separate register Then we apply laddered Toffoli operations (similar to a bit-
for each data point at different time: e.g., the quantum state adder: See quantum adders in Ref.[23]) to the second part of
|y1 i|y2 i is for the data measured at time t=1 and t=2. In this the state to convert into the following:
case we apply QF T to both registers and then use the SWAP  
test to create the distance measure between data points. y0
 y 
 1 
(QF T ⊗ QF T ) |y1 i |y2 i = QF T |y2 i QF T |y2 i (24)  . 
 . 
!  . 
In all cases, the Euclidean distance can be obtained by the 1 |yi y
 
√ =  n−1  (28)

measurement results. However, if d number of yi are given as
2 −Uadder |yi −yn−1 
classical input, the first approach requires O(log(dn)) number  −y0 
 
of qubits and the third approach requires O(dlog(n)) since  . 
 . 
a separate register is used for each yi . Since the second  . 
approach uses the superpositioned state, the number of qubits −yn−2
is O(log(n)) which is less than the first and third approaches.
On the other hand, the classical processing time is almost the Then, we apply a Hadamard gate again to the first qubit to
same for all approaches and is O(nd) since it requires linear obtain the following
scanning of the every data points. However, if |yi i are stored    
y0 y0 − yn−1
on a random accessed quantum memory (qRAM), then all  y   y −y 
 1  1 0
approaches requires only O(d) number of queries to qRAM
 
 .   .. 
 . 
to load the data into quantum registers.  .  .
 
 
 1 y  1
yn−1 − yn−2 
 
⊗n−1  n−1 
H ⊗I √  =   (29)
IV. T IME SERIES FORECASTING 2 −yn−1  2  y0 + yn−1 
 −y0   y1 + y0 
   
If the statistical parameters of time series such as the mean  . 
and variance do not change with time, it is called stationary:  . 
 .. 
 .  .
 
 
i.e. the probabilistic distribution of the parameters in any time −yn−2 yn−1 + yn−2
interval is the same as or very close to the time interval
found by shifting from this interval. And non-stationary if the Note that by applying Hadamard gates on different
qubits and
distributions change with time, which is mostly the case when some swap operations we can easily built y 00 or any high
we deal with real world data: e.g., if we look at the number of order differencing from the above state. For instance, one can
daily COVID cases, we see that the daily cases changes based obtain some seasonal differencing as follows: For a natural
on the season and different factors for the past few years. number s < i, let the seasonal difference be defined as
A nonstationary series can be made stationary by different
operations on the times series data. yi0 = yi − yi−s . (30)

In Eq.(29); instead of applying the Hadamard gate to the first


A. Differencing qubit, if we apply to the second or any other qubit based on the
Time series can be converted by considering differences value of s, we basically construct the equivalent quantum state
between the data yi at time ti and the data yi−1 at the previous for seasonal difference. Also note that the desired quantum
time ti−1 . This operation is called differencing and defined as state is the half part of the above state where the first qubit is
follows: in |0i state. In the further computations, one can either collapse
the state to the desired part by measuring the first qubit in |0i
yi0 = yi − yi−1 . (25) state1 .
If the data is still not stationary, then a second order
differencing also can be applied: B. Forecasting Models
yi00 = yi0 + yi−1
0
= yi − 2yi−1 + yi−2 (26) Differencing mean the time series is drifting in time; there-
fore, it can be modeled by simply the following:
To implement this as a quantum operation, first we ob-
serve that the closer y 0 to the equal superposition state, yi+1 = yi + c + ei+1 , (31)
|Hi = H ⊗n |0i, the more stationary it is. We can measure
the closeness of the two quantum state by the swap test[27] where yi+1 is the predicted value, ei+1 represents the white
or by some statistical sampling from the quantum state(i.e., noise with zero mean, and c is the time drift.
measuring some qubits to obtain probabilities of 0 and 1. The Forecasting on time series data can be made through au-
same or close values
indicate an equal superposition state). toregressive models. The forecasting with the univariate time
To construct y 0 first we use an auxiliary qubit in |1i, then series data is generally based on autoregression models where
apply a Hadamard gate to this qubit: the output value at time t, yt , is determined as a linear
!
 1 |yi 1
For positive y, we can also apply an oblivious-ampltidue amplification
H |1i |yi = √ (27)
2 − |yi [28] to the first qubit or consider a different order of differencing
6

combination of the values preceding window of some length auto-correlation between a value and its neighbors. In some
p (AR(p) model)[11]: ARIMA like models, p can be considered as the period of
p
X the time series. On quantum computers the value of p can be
yt = ai · yt−i + c + t , (32) found along with the other optimization parameters by using a
i=1 classical optimization. In addition, since it is the period of the
where c and the coefficients a1 . . . ap are learned through a data, one can also adapt quantum phase estimation algorithm
learning process. to find this period as done in Shor’s factoring algorithm.
In the moving average model (MA(q)), the behavioral
attribute value at any timestamp is determined from the un- C. Forecasting Models As Circulant Matrix Operators
expected historical variations in the time series data (shocks): If the equation for yt in any of the models is stacked up for
q the values t = 0, 1, . . . , T − 1, the above generic models can
X
yt = bi · t−i + c + t , (33) be rewritten as a system of equations constructed by circulant
i=1 matrices [30]. The circulant matrices can be implemented
very efficiently on quantum computers since their eigenspace
where c is the mean value and the coefficients bi s are learned
is formed by the Fourier transform (see Ref.[31] for circuit
from the data. A more powerful and general model can
implementation of these matrices). Therefore, one can use
be obtained by combining the two aforementioned models
those matrices in conjunction with the variational quantum
(ARMA(p, q)):
circuits to optimize the parameters of the forecasting models.
p
X q
X
yt = ai · yt−i + c + t + bi · t−i + c + t . (34)
V. D ISCUSSION AND F UTURE D IRECTIONS
i=1 i=1
In this paper, we go through data processing and forecasting
Autoregressive moving average models (ARMA) are used
with ARIMA models and show how to implement them on
best with the stationary data (when the mean, variance, and
quantum computers. As future direction, more example data
autocorrelation does not change in time). The non-stationary
analysis are needed to investigate algorithms and methods for
data are handled by integrating a series of differences into
preprocessing and transformation of time series data into the
the model (ARIMA(p, q, q)). For instance with the first-order
Hilbert space in which quantum computers work better than
difference (d = 1) the model can be written as:
classical computers; and investigate quantum machine learning
p q
X X methods and algorithms for univariate and/or multivariate time
yt0 = ai · 0
yt−i + c + t + bi · t−i + c + t . (35)
series data to forecast, classify and cluster the data. In addition,
i=1 i=1
for time series data, there are many classical packages such
We can easily generate the above by using a variational as Facebook’s- Prophet [32] and ARIMA (Autoregressive
quantum circuit including two and single qubit gates (as done Integrated Moving Average) models. There is also need to
in many of quantum machine learning models) and apply to design a quantum packages-that can be used with the current
the quantum state prepared in one of the following forms: quantum libraries such as IBM-QISKIT[33]-for time series
data to forecast and do other tasks.
!
 |yi
|yi |i , θy |yi + θ |i , or . (36)
|i
A. Data representation and Measurement
Note that |yi may also represent the state after differencing. Representation of data on quantum computers may not be
Also note that the angle values of the gates in the circuit without noise. Therefore, the small changes in any feature
are related to the parameters a and b so that we can obtain of data may not be observed on quantum state by using
the parameterized model in Eq.(35) or any of the previous simple measurements. If one qubit is used for one feature,
models. Furthermore, note that following the Ref.[29] where then we expect any change in the input data to impact the
it is shown how to map the Schmidt decomposition of a measurement statistics of a qubit. However, representing whole
vector into a quantum circuit, one can also prepare a quantum feature vector as a quantum state with fewer qubits enforces an
operator whose first row is the parameters a and b. automatic dimension reduction in the measurement and hence
 
a1 . . . ap b1 . . . bq may cause a significant data loss that impedes the prediction
 • ... •

... •
 results.
. (37)
.. .. .. .. .. 
. . . . B. Quantum optimization for autoregressive models
If the dimensions of a and b are k, this operator can be Note that autoregressive models can be also combined
generated by using O(k) number of quantum operations. We with quantum neural networks as done with the classical
map the parameters to the angles of the rotation gates. Then, neural networks [34]. Also note that autoregressive models
we adjust the rotation gate parameters by using a classical can be formulated as combinatorial optimization problems
optimization technique as done in the variational solvers. in the framework of graph learning [35]. In addition, it is
Since the value of p determines the seasonality (period) shown that quadratic unconstrained optimization formulation
in the time series. Therefore, ARIMA model indicates the can be used for forecasting in finance [36]. With the help
7

of these and other similar studies, autoregressive models can [14] M. S. Gashler and S. C. Ashmore, “Modeling time series
be formulated as combinatorial optimization in the form of data with deep fourier neural networks,” Neurocomput-
quadratic unconstrained binary optimization. Therefore, the ing, vol. 188, pp. 3–11, 2016.
quantum optimization approaches such as adiabatic quantum [15] V. Dunjko and P. Wittek, “A non-review of quantum
computation [37], quantum unconstrained bounded optimiza- machine learning: trends and explorations,” Quantum
tion algorithm [38], quantum power iteration [39] can be used Views, vol. 4, p. 32, 2020.
for these models. [16] V. Dunjko and H. J. Briegel, “Machine learning &
artificial intelligence in the quantum domain: a review of
recent progress,” Reports on Progress in Physics, vol. 81,
R EFERENCES no. 7, p. 074001, 2018.
[17] P. Singh, G. Dhiman, and A. Kaur, “A quantum approach
[1] R. P. Feynman, “Quantum mechanical computers,” Foun- for time series data based on graph and schrödinger
dations of physics, vol. 16, no. 6, pp. 507–531, 1986. equations methods,” Modern Physics Letters A, vol. 33,
[2] P. Benioff, “The computer as a physical system: A no. 35, p. 1850208, 2018.
microscopic quantum mechanical hamiltonian model of [18] D. Emmanoulopoulos and S. Dimoska, “Quantum ma-
computers as represented by turing machines,” Journal chine learning in finance: Time series forecasting,” arXiv
of statistical physics, vol. 22, no. 5, pp. 563–591, 1980. preprint arXiv:2202.00599, 2022.
[3] P. W. Shor, “Algorithms for quantum computation: dis- [19] M. Schuld, R. Sweke, and J. J. Meyer, “Effect of data en-
crete logarithms and factoring,” in Proceedings 35th coding on the expressive power of variational quantum-
annual symposium on foundations of computer science. machine-learning models,” Physical Review A, vol. 103,
Ieee, 1994, pp. 124–134. no. 3, p. 032430, 2021.
[4] L. K. Grover, “Quantum mechanics helps in searching for [20] N. Killoran, T. R. Bromley, J. M. Arrazola, M. Schuld,
a needle in a haystack,” Physical review letters, vol. 79, N. Quesada, and S. Lloyd, “Continuous-variable quantum
no. 2, p. 325, 1997. neural networks,” Physical Review Research, vol. 1,
[5] S. Lloyd, M. Mohseni, and P. Rebentrost, “Quantum no. 3, p. 033063, 2019.
principal component analysis,” Nature Physics, vol. 10, [21] A. Daskin, “A simple quantum neural net with a periodic
no. 9, pp. 631–633, 2014. activation function,” in 2018 IEEE International Confer-
[6] A. N. Kolmogorov, “On the representation of continuous ence on Systems, Man, and Cybernetics (SMC). IEEE,
functions of many variables by superposition of continu- 2018, pp. 2887–2891.
ous functions of one variable and addition,” in Doklady [22] V. Giovannetti, S. Lloyd, and L. Maccone, “Quantum
Akademii Nauk, vol. 114, no. 5. Russian Academy of random access memory,” Physical review letters, vol.
Sciences, 1957, pp. 953–956. 100, no. 16, p. 160501, 2008.
[7] R. Hecht-Nielsen, “Kolmogorov’s mapping neural net- [23] M. A. Nielsen and I. Chuang, “Quantum computation
work existence theorem,” in Proceedings of the interna- and quantum information,” 2002.
tional conference on Neural Networks, vol. 3. IEEE [24] A. Pérez-Salinas, A. Cervera-Lierta, E. Gil-Fuster, and
Press New York, NY, USA, 1987, pp. 11–14. J. I. Latorre, “Data re-uploading for a universal quantum
[8] G. Cybenko, “Approximation by superpositions of a classifier,” Quantum, vol. 4, p. 226, 2020.
sigmoidal function,” Mathematics of control, signals and [25] M. Schuld and N. Killoran, “Quantum machine learning
systems, vol. 2, no. 4, pp. 303–314, 1989. in feature hilbert spaces,” Physical review letters, vol.
[9] A. R. Gallant and H. White, “There exists a neural 122, no. 4, p. 040504, 2019.
network that does not make avoidable mistakes.” in [26] E. Grant, M. Benedetti, S. Cao, A. Hallam, J. Lockhart,
ICNN, 1988, pp. 657–664. V. Stojevic, A. G. Green, and S. Severini, “Hierarchical
[10] Z.-Q. John Xu, Y. Zhang, T. Luo, Y. Xiao, and quantum classifiers,” npj Quantum Information, vol. 4,
Z. Ma, “Frequency principle: Fourier analysis sheds no. 1, pp. 1–8, 2018.
light on deep neural networks,” Communications in [27] H. Buhrman, R. Cleve, J. Watrous, and R. De Wolf,
Computational Physics, vol. 28, no. 5, pp. 1746– “Quantum fingerprinting,” Physical Review Letters,
1767, 2020. [Online]. Available: http://global-sci.org/ vol. 87, no. 16, p. 167902, 2001.
intro/article detail/cicp/18395.html [28] D. W. Berry, A. M. Childs, R. Cleve, R. Kothari, and
[11] C. C. Aggarwal et al., Data mining: the textbook. R. D. Somma, “Exponential improvement in precision
Springer, 2015, vol. 1. for simulating sparse hamiltonians,” in Forum of Math-
[12] H. Ismail Fawaz, G. Forestier, J. Weber, L. Idoumghar, ematics, Sigma, vol. 5. Cambridge University Press,
and P.-A. Muller, “Deep learning for time series classifi- 2017.
cation: a review,” Data mining and knowledge discovery, [29] A. Daskin, A. Grama, G. Kollias, and S. Kais, “Universal
vol. 33, no. 4, pp. 917–963, 2019. programmable quantum circuit schemes to emulate an
[13] A. Abanda, U. Mori, and J. A. Lozano, “A review on operator,” The Journal of chemical physics, vol. 137,
distance based time series classification,” Data Mining no. 23, p. 234112, 2012.
and Knowledge Discovery, vol. 33, no. 2, pp. 378–412, [30] D. S. G. Pollock, “Circulant matrices and time-series
2019. analysis,” International Journal of Mathematical Edu-
8

cation in Science and Technology, vol. 33, no. 2, pp.


213–230, 2002.
[31] A. Daskin, “Quantum implementation of circulant ma-
trices and its use in quantum string processing,” arXiv
preprint arXiv:2206.09364, 2022.
[32] “Prophet,” accessed: Feb. 15, 2022. [Online]. Available:
https://facebook.github.io/prophet/
[33] “Qiskit,” accessed: Feb. 15, 2022. [Online]. Available:
https://qiskit.org/
[34] G. P. Zhang, “Time series forecasting using a hy-
brid arima and neural network model,” Neurocomputing,
vol. 50, pp. 159–175, 2003.
[35] Y. Peng, B. Choi, and J. Xu, “Graph learning for combi-
natorial optimization: A survey of state-of-the-art,” Data
Science and Engineering, vol. 6, no. 2, pp. 119–141,
2021.
[36] R. Orús, S. Mugel, and E. Lizaso, “Forecasting financial
crashes with quantum computing,” Physical Review A,
vol. 99, no. 6, p. 060301, 2019.
[37] E. Farhi, J. Goldstone, S. Gutmann, and M. Sipser,
“Quantum computation by adiabatic evolution,” arXiv
preprint quant-ph/0001106, 2000.
[38] E. Farhi, J. Goldstone, and S. Gutmann, “A quan-
tum approximate optimization algorithm,” arXiv preprint
arXiv:1411.4028, 2014.
[39] A. Daskin, “Combinatorial optimization through varia-
tional quantum power method,” Quantum Information
Processing, vol. 20, no. 10, pp. 1–14, 2021.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy