送男生什么生日礼物好| 流脑是什么| 属猴本命佛是什么佛| 狗狗能吃什么水果| 火龙果什么时候开花| 手机为什么没信号| 辛亥革命是什么时候| 神母是什么病| 屁股长痣代表什么| 骞字五行属什么| 11月出生是什么星座| ig是什么意思| 蟑螂对人体有什么危害| app有什么用途| 南瓜什么季节成熟| 沙示汽水有什么功效| ABB式的词语有什么| 什么护肤产品补水最好| 吃什么东西可以降压| 鼻涕倒流到咽喉老吐痰吃什么药能根治| 上厕所出血是什么原因| 卵巢筛查要做什么检查| 白带有血丝是什么情况| 减肥期间能吃什么水果| 羊水栓塞是什么原因引起的| 老人头晕是什么原因引起的| 赫尔墨斯是什么神| 早筛是检查什么项目| 地什么人什么| 眼神迷离什么意思| 什么叫做犯太岁| 头晕是什么病的征兆| 5.13是什么星座| 双鱼座的幸运色是什么颜色| 属狗男和什么属相最配| 什么猫最贵| 七月18日是什么星座| via什么意思| 景象什么意思| 胸下面是什么部位| 早上右眼跳是什么预兆| 02年的马是什么命| 生活质量是什么意思| 中秋节送礼送什么| 下面瘙痒用什么药| nb是什么牌子| 良缘是什么意思| 天鹅吃什么| 孕中期失眠是什么原因| 梦见房子倒塌是什么意思| 孕前检查一般有什么项目| 什么鬼什么神| 检查胸部应该挂什么科| 4级残疾证有什么优惠政策| 一什么香蕉| 玮五行属什么| 什么叫做犯太岁| 鲜为人知是什么意思| 心脏造影是什么意思| 餐标是什么意思| 狗狗流鼻涕吃什么药| 福相是什么意思| 武夷山在什么地方| 什么东西有脚却不能走路| 抗环瓜氨酸肽抗体高是什么意思| ich是什么意思| 西五行属什么| 秋高气爽是什么意思| 郡字五行属什么| 李子和什么不能一起吃| 打边炉是什么| 海龟汤是什么| 女人脚肿是什么原因| 男生做爱什么感觉| 乙肝复查检查什么项目| 辣椒炭疽病用什么药| 广州机场叫什么名字| 太公是什么意思| 不孝有三无后为大是什么意思| 南瓜与什么食物相克| 仓鼠不能吃什么| 浮肿是什么原因造成的| 取鱼刺挂什么科室| 西瓜配什么榨汁好喝| 总打嗝吃什么药| 家庭出身填什么| 属猴的幸运色是什么颜色| 汉武帝属什么生肖| 在减肥期间吃什么最好| 胃胀是什么症状| 做不好的梦预示着什么| gas是什么意思| 阴道口溃疡用什么药| 来大姨妈可以吃什么水果| 冲羊煞东是什么意思| 出生医学证明有什么用| 2020年是什么命| 小儿鼻炎用什么药好| 什么是过敏性紫癜| 检查肺部最好做什么检查最准确| 恶心吃什么药| 口什么心什么| 外阴白斑是什么原因| 发烧应该吃什么药| 前胸后背长痘痘用什么药| 卵巢囊肿有什么症状| 大运流年是什么意思| 什么球身上长毛| 香五行属什么| 81年属什么的| 小孩吃指甲是什么原因造成的| 血压太低会有什么危险| 十全十美是什么生肖| 儿童包皮过长挂什么科| 冠脉硬化什么意思| 咽口水喉咙痛吃什么药| 人绒毛膜促性腺激素是查什么的| 火烈鸟为什么是红色的| 尿酸高什么水果不能吃| 马子什么意思| 做梦掉牙齿是什么意思| 高铁座位为什么没有e座| 生目念什么| 撤退性出血什么意思| 吃什么能快速减肥| 吃虫草有什么好处| polo villae是什么档次| ct 是什么| 乳腺瘤不能吃什么| 洋地黄是什么药| 六月二号是什么星座| psa是什么| 肌酸粉有什么作用| 椎间盘突出挂什么科| 老公什么意思| 后背痛是什么原因| 梦见黑色的蛇是什么意思| 羊肉馅饺子放什么菜| 爱慕内衣什么档次| 日央读什么| 余沧海为什么是两个人| 月经第二天是什么期| 打猎是什么意思| 凌乱是什么意思| 97年的属什么生肖| 孕晚期宫缩是什么感觉| 元五行属什么| 精液是什么味| 养病是什么意思| 太阳筋疼是什么原因| 甲状腺结节是什么病| 什么是假性抑郁症| 迄今为止什么意思| 电解质氯高是什么原因| 什么情况下做心脏造影| 大人睡觉流口水是什么原因引起的| 加拿大属于什么洲| 甘油三酯高是什么原因| 男朋友有什么用| 等是什么生肖| 诸事不宜什么意思| 前额头疼是什么原因引起的| 急性扁桃体发炎吃什么药| 喉咙痒是什么原因引起的| 阿玛尼手表属于什么档次| 肝不好有什么症状有哪些表现| 五脏六腑指什么| 10.11是什么星座| 庸人自扰之是什么意思| 成人发烧吃什么药| 马铃薯什么时候传入中国| 什么汤养胃| penis什么意思| 降钙素原偏高说明什么| 莫逆是什么意思| 一个西一个米念什么| 付诸行动是什么意思| 室性早搏吃什么药最好| 鱼子酱是什么鱼的鱼子| 梦见红薯是什么意思| 黄体酮低吃什么补得快| 冰箱什么牌子的好| 促甲状腺素高是什么原因| 赵本山什么时候去世的| 阳光灿烂是什么意思| 生吃大蒜有什么好处| 小孩睡觉张开嘴巴是什么原因| 头疼是因为什么| 把碗打碎了有什么征兆| 月经量太少是什么原因引起的| 紫水晶五行属什么| 肝不好吃什么药| 前列腺炎是什么原因引起| 什么是八字生辰八字| 珍珠粉加蜂蜜做面膜有什么作用| 不将就什么意思| 血细胞分析是查什么的| 11月27日是什么星座| 什么是外阴炎| 深圳副市长什么级别| 逆生长是什么意思| 80年出生属什么生肖| 磨平了棱角是什么意思| 烦恼千千是什么生肖| 康养中心是做什么的| 牙膏什么牌子好| 我在你心里是什么颜色| 尿发黄是什么原因男性| 什么时候大暑| 草鱼吃什么| 毕业送什么花| 冬瓜有什么功效和作用| 镭射有什么危害| 宝宝不爱喝水有什么好的办法吗| 麝香是什么| 粿条是什么做的| 做梦梦到水是什么征兆| 3月9日什么星座| 免疫力低会引起什么病| 指骨属于什么骨| 什么农药最毒| 疝气是什么病怎样治疗| 肛门瘙痒用什么药| 乙肝表面抗体定量偏高什么意思| 赵本山什么时候去世的| 肌酸激酶是什么| 什么红| 啤酒鸭可以放什么配菜| 身上没力气没劲是什么原因| 胃火喝什么茶降火| 痛经喝什么| 流产后吃什么水果好| 黄色加蓝色等于什么颜色| 减肥挂什么科| ua是什么单位| 伦字五行属什么| 灰指甲长什么样子图片| 夏天穿什么鞋子| 子宫前置是什么意思| 黑醋是什么醋| 黑色五行属什么| 小腿经常抽筋是什么原因| 相性是什么意思| 虾仁炒什么菜好吃| 屁多吃什么药| 毛毛虫吃什么食物| 痣的位置代表什么| 做扩胸运动有什么好处| 江郎才尽是什么意思| 兵马未动粮草先行是什么意思| 脂肪肝喝什么茶最好最有效| 颈椎痛挂什么科| 嘴里苦是什么原因| 下巴脱臼是什么感觉| 甲胎蛋白偏低说明什么| 佛珠什么材质的最好| 煮玉米加什么才会香甜| 老鸨是什么意思| 利妥昔单抗是治什么病| 筋膜炎吃什么药好得快| 腱鞘炎什么症状| 心脏舒张功能减低是什么意思| 118号是什么星座| 干咳喝什么止咳糖浆好| 百度Jump to content

报告建议各方合力推动女性职业发展

From Wikipedia, the free encyclopedia
百度 工作室人手紧张,有的历史人物也真的不好找历史资料,于是这对父子只好采用了千人一面这种神操作。

In statistics, econometrics, and signal processing, an autoregressive (AR) model is a representation of a type of random process; as such, it can be used to describe certain time-varying processes in nature, economics, behavior, etc. The autoregressive model specifies that the output variable depends linearly on its own previous values and on a stochastic term (an imperfectly predictable term); thus the model is in the form of a stochastic difference equation (or recurrence relation) which should not be confused with a differential equation. Together with the moving-average (MA) model, it is a special case and key component of the more general autoregressive–moving-average (ARMA) and autoregressive integrated moving average (ARIMA) models of time series, which have a more complicated stochastic structure; it is also a special case of the vector autoregressive model (VAR), which consists of a system of more than one interlocking stochastic difference equation in more than one evolving random variable. Another important extension is the time-varying autoregressive (TVAR) model, where the autoregressive coefficients are allowed to change over time to model evolving or non-stationary processes. TVAR models are widely applied in cases where the underlying dynamics of the system are not constant, such as in sensors time series modelling[1][2], finance[3], climate science[4], economics[5], signal processing[6] and telecommunications[7], radar systems[8], and biological signals[9].

Unlike the moving-average (MA) model, the autoregressive model is not always stationary; non-stationarity can arise either due to the presence of a unit root or due to time-varying model parameters, as in time-varying autoregressive (TVAR) models.

Large language models are called autoregressive, but they are not a classical autoregressive model in this sense because they are not linear.

Definition

[edit]

The notation indicates an autoregressive model of order p. The AR(p) model is defined as

where are the parameters of the model, and is white noise.[10][11] This can be equivalently written using the backshift operator B as

so that, moving the summation term to the left side and using polynomial notation, we have

An autoregressive model can thus be viewed as the output of an all-pole infinite impulse response filter whose input is white noise.

Some parameter constraints are necessary for the model to remain weak-sense stationary. For example, processes in the AR(1) model with are not stationary. More generally, for an AR(p) model to be weak-sense stationary, the roots of the polynomial must lie outside the unit circle, i.e., each (complex) root must satisfy (see pages 89,92 [12]).

Intertemporal effect of shocks

[edit]

In an AR process, a one-time shock affects values of the evolving variable infinitely far into the future. For example, consider the AR(1) model . A non-zero value for at say time t=1 affects by the amount . Then by the AR equation for in terms of , this affects by the amount . Then by the AR equation for in terms of , this affects by the amount . Continuing this process shows that the effect of never ends, although if the process is stationary then the effect diminishes toward zero in the limit.

Because each shock affects X values infinitely far into the future from when they occur, any given value Xt is affected by shocks occurring infinitely far into the past. This can also be seen by rewriting the autoregression

(where the constant term has been suppressed by assuming that the variable has been measured as deviations from its mean) as

When the polynomial division on the right side is carried out, the polynomial in the backshift operator applied to has an infinite order—that is, an infinite number of lagged values of appear on the right side of the equation.

Characteristic polynomial

[edit]

The autocorrelation function of an AR(p) process can be expressed as [citation needed]

where are the roots of the polynomial

where B is the backshift operator, where is the function defining the autoregression, and where are the coefficients in the autoregression. The formula is valid only if all the roots have multiplicity 1.[citation needed]

The autocorrelation function of an AR(p) process is a sum of decaying exponentials.

  • Each real root contributes a component to the autocorrelation function that decays exponentially.
  • Similarly, each pair of complex conjugate roots contributes an exponentially damped oscillation.

Graphs of AR(p) processes

[edit]
"Figure has 5 plots of AR processes. AR(0) and AR(0.3) are white noise or look like white noise. AR(0.9) has some large scale oscillating structure."
AR(0); AR(1) with AR parameter 0.3; AR(1) with AR parameter 0.9; AR(2) with AR parameters 0.3 and 0.3; and AR(2) with AR parameters 0.9 and ?0.8

The simplest AR process is AR(0), which has no dependence between the terms. Only the error/innovation/noise term contributes to the output of the process, so in the figure, AR(0) corresponds to white noise.

For an AR(1) process with a positive , only the previous term in the process and the noise term contribute to the output. If is close to 0, then the process still looks like white noise, but as approaches 1, the output gets a larger contribution from the previous term relative to the noise. This results in a "smoothing" or integration of the output, similar to a low pass filter.

For an AR(2) process, the previous two terms and the noise term contribute to the output. If both and are positive, the output will resemble a low pass filter, with the high frequency part of the noise decreased. If is positive while is negative, then the process favors changes in sign between terms of the process. The output oscillates. This can be linked to edge detection or detection of change in direction.

Example: An AR(1) process

[edit]

An AR(1) process is given by:where is a white noise process with zero mean and constant variance . (Note: The subscript on has been dropped.) The process is weak-sense stationary if since it is obtained as the output of a stable filter whose input is white noise. (If then the variance of depends on time lag t, so that the variance of the series diverges to infinity as t goes to infinity, and is therefore not weak-sense stationary.) Assuming , the mean is identical for all values of t by definition of weak sense stationarity. If the mean is denoted by , it follows fromthatand hence

The variance is

where is the standard deviation of . This can be shown by noting that

and then by noticing that the quantity above is a stable fixed point of this relation.

The autocovariance is given by

It can be seen that the autocovariance function decays with a decay time (also called time constant) of .[13]

The spectral density function is the Fourier transform of the autocovariance function. In discrete terms this will be the discrete-time Fourier transform:

This expression is periodic due to the discrete nature of the , which is manifested as the cosine term in the denominator. If we assume that the sampling time () is much smaller than the decay time (), then we can use a continuum approximation to :

which yields a Lorentzian profile for the spectral density:

where is the angular frequency associated with the decay time .

An alternative expression for can be derived by first substituting for in the defining equation. Continuing this process N times yields

For N approaching infinity, will approach zero and:

It is seen that is white noise convolved with the kernel plus the constant mean. If the white noise is a Gaussian process then is also a Gaussian process. In other cases, the central limit theorem indicates that will be approximately normally distributed when is close to one.

For , the process will be a geometric progression (exponential growth or decay). In this case, the solution can be found analytically: whereby is an unknown constant (initial condition).

Explicit mean/difference form of AR(1) process

[edit]

The AR(1) model is the discrete-time analogy of the continuous Ornstein-Uhlenbeck process. It is therefore sometimes useful to understand the properties of the AR(1) model cast in an equivalent form. In this form, the AR(1) model, with process parameter , is given by

, where , is the model mean, and is a white-noise process with zero mean and constant variance .

By rewriting this as and then deriving (by induction) , one can show that

and
.

Choosing the maximum lag

[edit]

The partial autocorrelation of an AR(p) process equals zero at lags larger than p, so the appropriate maximum lag p is the one after which the partial autocorrelations are all zero.

Calculation of the AR parameters

[edit]

There are many ways to estimate the coefficients, such as the ordinary least squares procedure or method of moments (through Yule–Walker equations).

The AR(p) model is given by the equation

It is based on parameters where i = 1, ..., p. There is a direct correspondence between these parameters and the covariance function of the process, and this correspondence can be inverted to determine the parameters from the autocorrelation function (which is itself obtained from the covariances). This is done using the Yule–Walker equations.

Yule–Walker equations

[edit]

The Yule–Walker equations, named for Udny Yule and Gilbert Walker,[14][15] are the following set of equations.[16]

where m?=?0,?…,?p, yielding p?+?1 equations. Here is the autocovariance function of Xt, is the standard deviation of the input noise process, and is the Kronecker delta function.

Because the last part of an individual equation is non-zero only if m?=?0, the set of equations can be solved by representing the equations for m?>?0 in matrix form, thus getting the equation

which can be solved for all The remaining equation for m = 0 is

which, once are known, can be solved for

An alternative formulation is in terms of the autocorrelation function. The AR parameters are determined by the first p+1 elements of the autocorrelation function. The full autocorrelation function can then be derived by recursively calculating [17]

Examples for some Low-order AR(p) processes

  • p=1
    • Hence
  • p=2
    • The Yule–Walker equations for an AR(2) process are
      • Remember that
      • Using the first equation yields
      • Using the recursion formula yields

Estimation of AR parameters

[edit]

The above equations (the Yule–Walker equations) provide several routes to estimating the parameters of an AR(p) model, by replacing the theoretical covariances with estimated values.[18] Some of these variants can be described as follows:

  • Estimation of autocovariances or autocorrelations. Here each of these terms is estimated separately, using conventional estimates. There are different ways of doing this and the choice between these affects the properties of the estimation scheme. For example, negative estimates of the variance can be produced by some choices.
  • Formulation as a least squares regression problem in which an ordinary least squares prediction problem is constructed, basing prediction of values of Xt on the p previous values of the same series. This can be thought of as a forward-prediction scheme. The normal equations for this problem can be seen to correspond to an approximation of the matrix form of the Yule–Walker equations in which each appearance of an autocovariance of the same lag is replaced by a slightly different estimate.
  • Formulation as an extended form of ordinary least squares prediction problem. Here two sets of prediction equations are combined into a single estimation scheme and a single set of normal equations. One set is the set of forward-prediction equations and the other is a corresponding set of backward prediction equations, relating to the backward representation of the AR model:
Here predicted values of Xt would be based on the p future values of the same series.[clarification needed] This way of estimating the AR parameters is due to John Parker Burg,[19] and is called the Burg method:[20] Burg and later authors called these particular estimates "maximum entropy estimates",[21] but the reasoning behind this applies to the use of any set of estimated AR parameters. Compared to the estimation scheme using only the forward prediction equations, different estimates of the autocovariances are produced, and the estimates have different stability properties. Burg estimates are particularly associated with maximum entropy spectral estimation.[22]

Other possible approaches to estimation include maximum likelihood estimation. Two distinct variants of maximum likelihood are available: in one (broadly equivalent to the forward prediction least squares scheme) the likelihood function considered is that corresponding to the conditional distribution of later values in the series given the initial p values in the series; in the second, the likelihood function considered is that corresponding to the unconditional joint distribution of all the values in the observed series. Substantial differences in the results of these approaches can occur if the observed series is short, or if the process is close to non-stationarity.

Spectrum

[edit]

The power spectral density (PSD) of an AR(p) process with noise variance is[17]

AR(0)

[edit]

For white noise (AR(0))

AR(1)

[edit]

For AR(1)

  • If there is a single spectral peak at , often referred to as red noise. As becomes nearer 1, there is stronger power at low frequencies, i.e. larger time lags. This is then a low-pass filter, when applied to full spectrum light, everything except for the red light will be filtered.
  • If there is a minimum at , often referred to as blue noise. This similarly acts as a high-pass filter, everything except for blue light will be filtered.

AR(2)

[edit]

The behavior of an AR(2) process is determined entirely by the roots of it characteristic equation, which is expressed in terms of the lag operator as:

or equivalently by the poles of its transfer function, which is defined in the Z domain by:

It follows that the poles are values of z satisfying:

,

which yields:

.

and are the reciprocals of the characteristic roots, as well as the eigenvalues of the temporal update matrix:

AR(2) processes can be split into three groups depending on the characteristics of their roots/poles:

  • When , the process has a pair of complex-conjugate poles, creating a mid-frequency peak at:

with bandwidth about the peak inversely proportional to the moduli of the poles:

The terms involving square roots are all real in the case of complex poles since they exist only when .

Otherwise the process has real roots, and:

  • When it acts as a low-pass filter on the white noise with a spectral peak at
  • When it acts as a high-pass filter on the white noise with a spectral peak at .

The process is non-stationary when the poles are on or outside the unit circle, or equivalently when the characteristic roots are on or inside the unit circle. The process is stable when the poles are strictly within the unit circle (roots strictly outside the unit circle), or equivalently when the coefficients are in the triangle .

The full PSD function can be expressed in real form as:

Implementations in statistics packages

[edit]
  • R – the stats package includes ar function;[23] the astsa package includes sarima function to fit various models including AR.[24]
  • MATLAB – the Econometrics Toolbox[25] and System Identification Toolbox[26] include AR models.[27]
  • MATLAB and Octave – the TSA toolbox contains several estimation functions for uni-variate, multivariate, and adaptive AR models.[28]
  • PyMC3 – the Bayesian statistics and probabilistic programming framework supports AR modes with p lags.
  • bayesloop – supports parameter inference and model selection for the AR-1 process with time-varying parameters.[29]
  • Python – statsmodels.org hosts an AR model.[30]

Impulse response

[edit]

The impulse response of a system is the change in an evolving variable in response to a change in the value of a shock term k periods earlier, as a function of k. Since the AR model is a special case of the vector autoregressive model, the computation of the impulse response in vector autoregression#impulse response applies here.

n-step-ahead forecasting

[edit]

Once the parameters of the autoregression

have been estimated, the autoregression can be used to forecast an arbitrary number of periods into the future. First use t to refer to the first period for which data is not yet available; substitute the known preceding values Xt-i for i=1, ..., p into the autoregressive equation while setting the error term equal to zero (because we forecast Xt to equal its expected value, and the expected value of the unobserved error term is zero). The output of the autoregressive equation is the forecast for the first unobserved period. Next, use t to refer to the next period for which data is not yet available; again the autoregressive equation is used to make the forecast, with one difference: the value of X one period prior to the one now being forecast is not known, so its expected value—the predicted value arising from the previous forecasting step—is used instead. Then for future periods the same procedure is used, each time using one more forecast value on the right side of the predictive equation until, after p predictions, all p right-side values are predicted values from preceding steps.

There are four sources of uncertainty regarding predictions obtained in this manner: (1) uncertainty as to whether the autoregressive model is the correct model; (2) uncertainty about the accuracy of the forecasted values that are used as lagged values in the right side of the autoregressive equation; (3) uncertainty about the true values of the autoregressive coefficients; and (4) uncertainty about the value of the error term for the period being predicted. Each of the last three can be quantified and combined to give a confidence interval for the n-step-ahead predictions; the confidence interval will become wider as n increases because of the use of an increasing number of estimated values for the right-side variables.

See also

[edit]

Notes

[edit]
  1. ^ Souza, Douglas Baptista de; Leao, Bruno Paes (26 October 2023). "Data Augmentation of Sensor Time Series using Time-varying Autoregressive Processes". Annual Conference of the PHM Society. 15 (1). doi:10.36001/phmconf.2023.v15i1.3565.
  2. ^ Souza, Douglas Baptista de; Leao, Bruno Paes (5 November 2024). "Data Augmentation of Multivariate Sensor Time Series using Autoregressive Models and Application to Failure Prognostics". Annual Conference of the PHM Society. 16 (1). doi:10.36001/phmconf.2024.v16i1.4145.
  3. ^ Jia, Zhixuan; Li, Wang; Jiang, Yunlong; Liu, Xingshen (9 July 2025). "The Use of Minimization Solvers for Optimizing Time-Varying Autoregressive Models and Their Applications in Finance". Mathematics. 13 (14): 2230. doi:10.3390/math13142230.
  4. ^ Diodato, Nazzareno; Di Salvo, Cristina; Bellocchi, Gianni (18 March 2025). "Climate driven generative time-varying model for improved decadal storm power predictions in the Mediterranean". Communications Earth & Environment. 6 (1): 212. Bibcode:2025ComEE...6..212D. doi:10.1038/s43247-025-02196-2.
  5. ^ Inayati, Syarifah; Iriawan, Nur (31 December 2024). "Time-Varying Autoregressive Models for Economic Forecasting". Matematika: 131–142. doi:10.11113/matematika.v40.n3.1654.
  6. ^ Baptista de Souza, Douglas; Kuhn, Eduardo Vinicius; Seara, Rui (January 2019). "A Time-Varying Autoregressive Model for Characterizing Nonstationary Processes". IEEE Signal Processing Letters. 26 (1): 134–138. Bibcode:2019ISPL...26..134B. doi:10.1109/LSP.2018.2880086.
  7. ^ Wang, Shihan; Chen, Tao; Wang, Hongjian (17 March 2023). "IDBD-Based Beamforming Algorithm for Improving the Performance of Phased Array Radar in Nonstationary Environments". Sensors. 23 (6): 3211. Bibcode:2023Senso..23.3211W. doi:10.3390/s23063211. PMC 10052024. PMID 36991922.
  8. ^ Abramovich, Yuri I.; Spencer, Nicholas K.; Turley, Michael D. E. (April 2007). "Time-Varying Autoregressive (TVAR) Models for Multiple Radar Observations". IEEE Transactions on Signal Processing. 55 (4): 1298–1311. Bibcode:2007ITSP...55.1298A. doi:10.1109/TSP.2006.888064.
  9. ^ Gutierrez, D.; Salazar-Varas, R. (August 2011). "EEG signal classification using time-varying autoregressive models and common spatial patterns". 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society. pp. 6585–6588. doi:10.1109/IEMBS.2011.6091624. ISBN 978-1-4577-1589-1. PMID 22255848.
  10. ^ Box, George E. P. (1994). Time series analysis : forecasting and control. Gwilym M. Jenkins, Gregory C. Reinsel (3rd ed.). Englewood Cliffs, N.J.: Prentice Hall. p. 54. ISBN 0-13-060774-6. OCLC 28888762.
  11. ^ Shumway, Robert H. (2000). Time series analysis and its applications. David S. Stoffer. New York: Springer. pp. 90–91. ISBN 0-387-98950-1. OCLC 42392178.
  12. ^ Shumway, Robert H.; Stoffer, David (2010). Time series analysis and its applications : with R examples (3rd ed.). Springer. ISBN 978-1441978646.
  13. ^ Lai, Dihui; and Lu, Bingfeng; "Understanding Autoregressive Model for Time Series as a Deterministic Dynamic System" Archived 2025-08-14 at the Wayback Machine, in Predictive Analytics and Futurism, June 2017, number 15, June 2017, pages 7-9
  14. ^ Yule, G. Udny (1927) "On a Method of Investigating Periodicities in Disturbed Series, with Special Reference to Wolfer's Sunspot Numbers" Archived 2025-08-14 at the Wayback Machine, Philosophical Transactions of the Royal Society of London, Ser. A, Vol. 226, 267–298.]
  15. ^ Walker, Gilbert (1931) "On Periodicity in Series of Related Terms" Archived 2025-08-14 at the Wayback Machine, Proceedings of the Royal Society of London, Ser. A, Vol. 131, 518–532.
  16. ^ Theodoridis, Sergios (2025-08-14). "Chapter 1. Probability and Stochastic Processes". Machine Learning: A Bayesian and Optimization Perspective. Academic Press, 2015. pp. 9–51. ISBN 978-0-12-801522-3.
  17. ^ a b Von Storch, Hans; Zwiers, Francis W. (2001). Statistical analysis in climate research. Cambridge University Press. doi:10.1017/CBO9780511612336. ISBN 0-521-01230-9.[page needed]
  18. ^ Eshel, Gidon. "The Yule Walker Equations for the AR Coefficients" (PDF). stat.wharton.upenn.edu. Archived (PDF) from the original on 2025-08-14. Retrieved 2025-08-14.
  19. ^ Burg, John Parker (1968); "A new analysis technique for time series data", in Modern Spectrum Analysis (Edited by D. G. Childers), NATO Advanced Study Institute of Signal Processing with emphasis on Underwater Acoustics. IEEE Press, New York.
  20. ^ Brockwell, Peter J.; Dahlhaus, Rainer; Trindade, A. Alexandre (2005). "Modified Burg Algorithms for Multivariate Subset Autoregression" (PDF). Statistica Sinica. 15: 197–213. Archived from the original (PDF) on 2025-08-14.
  21. ^ Burg, John Parker (1967) "Maximum Entropy Spectral Analysis", Proceedings of the 37th Meeting of the Society of Exploration Geophysicists, Oklahoma City, Oklahoma.
  22. ^ Bos, Robert; De Waele, Stijn; Broersen, Piet M. T. (2002). "Autoregressive spectral estimation by application of the Burg algorithm to irregularly sampled data". IEEE Transactions on Instrumentation and Measurement. 51 (6): 1289. Bibcode:2002ITIM...51.1289B. doi:10.1109/TIM.2002.808031. Archived from the original on 2025-08-14. Retrieved 2025-08-14.
  23. ^ "Fit Autoregressive Models to Time Series" Archived 2025-08-14 at the Wayback Machine (in R)
  24. ^ Stoffer, David; Poison, Nicky (2025-08-14). "astsa: Applied Statistical Time Series Analysis". Retrieved 2025-08-14.
  25. ^ "Econometrics Toolbox". www.mathworks.com. Archived from the original on 2025-08-14. Retrieved 2025-08-14.
  26. ^ "System Identification Toolbox". www.mathworks.com. Archived from the original on 2025-08-14. Retrieved 2025-08-14.
  27. ^ "Autoregressive Model - MATLAB & Simulink". www.mathworks.com. Archived from the original on 2025-08-14. Retrieved 2025-08-14.
  28. ^ "The Time Series Analysis (TSA) toolbox for Octave and MATLAB". pub.ist.ac.at. Archived from the original on 2025-08-14. Retrieved 2025-08-14.
  29. ^ "christophmark/bayesloop". December 7, 2021. Archived from the original on September 28, 2020. Retrieved September 4, 2018 – via GitHub.
  30. ^ "statsmodels.tsa.ar_model.AutoReg — statsmodels 0.12.2 documentation". www.statsmodels.org. Archived from the original on 2025-08-14. Retrieved 2025-08-14.

References

[edit]
[edit]
什么体质不易怀孕 胆管炎吃什么药 六月十号是什么星座 ns是什么单位 空腹不能吃什么水果
cet什么意思 68年属猴的是什么命 医保和农村合作医疗有什么区别 茶花什么时候开花 梦见老人去世预示什么
湿疹吃什么药好 舌头痛吃什么药好得快 狗狗细小是什么症状 治便秘什么药最好 滴水不漏什么意思
男同是什么 肺看什么科室 注意身体是什么意思 喉咙痒咳嗽吃什么药好 泄泻病是什么意思
top1什么意思hcv8jop5ns8r.cn 438是什么意思qingzhougame.com 脚上有青筋是什么原因hcv8jop4ns8r.cn 汤姆猫是什么品种hcv8jop4ns1r.cn 夏天受凉感冒吃什么药hcv8jop5ns3r.cn
褥疮是什么yanzhenzixun.com 什么叫做hcv9jop3ns6r.cn 逆时针是什么方向hcv9jop6ns5r.cn 为什么总是头疼hcv9jop8ns1r.cn 接触性皮炎用什么药hcv7jop5ns0r.cn
90年属于什么生肖hcv8jop0ns4r.cn 尼古丁是什么东西shenchushe.com 蛋疼是什么原因引起的beikeqingting.com 四月初一是什么星座hcv8jop5ns3r.cn 叶酸片有什么作用hcv8jop4ns8r.cn
机化是什么意思hcv9jop4ns2r.cn 蛇鼠一窝指什么生肖hcv9jop3ns2r.cn 人流后可以吃什么hcv9jop8ns3r.cn 隆鼻后吃什么消肿快hcv8jop7ns8r.cn 栀子有什么作用与功效hcv9jop6ns9r.cn
百度