SlideShare a Scribd company logo
week 2 1
Stochastic Process - Introduction
• Stochastic processes are processes that proceed randomly in time.
• Rather than consider fixed random variables X, Y, etc. or even
sequences of i.i.d random variables, we consider sequences X0, X1,
X2, …. Where Xt represent some random quantity at time t.
• In general, the value Xt might depend on the quantity Xt-1 at time t-1,
or even the value Xs for other times s < t.
• Example: simple random walk .
week 2 2
STOCHASTIC PROCESSES
• DR.M.THIAGARAJAN ASSOCIATE
PROFESSOR OF MATHEMATICS
• ST.JOSEPH’S COLLEGE
• TRICHY-2
week 2 3
Stochastic Process - Definition
• A stochastic process is a family of time indexed random variables Xt
where t belongs to an index set. Formal notation, where I is
an index set that is a subset of R.
• Examples of index sets:
1) I = (-∞, ∞) or I = [0, ∞]. In this case Xt is a continuous time
stochastic process.
2) I = {0, ±1, ±2, ….} or I = {0, 1, 2, …}. In this case Xt is a discrete
time stochastic process.
• We use uppercase letter {Xt } to describe the process. A time series,
{xt } is a realization or sample function from a certain process.
• We use information from a time series to estimate parameters and
properties of process {Xt }.
 
I
t
Xt 
:
week 2 4
Probability Distribution of a Process
• For any stochastic process with index set I, its probability
distribution function is uniquely determined by its finite dimensional
distributions.
• The k dimensional distribution function of a process is defined by
for any and any real numbers x1, …, xk .
• The distribution function tells us everything we need to know about
the process {Xt }.
   
k
t
t
k
X
X x
X
x
X
P
x
x
F k
k
t
t


 ,...,
,..., 1
1
,..., 1
1
I
t
t k 
,...,
1
week 2 5
Moments of Stochastic Process
• We can describe a stochastic process via its moments, i.e.,
We often use the first two moments.
• The mean function of the process is
• The variance function of the process is
• The covariance function between Xt , Xs is
• The correlation function between Xt , Xs is
• These moments are often function of time.
      etc.
,
, 2
s
t
t
t X
X
E
X
E
X
E 
  .
t
t
X
E 

  .
2
t
t
X
Var 

    
 
s
s
t
t
s
t X
X
E
X
X 
 


,
Cov
 
 
2
2
,
Cov
,
s
t
s
t
s
t
X
X
X
X


 
week 2 6
Stationary Processes
• A process is said to be strictly stationary if has the same
joint distribution as . That is, if
• If {Xt } is a strictly stationary process and then, the mean
function is a constant and the variance function is also a constant.
• Moreover, for a strictly stationary process with first two moments
finite, the covariance function, and the correlation function depend
only on the time difference s.
• A trivial example of a strictly stationary process is a sequence of
i.i.d random variables.
 
k
t
t X
X ,...,
1
 



 k
t
t X
X ,...,
1
   
k
X
X
k
X
X x
x
F
x
x
F
k
t
t
k
t
t
,...,
,..., 1
,...,
1
,..., 1
1 




  

2
t
X
E
week 2 7
Weak Stationarity
• Strict stationarity is too strong of a condition in practice. It is often
difficult assumption to assess based on an observed time series x1,…,xk.
• In time series analysis we often use a weaker sense of stationarity in
terms of the moments of the process.
• A process is said to be nth-order weakly stationary if all its joint
moments up to order n exists and are time invariant, i.e., independent of
time origin.
• For example, a second-order weakly stationary process will have
constant mean and variance, with the covariance and the correlation
being functions of the time difference along.
• A strictly stationary process with the first two moments finite is also a
second-ordered weakly stationary. But a strictly stationary process may
not have finite moments and therefore may not be weakly stationary.
week 2 8
The Autocovariance and Autocorrelation Functions
• For a stationary process {Xt }, with constant mean μ and constant
variance σ2. The covariance between Xt and Xt+s is
• The correlation between Xt and Xt+s is
Where
• As functions of s, γ(s) is called the autocovariance function and ρ(s)
is called the autocorrelation function (ATF). They represent the
covariance and correlation between Xt and Xt+s from the same process,
separated only by s time lags.
      
 


 


  s
t
s
t
t X
X
E
X
X
s ,
cov
 
 
   
 
 
0
var
var
,
cov



s
X
X
X
X
s
s
t
s
t


     .
0
var
var 

 s
t
t X
X
week 2 9
Properties of γ(s) and ρ(s)
• For a stationary process, the autocovariance function γ(s) and the
autocorrelation function ρ(s) have the following properties:

 .
 .
 The autocovariance function γ(s) and the autocorrelation
function ρ(s) are positive semidefinite in the sense that
for any real numbers
      .
1
0
;
var
0 
 
 t
X
  1
1 

 s

       
s
s
s 


 


 s
and
  and
0
1 1



 
n
i
n
j
j
i j
i


   0
1 1



 
n
i
n
j
j
i j
i



.
,...,
, 2
1 n



week 2 10
Correlogram
• A correlogram is a plot of the autocorrelation function ρ(s) versus
the lag s where s = 0,1, ….
• Example…
week 2 11
Partial Autocorrelation Function
• Often we want to investigate the dependency / association between
Xt and Xt+k adjusting for their dependency on Xt+1, Xt+2,…, Xt+k-1.
• The conditional correlation Corr(Xt , Xt+k | Xt+1, Xt+2,…, Xt+k-1) is
usually referred to as the partial correlation in time series analysis.
• Partial autocorrelation is usually useful for identifying
autoregressive models.
week 2 12
Gaussian process
• A stochastic process is said to be a normal or Gaussian process if its
joint probability distribution is normal.
• A Gaussian process is strictly and weakly stationary because the
normal distribution is uniquely characterized by its first two
moments.
• The processes we will discuss are assumed to be Gaussian unless
mentioned otherwise.
• Like other areas in statistics, most time series results are established
for Gaussian processes.
week 2 13
White Noise Processes
• A process {Xt} is called white noise process if it is a sequence of
uncorrelated random variables from a fixed distribution with
constant mean μ (usually assume to be 0) and constant variance σ2
.
• A white noise process is stationary with autocovariance and
autocorrelation functions given by ….
• A white noise process is Gaussian if its joint distribution is normal.
week 2 14
Estimation of the mean
• Given a single realization {xt} of a stationary process {Xt}, a natural
estimator of the mean is the sample mean
which is the time average of n observations.
• It can be shown that the sample mean is unbiased and consistent
estimator for μ.
  

t
X
E



n
t
t
x
n
x
1
1
week 2 15
Sample Autocovariance Function
• Given a single realization {xt} of a stationary process {Xt}, the
sample autocovariance function given by
is an estimate of the autocivariance function.
    



 


k
n
t
k
t
t x
x
x
x
n
k
1
1
̂
week 2 16
Sample Autocorrelation Function
• For a given time series {xt}, the sample autocorrelation function is
given by
• The sample autocorrelation function is non-negative definite.
• The sample autocovariance and autocorrelation functions have the
same properties as the autocovariance and autocorrelation function
of the entire process.
 
  
 
 
 
.
0
ˆ
ˆ
ˆ
1
2
1



k
x
x
x
x
x
x
k n
t
t
k
n
t
k
t
t











week 2 17
Example

More Related Content

PPT
Introduction - Time Series Analysis
jaya gobi
 
PDF
Introduction to queueing systems with telecommunication applications
Springer
 
PDF
Introduction to queueing systems with telecommunication applications
Springer
 
PPTX
Dsp presentation
ILA SHARMA
 
PPTX
time series for agricultural research
ygarde
 
PDF
Lecture_Random Process_Part-1_July-Dec 2023.pdf
msc9219
 
PDF
Time Series Analysis with R
ARCHIT GUPTA
 
Introduction - Time Series Analysis
jaya gobi
 
Introduction to queueing systems with telecommunication applications
Springer
 
Introduction to queueing systems with telecommunication applications
Springer
 
Dsp presentation
ILA SHARMA
 
time series for agricultural research
ygarde
 
Lecture_Random Process_Part-1_July-Dec 2023.pdf
msc9219
 
Time Series Analysis with R
ARCHIT GUPTA
 

Similar to stochastic processes and properties -2.ppt (20)

PDF
time series analysis notes heheheheh.pdf
rv13iscool
 
PPT
probability Chapter Four-Random_Processes.ppt
getahunshanko2024
 
PPTX
Av 738- Adaptive Filtering - Background Material
Dr. Bilal Siddiqui, C.Eng., MIMechE, FRAeS
 
PPTX
Random vibrations
Koteswara Rao Unnam
 
PDF
Unit 1 PPT .d ocx.pdf ppt presentation
jayaage2716
 
PPT
Timeseries_presentation.ppt
M4zein
 
PPT
Timeseries_presentation.ppt
ssuserdca880
 
PDF
Stochastic Processes - part 3
HAmindavarLectures
 
PDF
Digital Communication - Stochastic Process
International Institute of Information Technology (I²IT)
 
PDF
Introduction to Statistical Methods for Financial Models 1st Severini Solutio...
kaelagerik3z
 
PDF
Time series for yotube_3_data anlysis.pdf
asmaamahmoudsaeed
 
PPT
Ch6 slides
fentaw leykun
 
PDF
MFx_Module_3_Properties_of_Time_Series.pdf
tilfani
 
PDF
ch9.pdf
KavS14
 
PPT
lecture about different methods and explainanation.ppt
sadafshahbaz7777
 
PPT
Scholastic process and explaination lectr14.ppt
sadafshahbaz7777
 
PDF
Introduction to Statistical Methods for Financial Models 1st Severini Solutio...
morinjnicau
 
PDF
stochastic-processes-1.pdf
oricho
 
PDF
Time Series for FRAM-Second_Sem_2021-22 (1).pdf
rembeauty4
 
time series analysis notes heheheheh.pdf
rv13iscool
 
probability Chapter Four-Random_Processes.ppt
getahunshanko2024
 
Av 738- Adaptive Filtering - Background Material
Dr. Bilal Siddiqui, C.Eng., MIMechE, FRAeS
 
Random vibrations
Koteswara Rao Unnam
 
Unit 1 PPT .d ocx.pdf ppt presentation
jayaage2716
 
Timeseries_presentation.ppt
M4zein
 
Timeseries_presentation.ppt
ssuserdca880
 
Stochastic Processes - part 3
HAmindavarLectures
 
Digital Communication - Stochastic Process
International Institute of Information Technology (I²IT)
 
Introduction to Statistical Methods for Financial Models 1st Severini Solutio...
kaelagerik3z
 
Time series for yotube_3_data anlysis.pdf
asmaamahmoudsaeed
 
Ch6 slides
fentaw leykun
 
MFx_Module_3_Properties_of_Time_Series.pdf
tilfani
 
ch9.pdf
KavS14
 
lecture about different methods and explainanation.ppt
sadafshahbaz7777
 
Scholastic process and explaination lectr14.ppt
sadafshahbaz7777
 
Introduction to Statistical Methods for Financial Models 1st Severini Solutio...
morinjnicau
 
stochastic-processes-1.pdf
oricho
 
Time Series for FRAM-Second_Sem_2021-22 (1).pdf
rembeauty4
 
Ad

Recently uploaded (20)

PDF
FLEX-LNG-Company-Presentation-Nov-2017.pdf
jbloggzs
 
PPTX
Chapter_Seven_Construction_Reliability_Elective_III_Msc CM
SubashKumarBhattarai
 
PDF
EVS+PRESENTATIONS EVS+PRESENTATIONS like
saiyedaqib429
 
PPTX
Color Model in Textile ( RGB, CMYK).pptx
auladhossain191
 
PPTX
22PCOAM21 Session 1 Data Management.pptx
Guru Nanak Technical Institutions
 
PDF
Introduction to Data Science: data science process
ShivarkarSandip
 
PDF
Top 10 read articles In Managing Information Technology.pdf
IJMIT JOURNAL
 
PPTX
EE3303-EM-I 25.7.25 electrical machines.pptx
Nagen87
 
PPT
Ppt for engineering students application on field effect
lakshmi.ec
 
PDF
Software Testing Tools - names and explanation
shruti533256
 
PPTX
Information Retrieval and Extraction - Module 7
premSankar19
 
PDF
flutter Launcher Icons, Splash Screens & Fonts
Ahmed Mohamed
 
PPTX
database slide on modern techniques for optimizing database queries.pptx
aky52024
 
PDF
Principles of Food Science and Nutritions
Dr. Yogesh Kumar Kosariya
 
PDF
Queuing formulas to evaluate throughputs and servers
gptshubham
 
PPTX
MSME 4.0 Template idea hackathon pdf to understand
alaudeenaarish
 
PDF
top-5-use-cases-for-splunk-security-analytics.pdf
yaghutialireza
 
PDF
dse_final_merit_2025_26 gtgfffffcjjjuuyy
rushabhjain127
 
PDF
67243-Cooling and Heating & Calculation.pdf
DHAKA POLYTECHNIC
 
PPTX
22PCOAM21 Session 2 Understanding Data Source.pptx
Guru Nanak Technical Institutions
 
FLEX-LNG-Company-Presentation-Nov-2017.pdf
jbloggzs
 
Chapter_Seven_Construction_Reliability_Elective_III_Msc CM
SubashKumarBhattarai
 
EVS+PRESENTATIONS EVS+PRESENTATIONS like
saiyedaqib429
 
Color Model in Textile ( RGB, CMYK).pptx
auladhossain191
 
22PCOAM21 Session 1 Data Management.pptx
Guru Nanak Technical Institutions
 
Introduction to Data Science: data science process
ShivarkarSandip
 
Top 10 read articles In Managing Information Technology.pdf
IJMIT JOURNAL
 
EE3303-EM-I 25.7.25 electrical machines.pptx
Nagen87
 
Ppt for engineering students application on field effect
lakshmi.ec
 
Software Testing Tools - names and explanation
shruti533256
 
Information Retrieval and Extraction - Module 7
premSankar19
 
flutter Launcher Icons, Splash Screens & Fonts
Ahmed Mohamed
 
database slide on modern techniques for optimizing database queries.pptx
aky52024
 
Principles of Food Science and Nutritions
Dr. Yogesh Kumar Kosariya
 
Queuing formulas to evaluate throughputs and servers
gptshubham
 
MSME 4.0 Template idea hackathon pdf to understand
alaudeenaarish
 
top-5-use-cases-for-splunk-security-analytics.pdf
yaghutialireza
 
dse_final_merit_2025_26 gtgfffffcjjjuuyy
rushabhjain127
 
67243-Cooling and Heating & Calculation.pdf
DHAKA POLYTECHNIC
 
22PCOAM21 Session 2 Understanding Data Source.pptx
Guru Nanak Technical Institutions
 
Ad

stochastic processes and properties -2.ppt

  • 1. week 2 1 Stochastic Process - Introduction • Stochastic processes are processes that proceed randomly in time. • Rather than consider fixed random variables X, Y, etc. or even sequences of i.i.d random variables, we consider sequences X0, X1, X2, …. Where Xt represent some random quantity at time t. • In general, the value Xt might depend on the quantity Xt-1 at time t-1, or even the value Xs for other times s < t. • Example: simple random walk .
  • 2. week 2 2 STOCHASTIC PROCESSES • DR.M.THIAGARAJAN ASSOCIATE PROFESSOR OF MATHEMATICS • ST.JOSEPH’S COLLEGE • TRICHY-2
  • 3. week 2 3 Stochastic Process - Definition • A stochastic process is a family of time indexed random variables Xt where t belongs to an index set. Formal notation, where I is an index set that is a subset of R. • Examples of index sets: 1) I = (-∞, ∞) or I = [0, ∞]. In this case Xt is a continuous time stochastic process. 2) I = {0, ±1, ±2, ….} or I = {0, 1, 2, …}. In this case Xt is a discrete time stochastic process. • We use uppercase letter {Xt } to describe the process. A time series, {xt } is a realization or sample function from a certain process. • We use information from a time series to estimate parameters and properties of process {Xt }.   I t Xt  :
  • 4. week 2 4 Probability Distribution of a Process • For any stochastic process with index set I, its probability distribution function is uniquely determined by its finite dimensional distributions. • The k dimensional distribution function of a process is defined by for any and any real numbers x1, …, xk . • The distribution function tells us everything we need to know about the process {Xt }.     k t t k X X x X x X P x x F k k t t    ,..., ,..., 1 1 ,..., 1 1 I t t k  ,..., 1
  • 5. week 2 5 Moments of Stochastic Process • We can describe a stochastic process via its moments, i.e., We often use the first two moments. • The mean function of the process is • The variance function of the process is • The covariance function between Xt , Xs is • The correlation function between Xt , Xs is • These moments are often function of time.       etc. , , 2 s t t t X X E X E X E    . t t X E     . 2 t t X Var          s s t t s t X X E X X      , Cov     2 2 , Cov , s t s t s t X X X X    
  • 6. week 2 6 Stationary Processes • A process is said to be strictly stationary if has the same joint distribution as . That is, if • If {Xt } is a strictly stationary process and then, the mean function is a constant and the variance function is also a constant. • Moreover, for a strictly stationary process with first two moments finite, the covariance function, and the correlation function depend only on the time difference s. • A trivial example of a strictly stationary process is a sequence of i.i.d random variables.   k t t X X ,..., 1       k t t X X ,..., 1     k X X k X X x x F x x F k t t k t t ,..., ,..., 1 ,..., 1 ,..., 1 1          2 t X E
  • 7. week 2 7 Weak Stationarity • Strict stationarity is too strong of a condition in practice. It is often difficult assumption to assess based on an observed time series x1,…,xk. • In time series analysis we often use a weaker sense of stationarity in terms of the moments of the process. • A process is said to be nth-order weakly stationary if all its joint moments up to order n exists and are time invariant, i.e., independent of time origin. • For example, a second-order weakly stationary process will have constant mean and variance, with the covariance and the correlation being functions of the time difference along. • A strictly stationary process with the first two moments finite is also a second-ordered weakly stationary. But a strictly stationary process may not have finite moments and therefore may not be weakly stationary.
  • 8. week 2 8 The Autocovariance and Autocorrelation Functions • For a stationary process {Xt }, with constant mean μ and constant variance σ2. The covariance between Xt and Xt+s is • The correlation between Xt and Xt+s is Where • As functions of s, γ(s) is called the autocovariance function and ρ(s) is called the autocorrelation function (ATF). They represent the covariance and correlation between Xt and Xt+s from the same process, separated only by s time lags.                  s t s t t X X E X X s , cov             0 var var , cov    s X X X X s s t s t        . 0 var var    s t t X X
  • 9. week 2 9 Properties of γ(s) and ρ(s) • For a stationary process, the autocovariance function γ(s) and the autocorrelation function ρ(s) have the following properties:   .  .  The autocovariance function γ(s) and the autocorrelation function ρ(s) are positive semidefinite in the sense that for any real numbers       . 1 0 ; var 0     t X   1 1    s          s s s         s and   and 0 1 1      n i n j j i j i      0 1 1      n i n j j i j i    . ,..., , 2 1 n   
  • 10. week 2 10 Correlogram • A correlogram is a plot of the autocorrelation function ρ(s) versus the lag s where s = 0,1, …. • Example…
  • 11. week 2 11 Partial Autocorrelation Function • Often we want to investigate the dependency / association between Xt and Xt+k adjusting for their dependency on Xt+1, Xt+2,…, Xt+k-1. • The conditional correlation Corr(Xt , Xt+k | Xt+1, Xt+2,…, Xt+k-1) is usually referred to as the partial correlation in time series analysis. • Partial autocorrelation is usually useful for identifying autoregressive models.
  • 12. week 2 12 Gaussian process • A stochastic process is said to be a normal or Gaussian process if its joint probability distribution is normal. • A Gaussian process is strictly and weakly stationary because the normal distribution is uniquely characterized by its first two moments. • The processes we will discuss are assumed to be Gaussian unless mentioned otherwise. • Like other areas in statistics, most time series results are established for Gaussian processes.
  • 13. week 2 13 White Noise Processes • A process {Xt} is called white noise process if it is a sequence of uncorrelated random variables from a fixed distribution with constant mean μ (usually assume to be 0) and constant variance σ2 . • A white noise process is stationary with autocovariance and autocorrelation functions given by …. • A white noise process is Gaussian if its joint distribution is normal.
  • 14. week 2 14 Estimation of the mean • Given a single realization {xt} of a stationary process {Xt}, a natural estimator of the mean is the sample mean which is the time average of n observations. • It can be shown that the sample mean is unbiased and consistent estimator for μ.     t X E    n t t x n x 1 1
  • 15. week 2 15 Sample Autocovariance Function • Given a single realization {xt} of a stationary process {Xt}, the sample autocovariance function given by is an estimate of the autocivariance function.             k n t k t t x x x x n k 1 1 ̂
  • 16. week 2 16 Sample Autocorrelation Function • For a given time series {xt}, the sample autocorrelation function is given by • The sample autocorrelation function is non-negative definite. • The sample autocovariance and autocorrelation functions have the same properties as the autocovariance and autocorrelation function of the entire process.            . 0 ˆ ˆ ˆ 1 2 1    k x x x x x x k n t t k n t k t t           