šŸ•·ļø Crawler Inspector

URL Lookup

Direct Parameter Lookup

Raw Queries and Responses

1. Shard Calculation

Query:
Response:
Calculated Shard: 115 (from laksa017)

2. Crawled Status Check

Query:
Response:

3. Robots.txt Check

Query:
Response:

4. Spam/Ban Check

Query:
Response:

5. Seen Status Check

ā„¹ļø Skipped - page is already crawled

šŸ“„
INDEXABLE
āœ…
CRAWLED
1 day ago
šŸ¤–
ROBOTS ALLOWED

Page Info Filters

FilterStatusConditionDetails
HTTP statusPASSdownload_http_code = 200HTTP 200
Age cutoffPASSdownload_stamp > now() - 6 MONTH0.1 months ago
History dropPASSisNull(history_drop_reason)No drop reason
Spam/banPASSfh_dont_index != 1 AND ml_spam_score = 0ml_spam_score=0
CanonicalPASSmeta_canonical IS NULL OR = '' OR = src_unparsedNot set

Page Details

PropertyValue
URLhttps://www.probabilitycourse.com/chapter11/11_2_4_classification_of_states.php
Last Crawled2026-04-13 16:42:21 (1 day ago)
First Indexed2018-12-15 13:22:05 (7 years ago)
HTTP Status Code200
Meta TitleClassification of States
Meta Descriptionnull
Meta Canonicalnull
Boilerpipe Text
To better understand Markov chains, we need to introduce some definitions. The first definition concerns the accessibility of states from each other: If it is possible to go from state i to state j , we say that state j is accessible from state i . In particular, we can provide the following definitions. We say that state j is accessible from state i , written as i → j , if p i j ( n ) > 0 for some n . We assume every state is accessible from itself since p i i ( 0 ) = 1 . Two states i and j are said to communicate , written as i ↔ j , if they are accessible from each other. In other words, i ↔ j Ā meansĀ  i → j Ā andĀ  j → i . Communication is an equivalence relation. That means that āˆ’ every state communicates with itself, i ↔ i ; āˆ’ if i ↔ j , then j ↔ i ; āˆ’ if i ↔ j and j ↔ k , then i ↔ k . Therefore, the states of a Markov chain can be partitioned into communicating classes such that only members of the same class communicate with each other. That is, two states i and j belong to the same class if and only if i ↔ j . Example Consider the Markov chain shown in Figure 11.9. It is assumed that when there is an arrow from state i to state j , then p i j > 0 . Find the equivalence classes for this Markov chain. Figure 11.9 - A state transition diagram. Solution There are four communicating classes in this Markov chain. Looking at Figure 11.10, we notice that states 1 and 2 communicate with each other, but they do not communicate with any other nodes in the graph. Similarly, nodes 3 and 4 communicate with each other, but they do not communicate with any other nodes in the graph. State 5 does not communicate with any other states, so it by itself is a class. Finally, states 6 , 7 , and 8 construct another class. Thus, here are the classes: Class 1 = { stateĀ  1 , stateĀ  2 } , Class 2 = { stateĀ  3 , stateĀ  4 } , Class 3 = { stateĀ  5 } , Class 4 = { stateĀ  6 , stateĀ  7 , stateĀ  8 } . Figure 11.10 - Equivalence classes. A Markov chain is said to be irreducible if it has only one communicating class. As we will see shortly, irreducibility is a desirable property in the sense that it can simplify analysis of the limiting behavior. A Markov chain is said to be irreducible if all states communicate with each other. Looking at Figure 11.10, we notice that there are two kinds of classes. In particular, if at any time the Markov chain enters Class 4 , it will always stay in that class. On the other hand, for other classes this is not true. For example, if X 0 = 1 , then the Markov chain might stay in Class 1 for a while, but at some point, it will leave that class and it will never return to that class again. The states in Class 4 are called recurrent states, while the other states in this chain are called transient . In general, a state is said to be recurrent if, any time that we leave that state, we will return to that state in the future with probability one. On the other hand, if the probability of returning is less than one, the state is called transient. Here, we provide a formal definition: For any state i , we define f i i = P ( X n = i , Ā for someĀ  n ≄ 1 | X 0 = i ) . State i is recurrent if f i i = 1 , and it is transient if f i i < 1 . It is relatively easy to show that if two states are in the same class, either both of them are recurrent, or both of them are transient. Thus, we can extend the above definitions to classes. A class is said to be recurrent if the states in that class are recurrent. If, on the other hand, the states are transient, the class is called transient. In general, a Markov chain might consist of several transient classes as well as several recurrent classes. Consider a Markov chain and assume X 0 = i . If i is a recurrent state, then the chain will return to state i any time it leaves that state. Therefore, the chain will visit state i an infinite number of times. On the other hand, if i is a transient state, the chain will return to state i with probability f i i < 1 . Thus, in that case, the total number of visits to state i will be a Geometric random variable with parameter 1 āˆ’ f i i . Consider a discrete-time Markov chain. Let V be the total number of visits to state i . If i is a recurrent state, then P ( V = āˆž | X 0 = i ) = 1. If i is a transient state, then V | X 0 = i ∼ G e o m e t r i c ( 1 āˆ’ f i i ) . Example Show that in a finite Markov chain, there is at least one recurrent class. Solution Consider a finite Markov chain with r states, S = { 1 , 2 , ⋯ , r } . Suppose that all states are transient. Then, starting from time 0 , the chain might visit state 1 several times, but at some point the chain will leave state 1 and will never return to it. That is, there exists an integer M 1 > 0 such that X n ≠ 1 , for all n ≄ M 1 . Similarly, there exists an integer M 2 > 0 such that X n ≠ 2 , for all n ≄ M 2 , and so on. Now, if you choose n ≄ max { M 1 , M 2 , ⋯ , M r } , then X n cannot be equal to any of the states 1 , 2 , ⋯ , r . This is a contradiction, so we conclude that there must be at least one recurrent state, which means that there must be at least one recurrent class. Periodicity: Consider the Markov chain shown in Figure 11.11. There is a periodic pattern in this chain. Starting from state 0 , we only return to 0 at times n = 3 , 6 , ⋯ . In other words, p 00 ( n ) = 0 , if n is not divisible by 3 . Such a state is called a periodic state with period d ( 0 ) = 3 . Figure 11.11 - A state transition diagram. The period of a state i is the largest integer d satisfying the following property: p i i ( n ) = 0 , whenever n is not divisible by d . The period of i is shown by d ( i ) . If p i i ( n ) = 0 , for all n > 0 , then we let d ( i ) = āˆž . āˆ’ If d ( i ) > 1 , we say that state i is periodic . āˆ’ If d ( i ) = 1 , we say that state i is aperiodic . You can show that all states in the same communicating class have the same period. A class is said to be periodic if its states are periodic. Similarly, a class is said to be aperiodic if its states are aperiodic. Finally, a Markov chain is said to be aperiodic if all of its states are aperiodic. If i ↔ j , then d ( i ) = d ( j ) . Why is periodicity important? As we will see shortly, it plays a role when we discuss limiting distributions. It turns out that in a typical problem, we are given an irreducible Markov chain, and we need to check if it is aperiodic. How do we check that a Markov chain is aperiodic? Here is a useful method. Remember that two numbers m and l are said to be co-prime if their greatest common divisor (gcd) is 1 , i.e., gcd ( l , m ) = 1 . Now, suppose that we can find two co-prime numbers l and m such that p i i ( l ) > 0 and p i i ( m ) > 0 . That is, we can go from state i to itself in l steps, and also in m steps. Then, we can conclude state i is aperiodic. If we have an irreducible Markov chain, this means that the chain is aperiodic. Since the number 1 is co-prime to every integer, any state with a self-transition is aperiodic. Consider a finite irreducible Markov chain X n : If there is a self-transition in the chain ( p i i > 0 for some i ), then the chain is aperiodic. Suppose that you can go from state i to state i in l steps, i.e., p i i ( l ) > 0 . Also suppose that p i i ( m ) > 0 . If gcd ( l , m ) = 1 , then state i is aperiodic. The chain is aperiodic if and only if there exists a positive integer n such that all elements of the matrix P n are strictly positive, i.e., p i j ( n ) > 0 , Ā for allĀ  i , j ∈ S . Example Consider the Markov chain in Example 11.6 . Is Class 1 = { stateĀ  1 , stateĀ  2 } aperiodic? Is Class 2 = { stateĀ  3 , stateĀ  4 } aperiodic? Is Class 4 = { stateĀ  6 , stateĀ  7 , stateĀ  8 } aperiodic? Solution Class 1 = { stateĀ  1 , stateĀ  2 } is aperiodic since it has a self-transition, p 22 > 0 . Class 2 = { stateĀ  3 , stateĀ  4 } is periodic with period 2 . Class 4 = { stateĀ  6 , stateĀ  7 , stateĀ  8 } is aperiodic . For example, note that we can go from state 6 to state 6 in two steps ( 6 āˆ’ 7 āˆ’ 6 ) and in three steps ( 6 āˆ’ 7 āˆ’ 8 āˆ’ 6 ). Since gcd ( 2 , 3 ) = 1 , we conclude state 6 and its class are aperiodic. The print version of the book is available on Amazon . Practical uncertainty: Useful Ideas in Decision-Making, Risk, Randomness, & AI
Markdown
- [HOME](https://www.probabilitycourse.com/) - [VIDEOS](https://www.probabilitycourse.com/videos/videos.php) - [CALCULATOR](https://www.probabilitycourse.com/calculator/calculator.php) - [COMMENTS](https://www.probabilitycourse.com/comments.php) - [COURSES](https://www.probabilitycourse.com/courses.php) - [FOR INSTRUCTOR](https://www.probabilitycourse.com/for_instructors.php) - [LOG IN](https://www.probabilitycourse.com/Login/mobile_login.php) [![](https://www.probabilitycourse.com/images/Template/logo_probability_website.png)](https://www.probabilitycourse.com/) Chapters Menu - [HOME](https://www.probabilitycourse.com/) - [VIDEOS](https://www.probabilitycourse.com/videos/videos.php) - [CALCULATOR](https://www.probabilitycourse.com/calculator/calculator.php) - [COMMENTS](https://www.probabilitycourse.com/comments.php) - [COURSES](https://www.probabilitycourse.com/courses.php) - [FOR INSTRUCTORS](https://www.probabilitycourse.com/for_instructors.php) - [Sign In]() [Forgot password?](https://www.probabilitycourse.com/Login/forgot_password.php) [←]() [previous](https://www.probabilitycourse.com/chapter11/chapter11/11_2_3_probability_distributions.php) [next](https://www.probabilitycourse.com/chapter11/chapter11/11_2_5_using_the_law_of_total_probability_with_recursion.php) [→]() *** ## 11\.2.4 Classification of States To better understand Markov chains, we need to introduce some definitions. The first definition concerns the accessibility of states from each other: If it is possible to go from state i i to state j j , we say that state j j is *accessible* from state i i . In particular, we can provide the following definitions. We say that state j j is **accessible** from state i i , written as i→j i → j , if p(n)ij\>0 p i j ( n ) \> 0 for some n n . We assume every state is accessible from itself since p(0)ii\=1 p i i ( 0 ) \= 1 . Two states i i and j j are said to **communicate**, written as i↔j i ↔ j , if they are **accessible** from each other. In other words, i↔j means i→j and j→i. i ↔ j means i → j and j → i . Communication is an *equivalence* relation. That means that āˆ’ āˆ’ every state communicates with itself, i↔i i ↔ i ; āˆ’ āˆ’ if i↔j i ↔ j , then j↔i j ↔ i ; āˆ’ āˆ’ if i↔j i ↔ j and j↔k j ↔ k , then i↔k i ↔ k . Therefore, the states of a Markov chain can be partitioned into communicating *classes* such that only members of the same class communicate with each other. That is, two states i i and j j belong to the same class if and only if i↔j i ↔ j . *** Example Consider the Markov chain shown in Figure 11.9. It is assumed that when there is an arrow from state i i to state j j , then pij\>0 p i j \> 0 . Find the equivalence classes for this Markov chain. ![MC-diagram-3](https://www.probabilitycourse.com/chapter11/images/chapter11/MC-diagram-3.png) Figure 11.9 - A state transition diagram. - [**Solution**]() - There are four communicating classes in this Markov chain. Looking at Figure 11.10, we notice that states 1 1 and 2 2 communicate with each other, but they do not communicate with any other nodes in the graph. Similarly, nodes 3 3 and 4 4 communicate with each other, but they do not communicate with any other nodes in the graph. State 5 5 does not communicate with any other states, so it by itself is a class. Finally, states 6 6 , 7 7 , and 8 8 construct another class. Thus, here are the classes: Class 1\={state 1,state 2},Class 2\={state 3,state 4},Class 3\={state 5},Class 4\={state 6,state 7,state 8}. Class 1 \= { state 1 , state 2 } , Class 2 \= { state 3 , state 4 } , Class 3 \= { state 5 } , Class 4 \= { state 6 , state 7 , state 8 } . ![MC-diagram-4](https://www.probabilitycourse.com/chapter11/images/chapter11/MC-diagram-4.png) Figure 11.10 - Equivalence classes. *** A Markov chain is said to be *irreducible* if it has only one communicating class. As we will see shortly, irreducibility is a desirable property in the sense that it can simplify analysis of the limiting behavior. A Markov chain is said to be **irreducible** if all states communicate with each other. Looking at Figure 11.10, we notice that there are two kinds of classes. In particular, if at any time the Markov chain enters Class 4 4, it will always stay in that class. On the other hand, for other classes this is not true. For example, if X0\=1 X 0 \= 1, then the Markov chain might stay in Class 1 1 for a while, but at some point, it will leave that class and it will never return to that class again. The states in Class 4 4 are called *recurrent* states, while the other states in this chain are called *transient*. In general, a state is said to be recurrent if, any time that we leave that state, we will return to that state in the future with probability one. On the other hand, if the probability of returning is less than one, the state is called transient. Here, we provide a formal definition: For any state i i , we define fii\=P(Xn\=i, for some n≄1\|X0\=i). f i i \= P ( X n \= i , for some n ≄ 1 \| X 0 \= i ) . State i i is **recurrent** if fii\=1 f i i \= 1 , and it is **transient** if fii\<1 f i i \< 1 . It is relatively easy to show that if two states are in the same class, either both of them are recurrent, or both of them are transient. Thus, we can extend the above definitions to classes. A class is said to be recurrent if the states in that class are recurrent. If, on the other hand, the states are transient, the class is called transient. In general, a Markov chain might consist of several transient classes as well as several recurrent classes. Consider a Markov chain and assume X0\=i X 0 \= i. If i i is a recurrent state, then the chain will return to state i i any time it leaves that state. Therefore, the chain will visit state i i an infinite number of times. On the other hand, if i i is a transient state, the chain will return to state i i with probability fii\<1 f i i \< 1. Thus, in that case, the total number of visits to state i i will be a Geometric random variable with parameter 1āˆ’fii 1 āˆ’ f i i. Consider a discrete-time Markov chain. Let V V be the total number of visits to state i i . 1. If i i is a recurrent state, then P(V\=āˆž\|X0\=i)\=1\. P ( V \= āˆž \| X 0 \= i ) \= 1\. 2. If i i is a transient state, then V\|X0\=i∼Geometric(1āˆ’fii). V \| X 0 \= i ∼ G e o m e t r i c ( 1 āˆ’ f i i ) . *** Example Show that in a finite Markov chain, there is at least one recurrent class. - [**Solution**]() - Consider a finite Markov chain with r r states, S\={1,2,⋯,r} S \= { 1 , 2 , ⋯ , r } . Suppose that all states are transient. Then, starting from time 0 0 , the chain might visit state 1 1 several times, but at some point the chain will leave state 1 1 and will never return to it. That is, there exists an integer M1\>0 M 1 \> 0 such that Xn≠1 X n ≠ 1 , for all n≄M1 n ≄ M 1 . Similarly, there exists an integer M2\>0 M 2 \> 0 such that Xn≠2 X n ≠ 2 , for all n≄M2 n ≄ M 2 , and so on. Now, if you choose n≄max{M1,M2,⋯,Mr}, n ≄ max { M 1 , M 2 , ⋯ , M r } , then Xn X n cannot be equal to any of the states 1,2,⋯,r 1 , 2 , ⋯ , r . This is a contradiction, so we conclude that there must be at least one recurrent state, which means that there must be at least one recurrent class. *** ### Periodicity: Consider the Markov chain shown in Figure 11.11. There is a periodic pattern in this chain. Starting from state 0 0 , we only return to 0 0 at times n\=3,6,⋯ n \= 3 , 6 , ⋯ . In other words, p(n)00\=0 p 00 ( n ) \= 0 , if n n is not divisible by 3 3 . Such a state is called a *periodic* state with period d(0)\=3 d ( 0 ) \= 3 . ![MC-diagram-periodic](https://www.probabilitycourse.com/chapter11/images/chapter11/MC-diagram-periodic.png) Figure 11.11 - A state transition diagram. The **period** of a state i i is the largest integer d d satisfying the following property: p(n)ii\=0 p i i ( n ) \= 0 , whenever n n is not divisible by d d . The period of i i is shown by d(i) d ( i ) . If p(n)ii\=0 p i i ( n ) \= 0 , for all n\>0 n \> 0 , then we let d(i)\=āˆž d ( i ) \= āˆž . āˆ’ āˆ’ If d(i)\>1 d ( i ) \> 1 , we say that state i i is **periodic**. āˆ’ āˆ’ If d(i)\=1 d ( i ) \= 1 , we say that state i i is **aperiodic**. You can show that all states in the same communicating class have the same period. A class is said to be periodic if its states are periodic. Similarly, a class is said to be aperiodic if its states are aperiodic. Finally, a Markov chain is said to be aperiodic if all of its states are aperiodic. If i↔j i ↔ j , then d(i)\=d(j) d ( i ) \= d ( j ) . Why is periodicity important? As we will see shortly, it plays a role when we discuss limiting distributions. It turns out that in a typical problem, we are given an irreducible Markov chain, and we need to check if it is aperiodic. How do we check that a Markov chain is aperiodic? Here is a useful method. Remember that two numbers m m and l l are said to be *co-prime* if their greatest common divisor (gcd) is 1 1, i.e., gcd(l,m)\=1 gcd ( l , m ) \= 1. Now, suppose that we can find two co-prime numbers l l and m m such that p(l)ii\>0 p i i ( l ) \> 0 and p(m)ii\>0 p i i ( m ) \> 0. That is, we can go from state i i to itself in l l steps, and also in m m steps. Then, we can conclude state i i is aperiodic. If we have an irreducible Markov chain, this means that the chain is aperiodic. Since the number 1 1 is co-prime to every integer, any state with a self-transition is aperiodic. Consider a finite irreducible Markov chain Xn X n : 1. If there is a self-transition in the chain ( pii\>0 p i i \> 0 for some i i ), then the chain is aperiodic. 2. Suppose that you can go from state i i to state i i in l l steps, i.e., p(l)ii\>0 p i i ( l ) \> 0 . Also suppose that p(m)ii\>0 p i i ( m ) \> 0 . If gcd(l,m)\=1 gcd ( l , m ) \= 1 , then state i i is aperiodic. 3. The chain is aperiodic if and only if there exists a positive integer n n such that all elements of the matrix Pn P n are strictly positive, i.e., *** Example Consider the Markov chain in [Example 11.6](https://www.probabilitycourse.com/chapter11/chapter11/11_2_4_classification_of_states.php#example11_6). 1. Is Class 1\={state 1,state 2} Class 1 \= { state 1 , state 2 } aperiodic? 2. Is Class 2\={state 3,state 4} Class 2 \= { state 3 , state 4 } aperiodic? 3. Is Class 4\={state 6,state 7,state 8} Class 4 \= { state 6 , state 7 , state 8 } aperiodic? - [**Solution**]() - 1. Class 1\={state 1,state 2} Class 1 \= { state 1 , state 2 } is aperiodic since it has a self-transition, p22\>0 p 22 \> 0 . 2. Class 2\={state 3,state 4} Class 2 \= { state 3 , state 4 } is periodic with period 2 2 . 3. Class 4\={state 6,state 7,state 8} Class 4 \= { state 6 , state 7 , state 8 } is aperiodic . For example, note that we can go from state 6 6 to state 6 6 in two steps ( 6āˆ’7āˆ’6 6 āˆ’ 7 āˆ’ 6 ) and in three steps ( 6āˆ’7āˆ’8āˆ’6 6 āˆ’ 7 āˆ’ 8 āˆ’ 6 ). Since gcd(2,3)\=1 gcd ( 2 , 3 ) \= 1 , we conclude state 6 6 and its class are aperiodic. *** [←]() [previous](https://www.probabilitycourse.com/chapter11/chapter11/11_2_3_probability_distributions.php) [next](https://www.probabilitycourse.com/chapter11/chapter11/11_2_5_using_the_law_of_total_probability_with_recursion.php) [→]() *** | | |---| | The print version of the book is available on [Amazon](https://www.amazon.com/Introduction-Probability-Statistics-Random-Processes/dp/0990637204/ref=sr_1_1?ie=UTF8&qid=1408880878&sr=8-1&keywords=pishro-nik). [![Book Cover](https://www.probabilitycourse.com/images/home/book_cover.jpg)](https://www.amazon.com/Introduction-Probability-Statistics-Random-Processes/dp/0990637204/ref=sr_1_1?ie=UTF8&qid=1408880878&sr=8-1&keywords=pishro-nik) | | **Practical uncertainty:** *Useful Ideas in Decision-Making, Risk, Randomness, & AI* [![ractical Uncertaintly Cover](https://www.probabilitycourse.com/images/home/practical_uncertaintly.PNG)](https://www.amazon.com/dp/B0CH2BHRVH/ref=tmm_pap_swatch_0?_encoding=UTF8&qid=1693837152&sr=8-1) | Open Menu - [0 Preface](https://www.probabilitycourse.com/preface.php) - [1 Basic Concepts]() - [1\.0 Introduction](https://www.probabilitycourse.com/chapter1/1_0_0_introduction.php) - [1\.1 Introduction]() - [1\.1.0 What Is Probability?](https://www.probabilitycourse.com/chapter1/1_1_0_what_is_probability.php) - [1\.1.1 Example](https://www.probabilitycourse.com/chapter1/1_1_1_example.php) - [1\.2 Review of Set Theory]() - [1\.2.0 Review](https://www.probabilitycourse.com/chapter1/1_2_0_review_set_theory.php) - [1\.2.1 Venn Diagrams](https://www.probabilitycourse.com/chapter1/1_2_1_venn.php) - [1\.2.2 Set Operations](https://www.probabilitycourse.com/chapter1/1_2_2_set_operations.php) - [1\.2.3 Cardinality](https://www.probabilitycourse.com/chapter1/1_2_3_cardinality.php) - [1\.2.4 Functions](https://www.probabilitycourse.com/chapter1/1_2_4_functions.php) - [1\.2.5 Solved Problems](https://www.probabilitycourse.com/chapter1/1_2_5_solved1.php) - [1\.3 Random Experiments and Probabilities]() - [1\.3.1 Random Experiments](https://www.probabilitycourse.com/chapter1/1_3_1_random_experiments.php) - [1\.3.2 Probability](https://www.probabilitycourse.com/chapter1/1_3_2_probability.php) - [1\.3.3 Finding Probabilities](https://www.probabilitycourse.com/chapter1/1_3_3_finding_probabilities.php) - [1\.3.4 Discrete Models](https://www.probabilitycourse.com/chapter1/1_3_4_discrete_models.php) - [1\.3.5 Continuous Models](https://www.probabilitycourse.com/chapter1/1_3_5_continuous_models.php) - [1\.3.6 Solved Problems](https://www.probabilitycourse.com/chapter1/1_3_6_solved2.php) - [1\.4 Conditional Probability]() - [1\.4.0 Conditional Probability](https://www.probabilitycourse.com/chapter1/1_4_0_conditional_probability.php) - [1\.4.1 Independence](https://www.probabilitycourse.com/chapter1/1_4_1_independence.php) - [1\.4.2 Law of Total Probability](https://www.probabilitycourse.com/chapter1/1_4_2_total_probability.php) - [1\.4.3 Bayes' Rule](https://www.probabilitycourse.com/chapter1/1_4_3_bayes_rule.php) - [1\.4.4 Conditional Independence](https://www.probabilitycourse.com/chapter1/1_4_4_conditional_independence.php) - [1\.4.5 Solved Problems](https://www.probabilitycourse.com/chapter1/1_4_5_solved3.php) - [1\.5 Problems]() - [1\.5.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter1/1_5_0_chapter1_problems.php) - [2 Combinatorics: Counting Methods]() - [2\.1 Combinatorics]() - [2\.1.0 Finding Probabilities with Counting Methods](https://www.probabilitycourse.com/chapter2/2_1_0_counting.php) - [2\.1.1 Ordered with Replacement](https://www.probabilitycourse.com/chapter2/2_1_1_ordered_with_replacement.php) - [2\.1.2 Ordered without Replacement](https://www.probabilitycourse.com/chapter2/2_1_2_ordered_without_replacement.php) - [2\.1.3 Unordered without Replacement](https://www.probabilitycourse.com/chapter2/2_1_3_unordered_without_replacement.php) - [2\.1.4 Unordered with Replacement](https://www.probabilitycourse.com/chapter2/2_1_4_unordered_with_replacement.php) - [2\.1.5 Solved Problems](https://www.probabilitycourse.com/chapter2/2_1_5_solved2_1.php) - [2\.2 Problems]() - [2\.2.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter2/2_3_0_chapter2_problems.php) - [3 Discrete Random Variables]() - [3\.1 Basic Concepts]() - [3\.1.1 Random Variables](https://www.probabilitycourse.com/chapter3/3_1_1_random_variables.php) - [3\.1.2 Discrete Random Variables](https://www.probabilitycourse.com/chapter3/3_1_2_discrete_random_var.php) - [3\.1.3 Probability Mass Function](https://www.probabilitycourse.com/chapter3/3_1_3_pmf.php) - [3\.1.4 Independent Random Variables](https://www.probabilitycourse.com/chapter3/3_1_4_independent_random_var.php) - [3\.1.5 Special Distributions](https://www.probabilitycourse.com/chapter3/3_1_5_special_discrete_distr.php) - [3\.1.6 Solved Problems](https://www.probabilitycourse.com/chapter3/3_1_6_solved3_1.php) - [3\.2 More about Discrete Random Variables]() - [3\.2.1 Cumulative Distribution Function](https://www.probabilitycourse.com/chapter3/3_2_1_cdf.php) - [3\.2.2 Expectation](https://www.probabilitycourse.com/chapter3/3_2_2_expectation.php) - [3\.2.3 Functions of Random Variables](https://www.probabilitycourse.com/chapter3/3_2_3_functions_random_var.php) - [3\.2.4 Variance](https://www.probabilitycourse.com/chapter3/3_2_4_variance.php) - [3\.2.5 Solved Problems](https://www.probabilitycourse.com/chapter3/3_2_5_solved3_2.php) - [3\.3 Problems]() - [3\.3.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter3/3_3_0_chapter3_problems.php) - [4 Continuous and Mixed Random Variables]() - [4\.0 Introduction](https://www.probabilitycourse.com/chapter4/4_0_0_intro.php) - [4\.1 Continuous Random Variables]() - [4\.1.0 Continuous Random Variables and their Distributions](https://www.probabilitycourse.com/chapter4/4_1_0_continuous_random_vars_distributions.php) - [4\.1.1 Probability Density Function](https://www.probabilitycourse.com/chapter4/4_1_1_pdf.php) - [4\.1.2 Expected Value and Variance](https://www.probabilitycourse.com/chapter4/4_1_2_expected_val_variance.php) - [4\.1.3 Functions of Continuous Random Variables](https://www.probabilitycourse.com/chapter4/4_1_3_functions_continuous_var.php) - [4\.1.4 Solved Problems](https://www.probabilitycourse.com/chapter4/4_1_4_solved4_1.php) - [4\.2 Special Distributions]() - [4\.2.1 Uniform Distribution](https://www.probabilitycourse.com/chapter4/4_2_1_uniform.php) - [4\.2.2 Exponential Distribution](https://www.probabilitycourse.com/chapter4/4_2_2_exponential.php) - [4\.2.3 Normal (Gaussian) Distribution](https://www.probabilitycourse.com/chapter4/4_2_3_normal.php) - [4\.2.4 Gamma Distribution](https://www.probabilitycourse.com/chapter4/4_2_4_Gamma_distribution.php) - [4\.2.5 Other Distributions](https://www.probabilitycourse.com/chapter4/4_2_5_other_distr.php) - [4\.2.6 Solved Problems](https://www.probabilitycourse.com/chapter4/4_2_6_solved4_2.php) - [4\.3 Mixed Random Variables]() - [4\.3.1 Mixed Random Variables](https://www.probabilitycourse.com/chapter4/4_3_1_mixed.php) - [4\.3.2 Using the Delta Function](https://www.probabilitycourse.com/chapter4/4_3_2_delta_function.php) - [4\.3.3 Solved Problems](https://www.probabilitycourse.com/chapter4/4_3_3_solved4_3.php) - [4\.4 Problems]() - [4\.4.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter4/4_4_0_chapter4_problems.php) - [5 Joint Distributions]() - [5\.1 Two Discrete Random Variables]() - [5\.1.0 Two Random Variables](https://www.probabilitycourse.com/chapter5/5_1_0_joint_distributions.php) - [5\.1.1 Joint Probability Mass Function (PMF)](https://www.probabilitycourse.com/chapter5/5_1_1_joint_pmf.php) - [5\.1.2 Joint Cumulative Distribution Function (CDF)](https://www.probabilitycourse.com/chapter5/5_1_2_joint_cdf.php) - [5\.1.3 Conditioning and Independence](https://www.probabilitycourse.com/chapter5/5_1_3_conditioning_independence.php) - [5\.1.4 Functions of Two Random Variables](https://www.probabilitycourse.com/chapter5/5_1_4_functions_two_variables.php) - [5\.1.5 Conditional Expectation](https://www.probabilitycourse.com/chapter5/5_1_5_conditional_expectation.php) - [5\.1.6 Solved Problems](https://www.probabilitycourse.com/chapter5/5_1_6_solved_prob.php) - [5\.2 Two Continuous Random Variables]() - [5\.2.0 Two Continuous Random Variables](https://www.probabilitycourse.com/chapter5/5_2_0_continuous_vars.php) - [5\.2.1 Joint Probability Density Function](https://www.probabilitycourse.com/chapter5/5_2_1_joint_pdf.php) - [5\.2.2 Joint Cumulative Distribution Function](https://www.probabilitycourse.com/chapter5/5_2_2_joint_cdf.php) - [5\.2.3 Conditioning and Independence](https://www.probabilitycourse.com/chapter5/5_2_3_conditioning_independence.php) - [5\.2.4 Functions of Two Continuous Random Variables](https://www.probabilitycourse.com/chapter5/5_2_4_functions.php) - [5\.2.5 Solved Problems](https://www.probabilitycourse.com/chapter5/5_2_5_solved_prob.php) - [5\.3 More Topics]() - [5\.3.1 Covariance and Correlation](https://www.probabilitycourse.com/chapter5/5_3_1_covariance_correlation.php) - [5\.3.2 Bivariate Normal Distribution](https://www.probabilitycourse.com/chapter5/5_3_2_bivariate_normal_dist.php) - [5\.3.3 Solved Problems](https://www.probabilitycourse.com/chapter5/5_3_3_solved_probs.php) - [5\.4 Problems]() - [5\.4.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter5/5_4_0_chapter_problems.php) - [6 Multiple Random Variables]() - [6\.0 Introduction](https://www.probabilitycourse.com/chapter6/6_0_0_intro.php) - [6\.1 Methods for More Than Two Random Variables]() - [6\.1.1 Joint Distributions and Independence](https://www.probabilitycourse.com/chapter6/6_1_1_joint_distributions_independence.php) - [6\.1.2 Sums of Random Variables](https://www.probabilitycourse.com/chapter6/6_1_2_sums_random_variables.php) - [6\.1.3 Moment Generating Functions](https://www.probabilitycourse.com/chapter6/6_1_3_moment_functions.php) - [6\.1.4 Characteristic Functions](https://www.probabilitycourse.com/chapter6/6_1_4_characteristic_functions.php) - [6\.1.5 Random Vectors](https://www.probabilitycourse.com/chapter6/6_1_5_random_vectors.php) - [6\.1.6 Solved Problems](https://www.probabilitycourse.com/chapter6/6_1_6_solved_probs.php) - [6\.2 Probability Bounds]() - [6\.2.0 Probability Bounds](https://www.probabilitycourse.com/chapter6/6_2_0_probability_bounds.php) - [6\.2.1 Union Bound and Extension](https://www.probabilitycourse.com/chapter6/6_2_1_union_bound_and_exten.php) - [6\.2.2 Markov Chebyshev Inequalities](https://www.probabilitycourse.com/chapter6/6_2_2_markov_chebyshev_inequalities.php) - [6\.2.3 Chernoff Bounds](https://www.probabilitycourse.com/chapter6/6_2_3_chernoff_bounds.php) - [6\.2.4 Cauchy Schwarz Inequality](https://www.probabilitycourse.com/chapter6/6_2_4_cauchy_schwarz.php) - [6\.2.5 Jensen's Inequality](https://www.probabilitycourse.com/chapter6/6_2_5_jensen's_inequality.php) - [6\.2.6 Solved Problems](https://www.probabilitycourse.com/chapter6/6_2_6_solved6_2.php) - [6\.3 Problems]() - [6\.3.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter6/6_3_0_chapter_problems.php) - [7 Limit Theorems and Convergence of Random Variables]() - [7\.0 Introduction](https://www.probabilitycourse.com/chapter7/7_0_0_intro.php) - [7\.1 Limit Theorems]() - [7\.1.0 Limit Theorems](https://www.probabilitycourse.com/chapter7/7_1_0_limit_theorems.php) - [7\.1.1 Law of Large Numbers](https://www.probabilitycourse.com/chapter7/7_1_1_law_of_large_numbers.php) - [7\.1.2 Central Limit Theorem (CLT)](https://www.probabilitycourse.com/chapter7/7_1_2_central_limit_theorem.php) - [7\.1.3 Solved Problems](https://www.probabilitycourse.com/chapter7/7_1_3_solved_probs.php) - [7\.2 Convergence of Random Variables]() - [7\.2.0 Convergence of Random Variables](https://www.probabilitycourse.com/chapter7/7_2_0_convergence_of_random_variables.php) - [7\.2.1 Convergence of Sequence of Numbers](https://www.probabilitycourse.com/chapter7/7_2_1_convergence_of_a_seq_of_nums.php) - [7\.2.2 Sequence of Random Variables](https://www.probabilitycourse.com/chapter7/7_2_2_sequence_of_random_variables.php) - [7\.2.3 Different Types of Convergence for Sequences of Random Variables](https://www.probabilitycourse.com/chapter7/7_2_3_different_types_of_convergence_for_sequences_of_random_variables.php) - [7\.2.4 Convergence in Distribution](https://www.probabilitycourse.com/chapter7/7_2_4_convergence_in_distribution.php) - [7\.2.5 Convergence in Probability](https://www.probabilitycourse.com/chapter7/7_2_5_convergence_in_probability.php) - [7\.2.6 Convergence in Mean](https://www.probabilitycourse.com/chapter7/7_2_6_convergence_in_mean.php) - [7\.2.7 Almost Sure Convergence](https://www.probabilitycourse.com/chapter7/7_2_7_almost_sure_convergence.php) - [7\.2.8 Solved Problems](https://www.probabilitycourse.com/chapter7/7_2_8_solved_probs.php) - [7\.3 Problems]() - [7\.3.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter7/7_3_0_chapter_problem.php) - [8 Statistical Inference I: Classical Methods]() - [8\.1 Introduction]() - [8\.1.0 Introduction](https://www.probabilitycourse.com/chapter8/8_1_0_intro.php) - [8\.1.1 Random Sampling](https://www.probabilitycourse.com/chapter8/8_1_1_random_sampling.php) - [8\.2 Point Estimation]() - [8\.2.0 Point Estimation](https://www.probabilitycourse.com/chapter8/8_2_0_point_estimation.php) - [8\.2.1 Evaluating Estimators](https://www.probabilitycourse.com/chapter8/8_2_1_evaluating_estimators.php) - [8\.2.2 Point Estimators for Mean and Variance](https://www.probabilitycourse.com/chapter8/8_2_2_point_estimators_for_mean_and_var.php) - [8\.2.3 Maximum Likelihood Estimation (MLE)](https://www.probabilitycourse.com/chapter8/8_2_3_max_likelihood_estimation.php) - [8\.2.4 Asymptotic Properties of MLEs](https://www.probabilitycourse.com/chapter8/8_2_4_asymptotic_probs_of_MLE.php) - [8\.2.5 Solved Problems](https://www.probabilitycourse.com/chapter8/8_2_5_solved_probs.php) - [8\.3 Interval Estimation (Confidence Intervals)]() - [8\.3.0 Interval Estimation (Confidence Intervals)](https://www.probabilitycourse.com/chapter8/8_3_0_interval_estimation.php) - [8\.3.1 The general framework of Interval Estimation](https://www.probabilitycourse.com/chapter8/8_3_1_gen_framework_of_int_estimation.php) - [8\.3.2 Finding Interval Estimators](https://www.probabilitycourse.com/chapter8/8_3_2_finding_interval_estimators.php) - [8\.3.3 Confidence Intervals for Normal Samples](https://www.probabilitycourse.com/chapter8/8_3_3_confidence_intervals_for_norm_samples.php) - [8\.3.4 Solved Problems](https://www.probabilitycourse.com/chapter8/8_3_4_solved_probs.php) - [8\.4 Hypothesis Testing]() - [8\.4.1 Introduction](https://www.probabilitycourse.com/chapter8/8_4_1_intro.php) - [8\.4.2 General Setting and Definitions](https://www.probabilitycourse.com/chapter8/8_4_2_general_setting_definitions.php) - [8\.4.3 Hypothesis Testing for the Mean](https://www.probabilitycourse.com/chapter8/8_4_3_hypothesis_testing_for_mean.php) - [8\.4.4 P-Values](https://www.probabilitycourse.com/chapter8/8_4_4_p_vals.php) - [8\.4.5 Likelihood Ratio Tests](https://www.probabilitycourse.com/chapter8/8_4_5_likelihood_ratio_tests.php) - [8\.4.6 Solved Problems](https://www.probabilitycourse.com/chapter8/8_4_6_solved_probs.php) - [8\.5 Linear Regression]() - [8\.5.0 Linear Regression](https://www.probabilitycourse.com/chapter8/8_5_0_linear_regression.php) - [8\.5.1 Simple Linear Regression Model](https://www.probabilitycourse.com/chapter8/8_5_1_simple_linear_regression_model.php) - [8\.5.2 The First Method for Finding beta](https://www.probabilitycourse.com/chapter8/8_5_2_first_method_for_finding_beta.php) - [8\.5.3 The Method of Least Squares](https://www.probabilitycourse.com/chapter8/8_5_3_the_method_of_least_squares.php) - [8\.5.4 Extensions and Issues](https://www.probabilitycourse.com/chapter8/8_5_4_extensions_and_issues.php) - [8\.5.5 Solved Problems](https://www.probabilitycourse.com/chapter8/8_5_5_solved_probs.php) - [8\.6 Problems]() - [8\.6.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter8/8_6_0_ch_probs.php) - [9 Statistical Inference II: Bayesian Inference]() - [9\.1 Bayesian Inference]() - [9\.1.0 Bayesian Inference](https://www.probabilitycourse.com/chapter9/9_1_0_bayesian_inference.php) - [9\.1.1 Prior and Posterior](https://www.probabilitycourse.com/chapter9/9_1_1_prior_and_posterior.php) - [9\.1.2 Maximum A Posteriori (MAP) Estimation](https://www.probabilitycourse.com/chapter9/9_1_2_MAP_estimation.php) - [9\.1.3 Comparison to ML Estimation](https://www.probabilitycourse.com/chapter9/9_1_3_comparison_to_ML_estimation.php) - [9\.1.4 Conditional Expectation (MMSE)](https://www.probabilitycourse.com/chapter9/9_1_4_conditional_expectation_MMSE.php) - [9\.1.5 Mean Squared Error (MSE)](https://www.probabilitycourse.com/chapter9/9_1_5_mean_squared_error_MSE.php) - [9\.1.6 Linear MMSE Estimation of Random Variables](https://www.probabilitycourse.com/chapter9/9_1_6_linear_MMSE_estimat_of_random_vars.php) - [9\.1.7 Estimation for Random Vectors](https://www.probabilitycourse.com/chapter9/9_1_7_estimation_for_random_vectors.php) - [9\.1.8 Bayesian Hypothesis Testing](https://www.probabilitycourse.com/chapter9/9_1_8_bayesian_hypothesis_testing.php) - [9\.1.9 Bayesian Interval Estimation](https://www.probabilitycourse.com/chapter9/9_1_9_bayesian_interval_estimation.php) - [9\.1.10 Solved Problems](https://www.probabilitycourse.com/chapter9/9_1_10_solved_probs.php) - [9\.2 Problems]() - [9\.2.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter9/9_2_0_ch_probs.php) - [10 Introduction to Random Processes]() - [10\.1 Basic Concepts]() - [10\.1.0 Basic Concepts](https://www.probabilitycourse.com/chapter10/10_1_0_basic_concepts.php) - [10\.1.1 PDFs and CDFs](https://www.probabilitycourse.com/chapter10/10_1_1_PDFs_and_CDFs.php) - [10\.1.2 Mean and Correlation Functions](https://www.probabilitycourse.com/chapter10/10_1_2_mean_and_correlation_functions.php) - [10\.1.3 Multiple Random Processes](https://www.probabilitycourse.com/chapter10/10_1_3_multiple_random_processes.php) - [10\.1.4 Stationary Processes](https://www.probabilitycourse.com/chapter10/10_1_4_stationary_processes.php) - [10\.1.5 Gaussian Random Processes](https://www.probabilitycourse.com/chapter10/10_1_5_gaussian_random_processes.php) - [10\.1.6 Solved Problems](https://www.probabilitycourse.com/chapter10/10_1_6_solved_probs.php) - [10\.2 Processing of Random Signals]() - [10\.2.0 Processing of Random Signals](https://www.probabilitycourse.com/chapter10/10_2_0_processing_of_random_signals.php) - [10\.2.1 Power Spectral Density](https://www.probabilitycourse.com/chapter10/10_2_1_power_spectral_density.php) - [10\.2.2 Linear Time-Invariant (LTI) Systems with Random Inputs](https://www.probabilitycourse.com/chapter10/10_2_2_LTI_systems_with_random_inputs.php) - [10\.2.3 Power in a Frequency Band](https://www.probabilitycourse.com/chapter10/10_2_3_power_in_a_frequency_band.php) - [10\.2.4 White Noise](https://www.probabilitycourse.com/chapter10/10_2_4_white_noise.php) - [10\.2.5 Solved Problems](https://www.probabilitycourse.com/chapter10/10_2_5_solved_probs.php) - [10\.3 Problems]() - [10\.3.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter10/10_3_0_ch_probs.php) - [11 Some Important Random Processes]() - [11\.1 Poisson Processes]() - [11\.1.0 Introduction](https://www.probabilitycourse.com/chapter11/11_0_0_intro.php) - [11\.1.1 Counting Processes](https://www.probabilitycourse.com/chapter11/11_1_1_counting_processes.php) - [11\.1.2 Basic Concepts of the Poisson Process](https://www.probabilitycourse.com/chapter11/11_1_2_basic_concepts_of_the_poisson_process.php) - [11\.1.3 Merging and Splitting Poisson Processes](https://www.probabilitycourse.com/chapter11/11_1_3_merging_and_splitting_poisson_processes.php) - [11\.1.4 Nonhomogeneous Poisson Processes](https://www.probabilitycourse.com/chapter11/11_1_4_nonhomogeneous_poisson_processes.php) - [11\.1.5 Solved Problems](https://www.probabilitycourse.com/chapter11/11_1_5_solved_probs.php) - [11\.2 Discrete-Time Markov Chains]() - [11\.2.1 Introduction](https://www.probabilitycourse.com/chapter11/11_2_1_introduction.php) - [11\.2.2 State Transition Matrix and Diagram](https://www.probabilitycourse.com/chapter11/11_2_2_state_transition_matrix_and_diagram.php) - [11\.2.3 Probability Distributions](https://www.probabilitycourse.com/chapter11/11_2_3_probability_distributions.php) - [11\.2.4 Classification of States](https://www.probabilitycourse.com/chapter11/11_2_4_classification_of_states.php) - [11\.2.5 Using the Law of Total Probability with Recursion](https://www.probabilitycourse.com/chapter11/11_2_5_using_the_law_of_total_probability_with_recursion.php) - [11\.2.6 Stationary and Limiting Distributions](https://www.probabilitycourse.com/chapter11/11_2_6_stationary_and_limiting_distributions.php) - [11\.2.7 Solved Problems](https://www.probabilitycourse.com/chapter11/11_2_7_solved_probs.php) - [11\.3 Continuous-Time Markov Chains]() - [11\.3.1 Introduction](https://www.probabilitycourse.com/chapter11/11_3_1_introduction.php) - [11\.3.2 Stationary and Limiting Distributions](https://www.probabilitycourse.com/chapter11/11_3_2_stationary_and_limiting_distributions.php) - [11\.3.3 The Generator Matrix](https://www.probabilitycourse.com/chapter11/11_3_3_the_generator_matrix.php) - [11\.3.4 Solved Problems](https://www.probabilitycourse.com/chapter11/11_3_4_solved_probs.php) - [11\.4 Brownian Motion (Wiener Process)]() - [11\.4.0 Brownian Motion (Wiener Process)](https://www.probabilitycourse.com/chapter11/11_4_0_brownian_motion_wiener_process.php) - [11\.4.1 Brownian Motion as the Limit of a Symmetric Random Walk](https://www.probabilitycourse.com/chapter11/11_4_1_brownian_motion_as_the_limit_of_a_symmetric_random_walk.php) - [1\.4.2 Definition and Some Properties](https://www.probabilitycourse.com/chapter11/11_4_2_definition_and_some_properties.php) - [11\.4.3 Solved Problems](https://www.probabilitycourse.com/chapter11/11_4_3_solved_probs.php) - [11\.5 Problems]() - [11\.5.0 End of Chapter Problems](https://www.probabilitycourse.com/chapter11/11_5_0_end_of_chapter_problems.php) - [12 Introduction to Simulation Using MATLAB](https://www.probabilitycourse.com/chapter12/chapter12.php) - [13 Introduction to Simulation Using R](https://www.probabilitycourse.com/chapter13/chapter13.php) - [14 Introduction to Simulation Using Python](https://www.probabilitycourse.com/chapter14/chapter14.php) - [15 Recursive Methods](https://www.probabilitycourse.com/chapter15/chapter15.php) - [Appendix]() - [Some Important Distributions](https://www.probabilitycourse.com/appendix/some_important_distributions.php) - [Review of the Fourier Transform](https://www.probabilitycourse.com/appendix/review_fourier_transform.php) - [Bibliography](https://www.probabilitycourse.com/bibliography.php) [![Creative Commons License](https://i.creativecommons.org/l/by-nc-nd/3.0/88x31.png)](https://creativecommons.org/licenses/by-nc-nd/3.0/deed.en_US) Introduction to Probability by [Hossein Pishro-Nik](https://websites.umass.edu/pishro/) is licensed under a [Creative Commons Attribution-NonCommercial-NoDerivs 3.0 Unported License](https://creativecommons.org/licenses/by-nc-nd/3.0/deed.en_US)
Readable Markdown
*** To better understand Markov chains, we need to introduce some definitions. The first definition concerns the accessibility of states from each other: If it is possible to go from state i to state j, we say that state j is *accessible* from state i. In particular, we can provide the following definitions. We say that state j is **accessible** from state i, written as i → j, if p i j ( n ) \> 0 for some n. We assume every state is accessible from itself since p i i ( 0 ) \= 1. Two states i and j are said to **communicate**, written as i ↔ j, if they are **accessible** from each other. In other words, i ↔ j means i → j and j → i . Communication is an *equivalence* relation. That means that āˆ’ every state communicates with itself, i ↔ i ; āˆ’ if i ↔ j , then j ↔ i ; āˆ’ if i ↔ j and j ↔ k , then i ↔ k . Therefore, the states of a Markov chain can be partitioned into communicating *classes* such that only members of the same class communicate with each other. That is, two states i and j belong to the same class if and only if i ↔ j. *** Example Consider the Markov chain shown in Figure 11.9. It is assumed that when there is an arrow from state i to state j, then p i j \> 0. Find the equivalence classes for this Markov chain. ![MC-diagram-3](https://www.probabilitycourse.com/images/chapter11/MC-diagram-3.png) Figure 11.9 - A state transition diagram. - [**Solution**]() - There are four communicating classes in this Markov chain. Looking at Figure 11.10, we notice that states 1 and 2 communicate with each other, but they do not communicate with any other nodes in the graph. Similarly, nodes 3 and 4 communicate with each other, but they do not communicate with any other nodes in the graph. State 5 does not communicate with any other states, so it by itself is a class. Finally, states 6 , 7 , and 8 construct another class. Thus, here are the classes: Class 1 \= { state 1 , state 2 } , Class 2 \= { state 3 , state 4 } , Class 3 \= { state 5 } , Class 4 \= { state 6 , state 7 , state 8 } . ![MC-diagram-4](https://www.probabilitycourse.com/images/chapter11/MC-diagram-4.png) Figure 11.10 - Equivalence classes. *** A Markov chain is said to be *irreducible* if it has only one communicating class. As we will see shortly, irreducibility is a desirable property in the sense that it can simplify analysis of the limiting behavior. A Markov chain is said to be **irreducible** if all states communicate with each other. Looking at Figure 11.10, we notice that there are two kinds of classes. In particular, if at any time the Markov chain enters Class 4, it will always stay in that class. On the other hand, for other classes this is not true. For example, if X 0 \= 1, then the Markov chain might stay in Class 1 for a while, but at some point, it will leave that class and it will never return to that class again. The states in Class 4 are called *recurrent* states, while the other states in this chain are called *transient*. In general, a state is said to be recurrent if, any time that we leave that state, we will return to that state in the future with probability one. On the other hand, if the probability of returning is less than one, the state is called transient. Here, we provide a formal definition: For any state i, we define f i i \= P ( X n \= i , for some n ≄ 1 \| X 0 \= i ) . State i is **recurrent** if f i i \= 1, and it is **transient** if f i i \< 1. It is relatively easy to show that if two states are in the same class, either both of them are recurrent, or both of them are transient. Thus, we can extend the above definitions to classes. A class is said to be recurrent if the states in that class are recurrent. If, on the other hand, the states are transient, the class is called transient. In general, a Markov chain might consist of several transient classes as well as several recurrent classes. Consider a Markov chain and assume X 0 \= i. If i is a recurrent state, then the chain will return to state i any time it leaves that state. Therefore, the chain will visit state i an infinite number of times. On the other hand, if i is a transient state, the chain will return to state i with probability f i i \< 1. Thus, in that case, the total number of visits to state i will be a Geometric random variable with parameter 1 āˆ’ f i i. Consider a discrete-time Markov chain. Let V be the total number of visits to state i. 1. If i is a recurrent state, then P ( V \= āˆž \| X 0 \= i ) \= 1\. 2. If i is a transient state, then V \| X 0 \= i ∼ G e o m e t r i c ( 1 āˆ’ f i i ) . *** Example Show that in a finite Markov chain, there is at least one recurrent class. - [**Solution**]() - Consider a finite Markov chain with r states, S \= { 1 , 2 , ⋯ , r } . Suppose that all states are transient. Then, starting from time 0 , the chain might visit state 1 several times, but at some point the chain will leave state 1 and will never return to it. That is, there exists an integer M 1 \> 0 such that X n ≠ 1 , for all n ≄ M 1 . Similarly, there exists an integer M 2 \> 0 such that X n ≠ 2 , for all n ≄ M 2 , and so on. Now, if you choose n ≄ max { M 1 , M 2 , ⋯ , M r } , then X n cannot be equal to any of the states 1 , 2 , ⋯ , r . This is a contradiction, so we conclude that there must be at least one recurrent state, which means that there must be at least one recurrent class. *** ### Periodicity: Consider the Markov chain shown in Figure 11.11. There is a periodic pattern in this chain. Starting from state 0, we only return to 0 at times n \= 3 , 6 , ⋯. In other words, p 00 ( n ) \= 0, if n is not divisible by 3. Such a state is called a *periodic* state with period d ( 0 ) \= 3. ![MC-diagram-periodic](https://www.probabilitycourse.com/images/chapter11/MC-diagram-periodic.png) Figure 11.11 - A state transition diagram. The **period** of a state i is the largest integer d satisfying the following property: p i i ( n ) \= 0, whenever n is not divisible by d. The period of i is shown by d ( i ). If p i i ( n ) \= 0, for all n \> 0, then we let d ( i ) \= āˆž. āˆ’ If d ( i ) \> 1 , we say that state i is **periodic**. āˆ’ If d ( i ) \= 1 , we say that state i is **aperiodic**. You can show that all states in the same communicating class have the same period. A class is said to be periodic if its states are periodic. Similarly, a class is said to be aperiodic if its states are aperiodic. Finally, a Markov chain is said to be aperiodic if all of its states are aperiodic. If i ↔ j , then d ( i ) \= d ( j ) . Why is periodicity important? As we will see shortly, it plays a role when we discuss limiting distributions. It turns out that in a typical problem, we are given an irreducible Markov chain, and we need to check if it is aperiodic. How do we check that a Markov chain is aperiodic? Here is a useful method. Remember that two numbers m and l are said to be *co-prime* if their greatest common divisor (gcd) is 1, i.e., gcd ( l , m ) \= 1. Now, suppose that we can find two co-prime numbers l and m such that p i i ( l ) \> 0 and p i i ( m ) \> 0. That is, we can go from state i to itself in l steps, and also in m steps. Then, we can conclude state i is aperiodic. If we have an irreducible Markov chain, this means that the chain is aperiodic. Since the number 1 is co-prime to every integer, any state with a self-transition is aperiodic. Consider a finite irreducible Markov chain X n: 1. If there is a self-transition in the chain ( p i i \> 0 for some i ), then the chain is aperiodic. 2. Suppose that you can go from state i to state i in l steps, i.e., p i i ( l ) \> 0 . Also suppose that p i i ( m ) \> 0 . If gcd ( l , m ) \= 1 , then state i is aperiodic. 3. The chain is aperiodic if and only if there exists a positive integer n such that all elements of the matrix P n are strictly positive, i.e., *** Example Consider the Markov chain in [Example 11.6](https://www.probabilitycourse.com/chapter11/11_2_4_classification_of_states.php#example11_6). 1. Is Class 1 \= { state 1 , state 2 } aperiodic? 2. Is Class 2 \= { state 3 , state 4 } aperiodic? 3. Is Class 4 \= { state 6 , state 7 , state 8 } aperiodic? - [**Solution**]() - 1. Class 1 \= { state 1 , state 2 } is aperiodic since it has a self-transition, p 22 \> 0 . 2. Class 2 \= { state 3 , state 4 } is periodic with period 2 . 3. Class 4 \= { state 6 , state 7 , state 8 } is aperiodic . For example, note that we can go from state 6 to state 6 in two steps ( 6 āˆ’ 7 āˆ’ 6 ) and in three steps ( 6 āˆ’ 7 āˆ’ 8 āˆ’ 6 ). Since gcd ( 2 , 3 ) \= 1 , we conclude state 6 and its class are aperiodic. *** *** | | |---| | The print version of the book is available on [Amazon](https://www.amazon.com/Introduction-Probability-Statistics-Random-Processes/dp/0990637204/ref=sr_1_1?ie=UTF8&qid=1408880878&sr=8-1&keywords=pishro-nik). [![Book Cover](https://www.probabilitycourse.com/images/home/book_cover.jpg)](https://www.amazon.com/Introduction-Probability-Statistics-Random-Processes/dp/0990637204/ref=sr_1_1?ie=UTF8&qid=1408880878&sr=8-1&keywords=pishro-nik) | | **Practical uncertainty:** *Useful Ideas in Decision-Making, Risk, Randomness, & AI* [![ractical Uncertaintly Cover](https://www.probabilitycourse.com/images/home/practical_uncertaintly.PNG)](https://www.amazon.com/dp/B0CH2BHRVH/ref=tmm_pap_swatch_0?_encoding=UTF8&qid=1693837152&sr=8-1) |
Shard115 (laksa)
Root Hash8531609719846996715
Unparsed URLcom,probabilitycourse!www,/chapter11/11_2_4_classification_of_states.php s443