官术网_书友最值得收藏!

Creating a Markov chain

Let's get started by creating a Markov chain, on which the MDP is developed.

A Markov chain describes a sequence of events that comply with the Markov property. It is defined by a set of possible states, S = {s0, s1, ... , sm}, and a transition matrix, T(s, s'), consisting of the probabilities of state s transitioning to state s'. With the Markov property, the future state of the process, given the present state, is conditionally independent of past states. In other words, the state of the process at t+1 is dependent only on the state at t. Here, we use a process of study and sleep as an example and create a Markov chain based on two states, s0 (study) and s1 (sleep). Let's say we have the following transition matrix:

In the next section, we will compute the transition matrix after k steps, and the probabilities of being in each state given an initial distribution of states, such as [0.7, 0.3], meaning there is a 70% chance that the process starts with study and a 30% chance that it starts with sleep.

主站蜘蛛池模板: 绥宁县| 涟水县| 神木县| 英吉沙县| 海原县| 毕节市| 河南省| 手游| 纳雍县| 吉木萨尔县| 临西县| 沂水县| 弥勒县| 白山市| 绥中县| 荥阳市| 奉新县| 嘉鱼县| 新乡县| 天等县| 河北区| 乌海市| 清河县| 南投县| 泽州县| 大同市| 南宫市| 洛浦县| 耒阳市| 扎兰屯市| 沙田区| 海南省| 铁力市| 临沭县| 曲沃县| 石屏县| 通辽市| 雷州市| 建湖县| 汉源县| 长沙市|