Stochastic Processing

Unfolding Stochasticity Sequentially

Profile Picture
Kaiwen Bian

Random Walk: Key Example

Random Walk: Key Example

Setting Bases of Sequential Process In Nature

Stochastic processes is abut understanding from local to gloabl, by having single transition probability matrix (discrete) or transition probability function (continuous) and trying to understand some global behavior of such chain or processes (i.e. absorption time, expected time of termination, or stationary distribution). WHat I found stochastic processing to be really amazing is how it is trying to model how randomness and stochasticity may play out in nature, and not at one time stamp but rather across different time stamp sequentially. It try to reason and capture some details about the nature, to undertsand a chain of "randomness" seuqntiall (not so random when you cna model it). Compare to random variables that reasons about how "randomness" plays out at one time stamp, stochatsic processes may dive into the interactions between "randomness" across time, to dive into their interactions, seeing how they sequentially interact and the prior effect the laters (of course under markov property only one layer of looking backward is needed).

In this section I want to discuss (not techniqually) about how amazing this field may be with one classic example that bridges across discrete Markove Chains (MC), Continuous Time Markov Chains (CTMC) and COntinuous Time Continuous State Brownian Motion (BM): Random Walk (RW).

Some Notes on Stochastic Processes

Markov Chain: Analysis Into Stochaticity

When considering about a sequential process that have "randomness' depending on teh previous "randomness", onemust consider using ideas from conditional probability, which discusses about a distribution given that a different distribution has occured. No matter MC, CTMC, or BM, they all hold/are established on a very powerful property known as the Markov Property that makes many of the techniqual details working with conditional probability much easier. Mathamatically: $$ P(X_n = x | X_{n-1} = x_{n-1}, \dots, X_0 = x_0) = P(X_n = x | X_{n-1} = x_{n-1}) $$ $$ P(X_{n+1} = x_{n+1} | X_n = x_n) = P(X_{n+1} = x_{n+1} | X_n = x_n, \dots, X_0 = x_0) $$ Saying that the "current" \(X_n\) does not depends on distance past but rather only the nearest past \(X_{n-1}\) and that the "future" \(X_{n+1}\) does not depend on the past but only the "present" \(X_n\).

With Markov Property holding, mathamatician can reason with the chain or such sequential process much more easily and some global probability such as expected time of termination, return probability, stationary distribution can be analytically calculated. One key theme in stochastic processes is that one might want to setup an recureent (recursive) system to reason about how things unfold, then finding generalized pattern (explicit solution) from such system of equation. Remanber that a recurrence system is equivalently represented in linear algebra as a system of equation when flatening all of it out, then numerical optimization can also be conducted. Taking an example of the return probability: $$ u_{ik} = P_{ik} + \sum_{j=0}^{k-1} P_{ij} u_{jk} = P(X_{T} = k | X_0 = i) $$ which talks about teh probability of terminating the chain at point \(k\) given that it starts on point \(i\). Equivalently, it can be written in matrix format for numerical optimization: $$ u^{(k)} = (I - Q)^{-1} R^{(k)} $$ Same idea can be used to reaosn with expected time of termination: $$ E[T | X_0 = i] = 1 + \sum_{j=0}^{r-1} P_{ij} \omega_j = \omega_j $$ These types of reaosnings are known to be First Step Analysis and it is one of the mos commonly used idea to reason with a sequential event in stochastic processing.

To the core example that I want to cover in this section: Random Walk (RW). RW in the discrete state + discrete time condition is essentially a Markov Chain with the transition probability matrix: $$ \Pr(X_{n+1} = j | X_n = i) = \begin{cases} q_i & \text{if } j = i-1 \\ r_i & \text{if } j = i \\ p_i & \text{if } j = i+1 \end{cases} $$ under the constrained that \(q_i + r_i + p_i = 1\).

Knowing that this is a Markov Chain and its transition probability matrix, much can be conducted and all of the analysis that was mentioned earlier can be used to help finding a global state of the chain. It is worth notice that though vanilla RW is only in discrete state + discrete time condition, it is actually a core example taht can be carried over to more complicated situations. Even under discrete condition, it can serev a pretty fine model for stochastic modeling in some practical senerios. For instance, gamble's ruin modeling or modeling stochasticity to see if learning has occured.

Continuous Time Markov Chain: Discretize + Functional Analysis

Now this is where tings gets a lot more complicated because CTMC is trying to reason under the condition that time is continuous, which introduces many more deficulties mathamaticaly. Remenber that we said one core idea in stochastic processes is to setup a recurence system and try to solve such system? When state is discrete, we can count them as steps and do recursion in that fashion, but when states are continuous, recursion can still be conducted, but sometimes with differential equation, which is not something that w want to do, the complexity is very high. When studying more into CTMC, one may encounter many classic processes that are CTMC, namely most of the classic examples deals with reasoning "when customer come into and leave a store":

These process in nature are easier to be described by rate diagram, which is how they are defined (rate diagram description is one of the three ways to interpret an CTMC) and we can use random variables to monitor such process or deducing transition probability function (i.e. increments of an Poisson Process follows Poison distribution random variable). Maybe in some sense, these random variavbles may be a good abstraction to represent the process just like how weights can be used as a representation of a predictive function? (ideas in Jump & Hold description?)

Problem with description comes where it becomes kind of hard trying to find some global property when reasoning under this paradigm. Thus, we can extend to the second and third interpretation of an CTMC: Jump & Hold Description and Infinite Decimal Generator Decription. This is where things really gets complicated because the first need to project an contnious chain to an discrete chain by discretize the continuous scale into a discrete numbers of continuous random variable depending on teh process in interest and the second is invlolved in using Functional Analysis techniques and theorems to find an Q Matrix that captures time dependent information while being itself a time independent matrix (hence the name genertaor). Here is a few names that might be worth looking into:

Different Description of CTMC

Different Description of Continuous Time Markov Chain

However, though being extremely complex, when dive into the analysis, it has a deep connection back with random walk, but just under a much complicated condition.

Brownian Motion: Bridging Analyst & Probabilist

Moving to Brownain motion, this is modeling stochasticity under continuous time and continuous state, which has a very deep theoritical root connecting back to the Analysis aspects of mathamatics. This is also where the quote I put on the left comes from where equality equation refers to the Analysis aspects of mathamatics and inequality equation refers to the modeling aspects of mathamatics.

Inspired by Prof.Carfagnini: "In math you either deal with equality equation or inequality equation, but sometimes they bridge"

Essentially in Brownian Motion, wes ay that modeling a system of differentiual equation to know everything about how heat dissapate in a region is the eqivalent with deploying a stochaastic agent into the environment, let it play around and see how the probabilistic rollout would be to map out the boundary condition (analyst's approach v.s. a probabilist's approach). And, again, when looking it from certain angle, the connection to Random Walk als shows up again.