MarketStochastic process
Company Profile

Stochastic process

In probability theory and related fields, a stochastic or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Stochastic processes are widely used as mathematical models of systems and phenomena that appear to vary in a random manner. Examples include the growth of a bacterial population, an electrical current fluctuating due to thermal noise, or the movement of a gas molecule. Stochastic processes have applications in many disciplines such as biology, chemistry, ecology, neuroscience, physics, image processing, signal processing, control theory, information theory, computer science, and telecommunications. Furthermore, seemingly random changes in financial markets have motivated the extensive use of stochastic processes in finance.

Introduction
A stochastic or random process can be defined as a collection of random variables that is indexed by some mathematical set, meaning that each random variable of the stochastic process is uniquely associated with an element in the set. When interpreted as time, if the index set of a stochastic process has a finite or countable number of elements, such as a finite set of numbers, the set of integers, or the natural numbers, then the stochastic process is said to be in discrete time. If the index set is some interval of the real line, then time is said to be continuous. The two types of stochastic processes are respectively referred to as discrete-time and continuous-time stochastic processes. Discrete-time stochastic processes are considered easier to study because continuous-time processes require more advanced mathematical techniques and knowledge, particularly due to the index set being uncountable. If the index set is the integers, or some subset of them, then the stochastic process can also be called a random sequence. In his work on probability Ars Conjectandi, originally published in Latin in 1713, Jakob Bernoulli used the phrase "Ars Conjectandi sive Stochastice", which has been translated to "the art of conjecturing or stochastics". This phrase was used, with reference to Bernoulli, by Ladislaus Bortkiewicz who in 1917 wrote in German the word stochastik with a sense meaning random. The term stochastic process first appeared in English in a 1934 paper by Joseph Doob. though the German term had been used earlier, for example, by Andrei Kolmogorov in 1931. According to the Oxford English Dictionary, early occurrences of the word random in English with its current meaning, which relates to chance or luck, date back to the 16th century, while earlier recorded usages started in the 14th century as a noun meaning "impetuosity, great speed, force, or violence (in riding, running, striking, etc.)". The word itself comes from a Middle French word meaning "speed, haste", and it is probably derived from a French verb meaning "to run" or "to gallop". The first written appearance of the term random process pre-dates stochastic process, which the Oxford English Dictionary also gives as a synonym, and was used in an article by Francis Edgeworth published in 1888. Terminology The definition of a stochastic process varies, but a stochastic process is traditionally defined as a collection of random variables indexed by some set. Both "collection", while instead of "index set", sometimes the terms "parameter set" though sometimes it is only used when the stochastic process takes real values. while the terms stochastic process and random process are usually used when the index set is interpreted as time, and other terms are used such as random field when the index set is n-dimensional Euclidean space \mathbb{R}^n or a manifold. \{X(t)\} or simply as X. Some authors mistakenly write X(t) even though it is an abuse of function notation. For example, X(t) or X_t are used to refer to the random variable with the index t, and not the entire stochastic process. If the index set is T=[0,\infty), then one can write, for example, (X_t , t \geq 0) to denote the stochastic process. ==Examples==
Examples
Bernoulli process One of the simplest stochastic processes is the Bernoulli process, In other words, a Bernoulli process is a sequence of iid Bernoulli random variables, where each idealised coin flip is an example of a Bernoulli trial. Random walk Random walks are stochastic processes that are usually defined as sums of iid random variables or random vectors in Euclidean space, so they are processes that change in discrete time. But some also use the term to refer to processes that change in continuous time, particularly the Wiener process used in financial models, which has led to some confusion, resulting in its criticism. There are various other types of random walks, defined so their state spaces can be other mathematical objects, such as lattices and groups, and in general they are highly studied and have many applications in different disciplines. A classic example of a random walk is known as the simple random walk, which is a stochastic process in discrete time with the integers as the state space, and is based on a Bernoulli process, where each Bernoulli variable takes either the value positive one or negative one. In other words, the simple random walk takes place on the integers, and its value increases by one with probability, say, p, or decreases by one with probability 1-p, so the index set of this random walk is the natural numbers, while its state space is the integers. If p=0.5, this random walk is called a symmetric random walk. Wiener process The Wiener process is a stochastic process with stationary and independent increments that are normally distributed based on the size of the increments. The Wiener process is named after Norbert Wiener, who proved its mathematical existence, but the process is also called the Brownian motion process or just Brownian motion due to its historical connection as a model for Brownian movement in liquids. Playing a central role in the theory of probability, the Wiener process is often considered the most important and studied stochastic process, with connections to other stochastic processes. Its index set and state space are the non-negative numbers and real numbers, respectively, so it has both continuous index set and states space. But the process can be defined more generally so its state space can be n-dimensional Euclidean space. If the mean of any increment is zero, then the resulting Wiener or Brownian motion process is said to have zero drift. If the mean of the increment for any two points in time is equal to the time difference multiplied by some constant \mu, which is a real number, then the resulting stochastic process is said to have drift \mu. Almost surely, a sample path of a Wiener process is continuous everywhere but nowhere differentiable. It can be considered as a continuous version of the simple random walk. The process arises as the mathematical limit of other stochastic processes such as certain random walks rescaled, which is the subject of Donsker's theorem or invariance principle, also known as the functional central limit theorem. The Wiener process is a member of some important families of stochastic processes, including Markov processes, Lévy processes and Gaussian processes. It plays a central role in quantitative finance, where it is used, for example, in the Black–Scholes–Merton model. The process is also used in different fields, including the majority of natural sciences as well as some branches of social sciences, as a mathematical model for various random phenomena. Poisson process The Poisson process is a stochastic process that has different forms and definitions. It can be defined as a counting process, which is a stochastic process that represents the random number of points or events up to some time. The number of points of the process that are located in the interval from zero to some given time is a Poisson random variable that depends on that time and some parameter. This process has the natural numbers as its state space and the non-negative numbers as its index set. This process is also called the Poisson counting process, since it can be interpreted as an example of a counting process. The homogeneous Poisson process is a member of important classes of stochastic processes such as Markov processes and Lévy processes. If the parameter constant of the Poisson process is replaced with some non-negative integrable function of t, the resulting process is called an inhomogeneous or nonhomogeneous Poisson process, where the average density of points of the process is no longer constant. Serving as a fundamental process in queueing theory, the Poisson process is an important process for mathematical models, where it finds applications for models of events randomly occurring in certain time windows. Defined on the real line, the Poisson process can be interpreted as a stochastic process, among other random objects. But then it can be defined on the n-dimensional Euclidean space or other mathematical spaces, where it is often interpreted as a random set or a random counting measure, instead of a stochastic process. But it has been remarked that the Poisson process does not receive as much attention as it should, partly due to it often being considered just on the real line, and not on other mathematical spaces. ==Definitions==
Definitions
Stochastic process A stochastic process is defined as a collection of random variables defined on a common probability space (\Omega, \mathcal{F}, P), where \Omega is a sample space, \mathcal{F} is a \sigma-algebra, and P is a probability measure; and the random variables, indexed by some set T, all take values in the same mathematical space S, which must be measurable with respect to some \sigma-algebra \Sigma. \{X(t):t\in T \}. Historically, in many problems from the natural sciences a point t\in T had the meaning of time, so X(t) is a random variable representing a value observed at time t. A stochastic process can also be written as \{X(t,\omega):t\in T \} to reflect that it is actually a function of two variables, t\in T and \omega\in \Omega. There are other ways to consider a stochastic process, with the above definition being considered the traditional one. For example, a stochastic process can be interpreted or defined as a S^T-valued random variable, where S^T is the space of all the possible functions from the set T into the space S. Index set The set T is called the index set of the stochastic process. Often this set is some subset of the real line, such as the natural numbers or an interval, giving the set T the interpretation of time. such as the Cartesian plane \mathbb{R}^2 or n-dimensional Euclidean space, where an element t\in T can represent a point in space. That said, many results and theorems are only possible for stochastic processes with a totally ordered index set. State space The mathematical space S of a stochastic process is called its state space. This mathematical space can be defined using integers, real lines, n-dimensional Euclidean spaces, complex planes, or more abstract mathematical spaces. The state space is defined using elements that reflect the different values that the stochastic process can take. Sample function A sample function is a single outcome of a stochastic process, so it is formed by taking a single possible value of each random variable of the stochastic process. More precisely, if \{X(t,\omega):t\in T \} is a stochastic process, then for any point \omega\in\Omega, the mapping X(\cdot,\omega): T \rightarrow S, is called a sample function, a realization, or, particularly when T is interpreted as time, a sample path of the stochastic process \{X(t,\omega):t\in T \}. This means that for a fixed \omega\in\Omega, there exists a sample function that maps the index set T to the state space S. or path. Increment An increment of a stochastic process is the difference between two random variables of the same stochastic process. For a stochastic process with an index set that can be interpreted as time, an increment is how much the stochastic process changes over a certain time period. For example, if \{X(t):t\in T \} is a stochastic process with state space S and index set T=[0,\infty), then for any two non-negative numbers t_1\in [0,\infty) and t_2\in [0,\infty) such that t_1\leq t_2, the difference X_{t_2}-X_{t_1} is a S-valued random variable known as an increment. For a measurable subset B of S^T, the pre-image of X gives X^{-1}(B)=\{\omega\in \Omega: X(\omega)\in B \}, so the law of a X can be written as: Finite-dimensional probability distributions For a stochastic process X with law \mu, its finite-dimensional distribution for t_1,\dots,t_n\in T is defined as: \mu_{t_1,\dots,t_n} =P\circ (X({t_1}),\dots, X({t_n}))^{-1}, This measure \mu_{t_1,..,t_n} is the joint distribution of the random vector (X({t_1}),\dots, X({t_n})) ; it can be viewed as a "projection" of the law \mu onto a finite subset of T. For any measurable subset C of the n-fold Cartesian power S^n=S\times\dots \times S, the finite-dimensional distributions of a stochastic process X can be written as: But the concept of stationarity also exists for point processes and random fields, where the index set is not interpreted as time. When the index set T can be interpreted as time, a stochastic process is said to be stationary if its finite-dimensional distributions are invariant under translations of time. This type of stochastic process can be used to describe a physical system that is in steady state, but still experiences random fluctuations. A sequence of random variables forms a stationary stochastic process only if the random variables are identically distributed. Khinchin introduced the related concept of stationarity in the wide sense, which has other names including covariance stationarity or stationarity in the broad sense. Filtration A filtration is an increasing sequence of sigma-algebras defined in relation to some probability space and an index set that has some total order relation, such as in the case of the index set being some subset of the real numbers. More formally, if a stochastic process has an index set with a total order, then a filtration \{\mathcal{F}_t\}_{t\in T} , on a probability space (\Omega, \mathcal{F}, P) is a family of sigma-algebras such that \mathcal{F}_s \subseteq \mathcal{F}_t \subseteq \mathcal{F} for all s \leq t, where t, s\in T and \leq denotes the total order of the index set T. Modification A modification of a stochastic process is another stochastic process, which is closely related to the original stochastic process. More precisely, a stochastic process X that has the same index set T, state space S, and probability space (\Omega,{\cal F},P) as another stochastic process Y is said to be a modification of X if for all t\in T the following P(X_t=Y_t)=1 , holds. Two stochastic processes that are modifications of each other have the same finite-dimensional law and they are said to be stochastically equivalent or equivalent. Instead of modification, the term version is also used, however some authors use the term version when two stochastic processes have the same finite-dimensional distributions, but they may be defined on different probability spaces, so two processes that are modifications of each other, are also versions of each other, in the latter sense, but not the converse. The theorem can also be generalized to random fields so the index set is n-dimensional Euclidean space as well as to stochastic processes with metric spaces as their state spaces. Indistinguishable Two stochastic processes X and Y defined on the same probability space (\Omega,\mathcal{F},P) with the same index set T and set space S are said be indistinguishable if the following P(X_t=Y_t \text{ for all } t\in T )=1 , holds. Separability Separability is a property of a stochastic process based on its index set in relation to the probability measure. The property is assumed so that functionals of stochastic processes or random fields with uncountable index sets can form random variables. For a stochastic process to be separable, in addition to other conditions, its index set must be a separable space, which means that the index set has a dense countable subset. More precisely, a real-valued continuous-time stochastic process X on a probability space (\Omega,{\cal F},P) is separable iff its index set T has a dense countable subset U\subset T and there is a set \Omega_0 \subset \Omega of probability zero, so P(\Omega_0)=0, such that for every open set G\subset T and every closed set F\subset \mathbb{R} =(-\infty,\infty) , the two events \{ X_t \in F \text{ for all } t \in G\cap U\} and \{ X_t \in F \text{ for all } t \in G\} differ from each other at most on a subset of \Omega_0. The definition of separability can also be stated for other index sets and state spaces, such as in the case of random fields, where the index set as well as the state space can be n-dimensional Euclidean space. A theorem by Doob, sometimes known as Doob’s separability theorem, says that any real-valued continuous-time stochastic process has a separable modification. Versions of this theorem also exist for more general stochastic processes with index sets and state spaces other than the real line. Uncorrelatedness Two stochastic processes \left\{X_t\right\} and \left\{Y_t\right\} are called uncorrelated if their cross-covariance \operatorname{K}_{\mathbf{X}\mathbf{Y}}(t_1,t_2) = \operatorname{E} \left[ \left( X(t_1)- \mu_X(t_1) \right) \left( Y(t_2)- \mu_Y(t_2) \right) \right] is zero for all times. Formally: :\left\{X_t\right\},\left\{Y_t\right\} \text{ uncorrelated} \quad \iff \quad \operatorname{K}_{\mathbf{X}\mathbf{Y}}(t_1,t_2) = 0 \quad \forall t_1,t_2. Independence implies uncorrelatedness If two stochastic processes X and Y are independent, then they are also uncorrelated. Such functions are known as càdlàg or cadlag functions, based on the acronym of the French phrase continue à droite, limite à gauche. A Skorokhod function space, introduced by Anatoliy Skorokhod, The notation of this function space can also include the interval on which all the càdlàg functions are defined, so, for example, D[0,1] denotes the space of càdlàg functions defined on the unit interval [0,1]. Skorokhod function spaces are frequently used in the theory of stochastic processes because it often assumed that the sample functions of continuous-time stochastic processes belong to a Skorokhod space. Regularity In the context of mathematical construction of stochastic processes, the term regularity is used when discussing and assuming certain conditions for a stochastic process to resolve possible construction issues. For example, to study stochastic processes with uncountable index sets, it is assumed that the stochastic process adheres to some type of regularity condition such as the sample functions being continuous. ==Further examples==
Further examples
Markov processes and chains Markov processes are stochastic processes, traditionally in discrete or continuous time, that have the Markov property, which means the next value of the Markov process depends on the current value, but it is conditionally independent of the previous values of the stochastic process. In other words, the behavior of the process in the future is stochastically independent of its behavior in the past, given the current state of the process. The Brownian motion process and the Poisson process (in one dimension) are both examples of Markov processes in continuous time, while random walks on the integers and the gambler's ruin problem are examples of Markov processes in discrete time. A Markov chain is a type of Markov process that has either discrete state space or discrete index set (often representing time), but the precise definition of a Markov chain varies. For example, it is common to define a Markov chain as a Markov process in either discrete or continuous time with a countable state space (thus regardless of the nature of time), but it has been also common to define a Markov chain as having discrete time in either countable or continuous state space (thus regardless of the state space). Markov processes form an important class of stochastic processes and have applications in many areas. For example, they are the basis for a general stochastic simulation method known as Markov chain Monte Carlo, which is used for simulating random objects with specific probability distributions, and has found application in Bayesian statistics. The concept of the Markov property was originally for stochastic processes in continuous and discrete time, but the property has been adapted for other index sets such as n-dimensional Euclidean space, which results in collections of random variables known as Markov random fields. Martingale A martingale is a discrete-time or continuous-time stochastic process with the property that, at every instant, given the current value and all the past values of the process, the conditional expectation of every future value is equal to the current value. In discrete time, if this property holds for the next value, then it holds for all future values. The exact mathematical definition of a martingale requires two other conditions coupled with the mathematical concept of a filtration, which is related to the intuition of increasing available information as time passes. Martingales are usually defined to be real-valued, but they can also be complex-valued or even more general. A symmetric random walk and a Wiener process (with zero drift) are both examples of martingales, respectively, in discrete and continuous time. In this aspect, discrete-time martingales generalize the idea of partial sums of independent random variables. Martingales can also be created from stochastic processes by applying some suitable transformations, which is the case for the homogeneous Poisson process (on the real line) resulting in a martingale called the compensated Poisson process. Martingales mathematically formalize the idea of a 'fair game' where it is possible form reasonable expectations for payoffs, and they were originally developed to show that it is not possible to gain an 'unfair' advantage in such a game. Many problems in probability have been solved by finding a martingale in the problem and studying it. Martingales will converge, given some conditions on their moments, so they are often used to derive convergence results, due largely to martingale convergence theorems. Martingales have many applications in statistics, but it has been remarked that its use and application are not as widespread as it could be in the field of statistics, particularly statistical inference. They have found applications in areas in probability theory such as queueing theory and Palm calculus and other fields such as economics and finance. These processes have many applications in fields such as finance, fluid mechanics, physics and biology. The main defining characteristics of these processes are their stationarity and independence properties, so they were known as processes with stationary and independent increments. In other words, a stochastic process X is a Lévy process if for n non-negatives numbers, 0\leq t_1\leq \dots \leq t_n, the corresponding n-1 increments X_{t_2}-X_{t_1}, \dots , X_{t_n}-X_{t_{n-1}}, are all independent of each other, and the distribution of each increment only depends on the difference in time. If the specific definition of a stochastic process requires the index set to be a subset of the real line, then the random field can be considered as a generalization of stochastic process. Point process A point process is a collection of points randomly located on some mathematical space such as the real line, n-dimensional Euclidean space, or more abstract spaces. Sometimes the term point process is not preferred, as historically the word process denoted an evolution of some system in time, so a point process is also called a random point field. There are different interpretations of a point process, such a random counting measure or a random set. Some authors regard a point process and stochastic process as two different objects such that a point process is a random object that arises from or is associated with a stochastic process, though it has been remarked that the difference between point processes and stochastic processes is not clear. Other stochastic processes such as renewal and counting processes are studied in the theory of point processes. ==History==
History
Early probability theory Probability theory has its origins in games of chance, which have a long history, with some games being played thousands of years ago, but very little analysis on them was done in terms of probability. The year 1654 is often considered the birth of probability theory when French mathematicians Pierre Fermat and Blaise Pascal had a written correspondence on probability, motivated by a gambling problem. But there was earlier mathematical work done on the probability of gambling games such as Liber de Ludo Aleae by Gerolamo Cardano, written in the 16th century but posthumously published later in 1663. After Cardano, Jakob Bernoulli wrote Ars Conjectandi, which is considered a significant event in the history of probability theory. Bernoulli's book was published, also posthumously, in 1713 and inspired many mathematicians to study probability. But despite some renowned mathematicians contributing to probability theory, such as Pierre-Simon Laplace, Abraham de Moivre, Carl Gauss, Siméon Poisson and Pafnuty Chebyshev, most of the mathematical community did not consider probability theory to be part of mathematics until the 20th century. Statistical mechanics In the physical sciences, scientists developed in the 19th century the discipline of statistical mechanics, where physical systems, such as containers filled with gases, are regarded or treated mathematically as collections of many moving particles. Although there were attempts to incorporate randomness into statistical physics by some scientists, such as Rudolf Clausius, most of the work had little or no randomness. This changed in 1859 when James Clerk Maxwell contributed significantly to the field, more specifically, to the kinetic theory of gases, by presenting work where he modelled the gas particles as moving in random directions at random velocities. The kinetic theory of gases and statistical physics continued to be developed in the second half of the 19th century, with work done chiefly by Clausius, Ludwig Boltzmann and Josiah Gibbs, which would later have an influence on Albert Einstein's mathematical model for Brownian movement. Measure theory and probability theory At the International Congress of Mathematicians in Paris in 1900, David Hilbert presented a list of mathematical problems, where his sixth problem asked for a mathematical treatment of physics and probability involving axioms. In the early 1930s, Khinchin and Kolmogorov set up probability seminars, which were attended by researchers such as Eugene Slutsky and Nikolai Smirnov, and Khinchin gave the first mathematical definition of a stochastic process as a set of random variables indexed by the real line. Birth of modern probability theory In 1933, Andrei Kolmogorov published in German, his book on the foundations of probability theory titled Grundbegriffe der Wahrscheinlichkeitsrechnung, where Kolmogorov used measure theory to develop an axiomatic framework for probability theory. The publication of this book is now widely considered to be the birth of modern probability theory, when the theories of probability and stochastic processes became parts of mathematics. did early work on the theory of stochastic processes, making fundamental contributions, particularly in the theory of martingales. Starting in the 1940s, Kiyosi Itô published papers developing the field of stochastic calculus, which involves stochastic integrals and stochastic differential equations based on the Wiener or Brownian motion process. Also starting in the 1940s, connections were made between stochastic processes, particularly martingales, and the mathematical field of potential theory, with early ideas by Shizuo Kakutani and then later work by Joseph Doob. In 1953, Doob published his book Stochastic processes, which had a strong influence on the theory of stochastic processes and stressed the importance of measure theory in probability. Doob also chiefly developed the theory of martingales, with later substantial contributions by Paul-André Meyer. Earlier work had been carried out by Sergei Bernstein, Paul Lévy and Jean Ville, the latter adopting the term martingale for the stochastic process. Methods from the theory of martingales became popular for solving various probability problems. Techniques and theory were developed to study Markov processes and then applied to martingales. Conversely, methods from the theory of martingales were established to treat Markov processes. which would later result in Varadhan winning the 2007 Abel Prize. In the 1990s and 2000s the theories of Schramm–Loewner evolution and rough paths were introduced and developed to study stochastic processes and other mathematical objects in probability theory, which respectively resulted in Fields Medals being awarded to Wendelin Werner in 2008 and to Martin Hairer in 2014. The theory of stochastic processes still continues to be a focus of research, with yearly international conferences on the topic of stochastic processes. Bernoulli process The Bernoulli process, which can serve as a mathematical model for flipping a biased coin, is possibly the first stochastic process to have been studied. Bernoulli's work, including the Bernoulli process, were published in his book Ars Conjectandi in 1713. Random walks In 1905, Karl Pearson coined the term random walk while posing a problem describing a random walk on the plane, which was motivated by an application in biology, but such problems involving random walks had already been studied in other fields. Certain gambling problems that were studied centuries earlier can be considered as problems involving random walks. and is an example of a random walk with absorbing barriers. Pascal, Fermat and Huyens all gave numerical solutions to this problem without detailing their methods, and then more detailed solutions were presented by Jakob Bernoulli and Abraham de Moivre. For random walks in n-dimensional integer lattices, George Pólya published, in 1919 and 1921, work where he studied the probability of a symmetric random walk returning to a previous position in the lattice. Pólya showed that a symmetric random walk, which has an equal probability to advance in any direction in the lattice, will return to a previous position in the lattice an infinite number of times with probability one in one and two dimensions, but with probability zero in three or higher dimensions. Wiener process The Wiener process or Brownian motion process has its origins in different fields including statistics, finance and physics. The work is now considered as an early discovery of the statistical method known as Kalman filtering, but the work was largely overlooked. It is thought that the ideas in Thiele's paper were too advanced to have been understood by the broader mathematical and statistical community at the time. in order to model price changes on the Paris Bourse, a stock exchange, without knowing the work of Thiele. and Bachelier's thesis is now considered pioneering in the field of financial mathematics. Einstein's work, as well as experimental results obtained by Jean Perrin, later inspired Norbert Wiener in the 1920s to use a type of measure theory, developed by Percy Daniell, and Fourier analysis to prove the existence of the Wiener process as a mathematical object. There are a number of claims for early uses or discoveries of the Poisson process. Another discovery occurred in Denmark in 1909 when A.K. Erlang derived the Poisson distribution when developing a mathematical model for the number of incoming phone calls in a finite time interval. Erlang was not at the time aware of Poisson's earlier work and assumed that the number phone calls arriving in each interval of time were independent to each other. He then found the limiting case, which is effectively recasting the Poisson distribution as a limit of the binomial distribution. which had been commonly regarded as a requirement for such mathematical laws to hold. Starting in 1928, Maurice Fréchet became interested in Markov chains, eventually resulting in him publishing in 1938 a detailed study on Markov chains. Andrei Kolmogorov developed in a 1931 paper a large part of the early theory of continuous-time Markov processes. He introduced and studied a particular set of Markov processes known as diffusion processes, where he derived a set of differential equations describing the processes. Independent of Kolmogorov's work, Sydney Chapman derived in a 1928 paper an equation, now called the Chapman–Kolmogorov equation, in a less mathematically rigorous way than Kolmogorov, while studying Brownian movement. The differential equations are now called the Kolmogorov equations or the Kolmogorov–Chapman equations. Other mathematicians who contributed significantly to the foundations of Markov processes include William Feller, starting in the 1930s, and then later Eugene Dynkin, starting in the 1950s. In addition to Lévy, Khinchin and Kolomogrov, early fundamental contributions to the theory of Lévy processes were made by Bruno de Finetti and Kiyosi Itô. ==Mathematical construction==
Mathematical construction
In mathematics, constructions of mathematical objects are needed, which is also the case for stochastic processes, to prove that they exist mathematically. Another approach involves defining a collection of random variables to have specific finite-dimensional distributions, and then using Kolmogorov's existence theorem to prove a corresponding stochastic process exists. says that if any finite-dimensional distributions satisfy two conditions, known as consistency conditions, then there exists a stochastic process with those finite-dimensional distributions. This means that the distribution of the stochastic process does not, necessarily, specify uniquely the properties of the sample functions of the stochastic process. Another problem is that functionals of continuous-time process that rely upon an uncountable number of points of the index set may not be measurable, so the probabilities of certain events may not be well-defined. Separability ensures that infinite-dimensional distributions determine the properties of sample functions by requiring that sample functions are essentially determined by their values on a dense countable set of points in the index set. Furthermore, if a stochastic process is separable, then functionals of an uncountable number of points of the index set are measurable and their probabilities can be studied. for a continuous-time stochastic process with any metric space as its state space. For the construction of such a stochastic process, it is assumed that the sample functions of the stochastic process belong to some suitable function space, which is usually the Skorokhod space consisting of all right-continuous functions with left limits. This approach is now more used than the separability assumption, but such a stochastic process based on this approach will be automatically separable. Although less used, the separability assumption is considered more general because every stochastic process has a separable version. == Application ==
Application
Applications in Finance Black-Scholes Model One of the most famous applications of stochastic processes in finance is the Black-Scholes model for option pricing. Developed by Fischer Black, Myron Scholes, and Robert Merton (whose contributions led to the 1997 Nobel Memorial Prize in Economic Sciences), this model uses Geometric Brownian motion, a specific type of stochastic process, to describe the dynamics of asset prices. The model assumes that the price of a stock follows a continuous-time stochastic process driven by a stochastic differential equation (SDE): dS_t = \mu S_t\,dt + \sigma S_t\,dW_t Where: • S_t is the price of the underlying asset at time . • is the drift rate, representing the expected return of the asset. • is the volatility of the asset's returns. • W_t is a standard Wiener process (Brownian motion) introducing the random market shocks. The key assumption of the Black-Scholes model is that the price of a financial asset, such as a stock, follows a log-normal distribution, with its continuous returns following a normal distribution. Because of these properties, the model provides a closed-form solution for pricing European-style options. The Black-Scholes formula has had a profound impact on financial markets, forming the basis for much of modern options trading. Although the model has limitations, such as the assumption of constant volatility and no transaction costs, it remains widely used due to its simplicity and practical relevance. Stochastic Volatility Models Another significant application of stochastic processes in finance is in stochastic volatility models, which aim to capture the time-varying nature of market volatility. The Heston model is a popular example, allowing for the volatility of asset prices to follow its own stochastic process rather than remaining strictly constant. In the Heston model, the asset price and its variance are modeled as a system of coupled stochastic differential equations: dS_t = \mu S_t\,dt + \sqrt{v_t} S_t\,dW_t^S dv_t = \kappa(\theta - v_t)\,dt + \xi \sqrt{v_t}\,dW_t^v Where: • v_t is the instantaneous variance, which follows a mean-reverting Cox–Ingersoll–Ross process. • is the long-run average variance, and is the rate at which v_t reverts to . • is the "volatility of volatility". • W_t^S and W_t^v are two distinct Wiener processes that are mathematically correlated (with correlation coefficient ), which allows the model to capture leverage effects (e.g., volatility often increases when stock prices fall). Unlike the Black-Scholes model, which assumes constant volatility and results in a flat volatility surface, stochastic volatility models provide a more flexible framework for modeling market dynamics. They successfully reproduce the "volatility smile" observed in real-world options pricing, making them particularly crucial during periods of high uncertainty or market stress. Applications in Biology Population Dynamics One of the primary applications of stochastic processes in biology is in population dynamics. In contrast to deterministic models, which assume that populations change in predictable ways, stochastic models account for the inherent randomness in births, deaths, and migration. The birth-death process, a simple stochastic model, describes how populations fluctuate over time due to random births and deaths. These models are particularly important when dealing with small populations, where random events can have large impacts, such as in the case of endangered species or small microbial populations. Another example is the branching process, These methods balance computational efficiency with accuracy, making them invaluable for handling large datasets. Randomized algorithms are also extensively applied in areas such as cryptography, large-scale simulations, and artificial intelligence, where uncertainty must be managed effectively. This is particularly relevant in network traffic analysis and server management. For instance, queuing models help predict delays, manage resource allocation, and optimize throughput in web servers and communication networks. The flexibility of stochastic models allows researchers to simulate and improve the performance of high-traffic environments. For example, queueing theory is crucial for designing efficient data centers and cloud computing infrastructures. ==See also==
tickerdossier.comtickerdossier.substack.com