Continuous Semi markov Processes and Their Applications
Semi-Markov Process
Introduction
Shun-Zheng Yu , in Hidden Semi-Markov Models, 2016
1.1.2 Semi-Markov Process
A semi-Markov process is equivalent to a Markov renewal process in many aspects, except that a state is defined for every given time in the semi-Markov process, not just at the jump times. Therefore, the semi-Markov process is an actual stochastic process that evolves over time. Semi-Markov processes were introduced by Levy (1954) and Smith (1955) in 1950s and are applied in queuing theory and reliability theory.
For an actual stochastic process that evolves over time, a state must be defined for every given time. Therefore, the state at time is defined by for . The process is thus called a semi-Markov process. In this process, the times 0= are the jump times of , and are the sojourn times in the states. Every transition from a state to the next state is instantaneously made at the jump times.
For a time-homogeneous semi-Markov process, the transition density functions are
where is independent of the jumping time . It is the probability density function that after having entered state i at time zero the process transits to state j in between time and . They must satisfy
for all . That is, state i must transit to another state in the time .
If the number of jumps in the time interval is , then the sample path is equivalent to the sample path , , , …, , , with probability 1. Then the joint distribution of the process is
where is the probability that the process stays in state for at most time before transiting to another state, and is the probability that the process will not make transition from state i to any other state within time . The likelihood function corresponding to the sample path ( , , , …, , , ) is thus
Suppose the current time is t. The time that has been passed since last jump is defined by . Then the process is a continuous time homogeneous Markov process.
The semi-Markov process can be generated by different types of random mechanisms (Nunn and Desiderio, 1977), for instances:
- 1.
-
Usually, it is thought as such a stochastic process that after having entered state , it randomly determines the successor state based on the state transition probabilities , and then randomly determines the amount of time staying in state before going to state based on the holding time density function , where is the transition probability from state to state , s.t. , and
is the probability that the transition to the next state will occur in the time between and given that the current state is and the next state is . In this model,
- 2.
-
The semi-Markov process can be thought as a stochastic process that after having entered state , it randomly determines the waiting time for transition out of state based on the waiting time density function , and then randomly determines the successor state j based on the state transition probabilities , where is the density function of the waiting time for transition out of state i defined by
and
is the probability that the system will make the next transition to state , given time and current state i. In this model,
- 3.
-
The semi-Markov process can also be thought as such a process that after having entered state , it randomly draws the pair for all , based on , and then determines the successor state and length of time in state from the smallest draw. That is, if , then the next transition is to state and the length of time the process holds in state before going to state is . In this model,
where , and is the probability that the process will not transit to another state except by time . This type of semi-Markov process is applied to such as reliability analysis (Veeramany and Pandey, 2011). An example of this type of semi-Markov process is as follows.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128027677000012
General Hidden Semi-Markov Model
Shun-Zheng Yu , in Hidden Semi-Markov Models, 2016
2.1 A General Definition of HSMM
An HSMM allows the underlying process to be a semi-Markov chain with a variable duration or sojourn time for each state. State duration d is a random variable and assumes an integer value in the set , where D is the maximum duration of a state and can be infinite in some applications. Each state can emit a series of observations, and the number of observations produced while in state i is determined by the length of time spent in the state, that is, the duration d. Now we provide a unified description of HSMMs.
Assume a discrete-time semi-Markov process with a set of (hidden) states . The state sequence is denoted by , where is the state at time . A realization of is denoted as . For simplicity of notation in the following sections, we denote:
- •
-
—state i that the system stays in during the period from to . In other words, it means and . Note that the previous state and the next state may or may not be i.
- •
-
—state that starts at time and ends at with duration . This implies that the previous state and the next state must not be .
- •
-
—state that starts at time and lasts till , with , where means that at the system entered state from some other state, that is, the previous state must not be . The next state may or may not be .
- •
-
—state that lasts from to and ends at with , where means that at time the state will end and transit to some other state at time , that is, the next state must not be . The previous state may or may not be .
Based on these definitions, means state starting and ending at with duration 1, means state starting at , means state ending at , and means the state at being state .
Denote the observation sequence by , where is the observation at time and is the set of observable values. For observation sequence , the underlying state sequence is , , and the state transitions are , for , where , , and . Note that the first state is not necessarily starting at time 1 associated with the first observation and the last state is not necessarily ending at time associated with the last observation . As the states are hidden, the number N of hidden states in the underlying state sequence is also hidden/unknown.
We note that the observable values can be discrete, continuous, or have infinite support, and the observation can be a value, a vector, a symbol, or an event. The length T of the observation sequence can be very large, but is usually assumed to be finite except in the case of online learning. There are usually multiple observation sequences in practice, but we do not always explicitly mention this fact unless it is required. The formulas derived for the single observation sequence usually cannot be directly applied for the multiple observation sequences because the sequence lengths are different with different likelihood functions. Therefore, while applying the formulas derived for the single observation sequence into the case of multiple observation sequences, the formulas must be divided by the likelihood functions if they have not yet appeared in the formulas, where is the lth observation sequence of length T l .
Suppose the current time is , the process has made jumps, and the time spent since the previous jump is . As explained in Section 1.1.2, the process is a discrete-time homogeneous Markov process. Its subsequence is also a Markov process based on the Markov property. Then we can define the state transition probability from state i having duration h to state having duration d by
which is assumed independent of time , for , . The transition probabilities must satisfy , for all given and , with zero self-transition probabilities , for all and . In other words, when a state ends at time t, it cannot transit to the same state at the next time t+1 because the state durations are explicitly specified by some distributions other than geometric or exponential distributions. From the definition we can see that the previous state started at and ended at , with duration . Then it transits to state having duration , according to the state transition probability . State will start at and end at . This means both the state and the duration are dependent on both the previous state and its duration. While in state , there will be observations being emitted. Denote this emission/observation probability by
which is assumed to be independent of time , where is the observed values of . Let the distribution of the first state be
or
depending on the model assumption that the first state is starting at or before. We can equivalently let the initial distribution of the state be
It represents the probability of the initial state and its duration before time or before the first observation obtained. The relationship between the two definitions of initial state distribution is , where if the starting time of the first state must be then ; otherwise, if the first state can start at or before , then . Usually, the second definition of the initial state distribution, , makes the computation of the forward variables in the HSMM algorithms simpler. Then the set of model parameters for the HSMM is defined by
or
where represents an observable substring of length for . This general HSMM is shown in Figure 1.6.
The general HSMM is reduced to specific models of HSMM depending on the assumptions made. For instances,
- 1.
-
If the state duration is assumed to be independent of the previous state, then the state transition probability can be further specified as , where
(2.1)
is the transition probability from state that has stayed for duration to state that will start at , and
(2.2)
is the probability of duration that state will take. This is the model proposed by Marhasev et al. (2006). Compared with the general HSMM, the number of model parameters is reduced from to , and the state duration distributions can be explicitly expressed using probability density functions (e.g., Gaussian distributions) or a probability mass function.
- 2.
-
If a state transition is assumed to be independent of the duration of the previous state, then the state transition probability from (i,h) to (j,d) becomes , where
(2.3)
is the transition probability that state ended at and transits to state having duration . If it is assumed that a state transition for is and a self-transition is , for , then the model becomes the residual time HMM (Yu and Kobayashi, 2003a). In this model, the starting time of the state is not of concern, but the ending time is of interest. Therefore, d represents the remaining sojourn (or residual life) time of state . This model is obviously appropriate to applications for which the residual life is of the most concern. The number of model parameters is reduced to . More importantly, if the state duration is further assumed to be independent of the previous state, then the state transition probability can be specified as . In this case, the computational complexity will be the lowest among all HSMMs. The number of model parameters is further reduced to .
- 3.
-
If self-transition is allowed and the state duration is assumed to be independent of the previous state, then the state transition probability becomes
where ; is the self-transition probability when state has stayed for time units, that is,
and is the probability that state ends with duration . This is the variable transition HMM (Krishnamurthy et al., 1991; Vaseghi, 1991). In this model, a state transition is either for or for a self-transition. This process is similar to the standard discrete-time semi-Markov process. The concept of the discrete-time semi-Markov process can thus be used in modeling an application. This model has model parameters. The computational complexity is relatively high compared with other conventional HSMMs.
- 4.
-
If a transition to the current state is independent of the duration of the previous state and the duration of the current state is only conditioned on the current state itself, then
where is the transition probability from state to state , with the self-transition probability . This is the explicit duration HMM (Ferguson, 1980), with model parameters and lower computational complexity. This is the simplest and the most popular model among all HSMMs, with easily understandable formulas and modeling concepts.
Besides, the general form of observation distributions can be simplified and dedicated to applications. They can be parametric (e.g., a mixture of Gaussian distributions) or nonparametric (e.g., a probability mass function), discrete or continuous, and dependent on or independent of the state durations. The observations can be assumed dependent or conditionally independent for given states, that is, . The conditional independence makes HSMMs simpler and so is often assumed in the literature.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128027677000024
Stochastic Modeling Techniques for Secure and Survivable Systems
Kishor S. Trivedi , ... Selvamuthu Dharmaraja , in Information Assurance, 2008
Computations of Steady-State Probabilities
The embedded DTMC for the above discussed SMP is shown in Figure 7.2. As mentioned earlier in subsection 7.2.2, we need the information of sojourn times in each state and the transition probabilities. The corresponding parameters are listed as follows:
FIGURE 7.2. Embedded DTMC for the SMP model.
-
hG : Mean time for a system to resist becoming vulnerable to attacks.
-
hV : Mean time for a system to resist attacks when vulnerable.
-
hA : Mean time taken by a system to detect an attack and initiate triage actions.
-
hMC : Mean time a system can keep the effects of an attack masked.
-
hUC : Mean time that an attack remains undetected while doing damage.
-
hTR : Mean time a system takes to evaluate how best to handle an attack.
-
hFS : Mean time a system operates in a fail-secure mode in the presence of an attack.
-
hGD : Mean time a system is in the degraded state in the presence of an attack.
-
hF : Mean time a system is in the failed state despite detecting an attack.
-
pa : Probability of injecting a successful attack, given that a system is vulnerable.
-
pu : Probability that a successful attack has remain undetected.
-
pm : Probability that a system successfully masks an attack.
-
pg : Probability that a system resists an attack by graceful degradation.
-
ps : Probability that a system responds to an attack in a fail-secure manner.
For computing the security attributes in terms of availability, confidentiality, and integrity, we need to determine the steady-state probabilities {π i , i ∈ S} of the SMP states. These probabilities are to be determined in terms of the steady-state probabilities and the mean sojourn time hi of the states of the DTMC.
The matrix P that describes the state transition probabilities for this DTMC is written as:
where , , and On solving the equation
(7.4)
where the steady-state probabilities are obtained.
Next, we compute the mean sojourn time hi in each state i. The quantity hi is determined by a random time the process spends in state i. The attacker behavior is described by the transitions G → V and V → A. To model the wide range of attacks (from amateur mischief to cyber attacks), it is necessary to consider a variety of probability distributions. On the other hand, system response to an attack is algorithmic and automated. Based on the sojourn time distribution of state i, the mean sojourn time hi for this state is calculated. For example, if sojourn time distribution for state G is hypoexponential with parameters λg 1 and λ g 2, then its mean sojourn time hG is given as . Similarly, the mean sojourn times for the other states hA, hV, hMC, hGD, hTR, hFS, hGD, hF are also computed.
Steady-state probability for SMP states is expressed in terms of the steady-state probabilities of the DTMC and their sojourn times hi using Eq. 7.3. Substituting for and hi in Eq. 7.3, we get the steady-state probabilities for the SMP as:
(7.5)
Once the steady-state probabilities are known, we now compute the security attributes such as availability, confidentiality, and integrity.
For calculating availability, we observe that a system is not available in states FS, F, and UC and is available in all the other states. Then, the availability is given as:
(7.6)
In case of a DoS attack, a system is made to stop functioning, that is, bringing it to the FS state will accomplish the goal of a DoS attack. Therefore, the states FS and MC will not be part of the state transition diagram. For this attack, system availability is given as:
(7.7)
Similarly, confidentiality and integrity measures can be computed in the context of specific security attacks. For example, Microsoft IIS 4.0 suffered from the ASP vulnerability as documented in the Bugtraq ID 1002 [20]. Exploitation of this vulnerability allows an attacker to traverse the entire web server file system, thus compromising confidentiality. Therefore, in the context of this attack, states UC and F are identified with the loss of confidentiality. Similarly, if Code-Red worm is modified to inject a piece of code into a vulnerable IIS server to browse unauthorized files, states UC and F will imply loss of confidentiality. Therefore, the steady-state confidentiality measure is computed as:
(7.8)
Consider another example, where a Common Gateway Interface (CGI) vulnerability present in the Samber server as reported in Bugtraq ID 1002 was reported [20]. Exploitation of this vulnerability permits an attacker to execute any MS-DOS command including deletion and modification of files in an unauthorized manner, thus compromising the integrity of a system. Here also, states UC and F indicate the loss of integrity, and thus the steady-state measure of integrity is given as:
(7.9)
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780123735669500094
Dependable and Secure Systems Engineering
Kishor Trivedi , ... Fumio Machida , in Advances in Computers, 2012
3.2.3 Job Completion Time
Yet another example of combined performance and reliability analysis is computation of the job completion time on a system subject to component failure and repair. The distribution of the job completion time on a computer system considering CPU failure and repair was originally studied in Ref. [51]. The CPU or the server model used in the study was a three-state SMP, and the job completion time was analyzed in a general manner [29,52]. Figure 35 shows the SMP for the CPU model where state 1 represents the server up state; state 2 is the state where the server is recovering from a nonfatal failure; and state 3 is the state where the server is recovering from a fatal failure.
Fig. 35. Three-state CPU model.
The state 1 and the state 2 are categorized as pre-emptive resume (prs) states in which the job execution is resumed from the interrupted point. On the other hand, the state 3 is categorized as pre-emptive repeat identical (pri) state in which the job execution is restarted from the beginning. A job that started execution when the server is in state 1 may encounter a nonfatal error that leads to the server state change from 1 to 2. The job execution also faces a fatal error that causes the server state change from 1 to 3. Both of nonfatal error and fatal error are repairable and their times to recovery follow general distribution G 2(t) and G 3(t), respectively. Assuming that failures are exponentially distributed with rate λ and each failure is either nonfatal with probability p nf or fatal with probability p f = 1 − p nf, the SMP kernel distributions are given by following expressions.
Using the analysis method developed in Ref. [52] to the SMP model, we can obtain the Laplace–Stieltjes transforms (LSTs) of the job completion time distribution F 1 ∼(s, x) for fixed work amount x:
where Q 21 ∼(s) and G 3 ∼(s) are the LSTs of Q 21(t) and G 3(t).
Then LST can be numerically inverted, or by taking derivatives expected completion time determined.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780123965257000010
Non-Markovian Queueing Systems
J. MEDHI , in Stochastic Models in Queueing Theory (Second Edition), 2003
6.3.4 Semi-Markov process approach
The system-size {N(t)} is not semi-Markovian. Consider that a transition occurs with a service completion (departure of a unit)— that is, tn, n = 0,1,2,… are the nth departure epochs. Using the notation of Section 1.9,
we see that {Y(t), t ≥ 0} is a semi-Markov process that has { Xn, n ≥ 0} for its embedded Markov chain. We get
One can proceed, as in Section 1.9, to find
and
See Fabens (1961) for a relationship between νj and πj , where
See also Neuts (1967) for semi-Markov analysis of the more general model M/G(a,b)/1.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780124874626500060
Spatial Choice Models
H. Timmermans , in International Encyclopedia of the Social & Behavioral Sciences, 2001
5 Complex Spatial Choice Behavior
The previous models are all typically based on single-purpose, single-stop behavior. Over the years, however, it became clear that an increasing proportion of trips involved multistop behavior. Moreover, Hägerstrand's time geography had convincingly argued that behavior does not reflect preferences only, but also constraints. Thus, various attempts have been made to develop models of trip chaining and activity-travel patterns.
Originally, most models relied on semi-Markov process models or Monte Carlo simulations. More recently, utility-maximizing models have dominated the field. An important contribution in this regard was made by Kitamura ( 1984), who introduced the concept of prospective utility. It states that the utility of a destination is not only a function of its inherent attributes and the distance to that destination, but also of the utility of continuing the trip from that destination. Dellaert et al. (1998) generalized Kitamura's approach to account for multipurpose aspects of the trip chain.
Trip chaining is only one aspect of multiday activity/travel patterns. Several models have recently been suggested to predict more comprehensive activity patterns. These models can be differentiated into the older constraints-based models (e.g., Lenntorp 1976), utility-maximizing models (e.g., Recker et al. 1986), rule-based or computational process models (e.g., Golledge et al. 1994, ALBATROSS – Arentze and Timmermans 2000).
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B0080430767025201
Markov chain models and applications
Kishor S. Trivedi , ... Dharmaraja Selvamuthu , in Modeling and Simulation of Computer Networks and Systems, 2015
2 Strengths of Markov models
The essential necessity for a stochastic process to be a homogeneous continuous time Markov chain (CTMC) is that the sojourn time in each state must be exponentially distributed. However, the sojourn times in each state may not follow exponential distribution while modeling practical or real-time situations. The existence of the non-exponentially distributed event time gives rise to non-Markovian models. A non-Markovian model can be modeled using phase-type approximation. However, phase-type expansion increases the already large state-space of a real system model. The problem becomes really severe when mixing deterministic times with exponential ones. The strict Markovian constraints are relaxed by using Markov regenerative processes (MRGP). A generalization of CTMC where the time spent by the process in a given state is allowed to follow non-exponential (general) distribution is a semi-Markov process (SMP). Further generalization is provided by MRGP. This concept is used in extending the CTMC model by allowing general distribution for all the event times other than failure times in the given examples. As a result the stochastic process under consideration becomes MRGP.
A Markov renewal process becomes a Markov process when the transition times are independent exponential and are independent of the next state visited. It becomes a Markov chain when the transition times are all identically equal to 1. It reduces to a renewal process if there is only one state and then only transition time becomes relevant. Renewal theory is used to analyze stochastic processes that regenerate themselves from time-to-time. The long-run behavior of a regenerative stochastic process can be studied in terms of its behavior during a single regeneration cycle. Semi-Markov processes are used in the study of certain queuing systems.
Numerous studies have described and reported the occurrence of "software aging" [2–4] in which the state of software degrades with time. This degradation is caused primarily by the exhaustion of operating system resources, data corruption and numerical error accumulation. If untreated, this may lead to performance degradation of the software or crash/hang failure, or both in the long run. Examples of software aging are memory bloating and leaking, unreleased file-locks, data corruption, storage space fragmentation and accumulation of round-off errors [3,4]. Aging has not only been observed in software used on a mass scale but also in specialized software used in high-availability and safety-critical applications [2,5]. To counteract software aging, a preventive maintenance technique called "software rejuvenation" has been proposed [2,6,7], which involves periodically stopping the system, cleaning up, and restarting it from a clean internal state. This "renewal" of software prevents (or at least postpones) a crash failure. The internal state of the software can be cleaned by techniques like garbage collection, flushing operating system kernel tables and reinitializing internal data structures.
Rejuvenation has been implemented in various types of systems, from telecommunication systems [2,8,9], operating systems [10], transaction processing systems [11], web servers [12–14], cluster servers [15–17], cable modem systems [18], spacecraft systems [19], safety-critical systems [5,20], to biomedical applications [21]. Preventive maintenance, however, incurs an overhead (lost transactions, downtime, additional resources, etc.) which should be balanced against the cost incurred due to unexpected outage caused by failure. This in turn demands a quantitative analysis, which in the context of software systems has only recently started to receive attention.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128008874000134
Implementation of HSMM Algorithms
Shun-Zheng Yu , in Hidden Semi-Markov Models, 2016
4.4.4 Unknown Observation Distribution
Usually, based on the empirical knowledge on the stochastic process, the observation distribution can be determined whether they are parametric or nonparametric. If they are assumed to be parametric, their probability density distribution functions can be correspondingly determined. When the parametric distribution is unknown, the most popular ones that are often used in practice are a mixture of Gaussian distributions.
Example 4.4 Parametric Distribution of Observations
Use Example 1.4 . Assume the observation distributions are parametric, and the request arrivals is characterized as a Poisson process modulated by an underlying (hidden state) semi-Markov process. The finite number of discrete states are defined by the discrete mean arrival rates. Let be the mean arrival rate for given state . Then the number of arrivals in a time interval and the Markov state are related through the conditional probability distribution
where is assumed conditionally independent.
Note that when the observation distributions are parametric, the new parameters for state j can be found by maximizing subject to the constraint . For instance, if the probability density function , for , is Poisson with mean , then the parameter can be estimated by or, equivalently,
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128027677000048
Mitosis detection in biomedical images
Yao Lu , ... Yu-Ting Su , in Computer Vision for Microscopy Image Analysis, 2021
3.3.7 Hidden conditional random field and semi-Markov model (HCRF and SMM)
The previous methods discussed in this chapter mainly focus on the identification of mitotic sequences. The graph model can also be used for cell state segmentation. A semi-Markov model (SMM) has been applied to model the mitosis process with different stages [6] . SMM is a random field model for sequence segmentation by modeling the state transition as a semi-Markov process. In the sequence segmentation framework, mitotic sequences are first identified by the HCRF method, as defined in Section 3.3.4. Based on the detection results from the HCRF model, SMM further segments the sequence into different stages, which are defined by the transition of cell appearance during the mitosis process. Generally, we can define four stages of the cell appearance transition: (1) interphase, (2) the start of mitosis, (3) formation of daughter cells, and (4) separation of daughter cells. The four stages of mitosis are illustrated in Fig. 6.14.
Fig. 6.14. The four stages of the mitosis process.
(This image is from "Anan Liu, Kang Li, Takeo Kanade, A semi-Markov model for mitosis segmentation in time-lapse phase contrast microscopy image sequences of stem cell populations, IEEE Trans. Med. Imaging 31(2) (2012) 359–369.")As shown in Fig. 6.15, given the input sequence X = {x t } t = 1 T, SMM divides the sequence into segments S = {s i } i = 1 s . Each segment s i is represented by a pair of integers (u i , l i ), where u i indicates the position of the last frame in s i and l i represents the state of s i .
Fig. 6.15. The HMM model structure.
The SMM also generates an overall prediction of the whole sequence as y ′ ∈ {0, 1}:
(6.33)
where γ is the parameter vector for the SMM learned during training.
The sequence is considered to be a valid mitosis process only if the sequence contains the complete state transition process from stage 1 to stage 4. Thus, we define y′ = 1 when S contains one (and only one) complete stage transition process, and y′ = 0 in other cases.
With this definition, Eq. (6.33) can be simplified as
(6.34)
The potential function γ T · ψ(X, S) can be defined as
(6.35)
where ψ 1(·) indicates the averaged feature vectors, ψ 2(·) indicates the standard deviation of feature vectors, and [·] represents vector concatenation operation.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128149720000060
Applications of HSMMs
Shun-Zheng Yu , in Hidden Semi-Markov Models, 2016
9.3 Network Traffic Characterization and Anomaly Detection
In this application, HSMMs are applied to characterize the network traffic. Measurements of real traffic often indicate that a significant amount of variability is presented in the traffic observed over a wide range of time scales, exhibiting self-similar or long-range dependent characteristics (Leland et al., 1994). Such characteristics can have a significant impact on the performance of networks and systems (Tuan and Park, 1999; Park et al., 1997). Therefore, better understanding the nature of network traffic is critical for network design, planning, management, and security. A major advantage of using an HSMM is the capability of capturing various statistical properties of the traffic, including the long-range dependence (Yu et al., 2002). It can also be used together with, for example, matrix-analytic methods to obtain analytically tractable solutions to queueing-theoretic models of server performance (Riska et al., 2002).
In this application, an observation in the observation sequence represents the number of user requests/clicks, packets, bytes, connections, etc., arriving in a time unit. It can also be the inter-arrival time between requests, packets, URLs, or protocol keywords. The observation sequence is characterized as a discrete-time random process modulated by an underlying (hidden state) semi-Markov process. The hidden state represents the density of traffic, mass of active users, or a web page that is hyperlinked with others.
Using the HSMM trained by the normal behavior, one can detect anomaly embedded in the network behavior according to its likelihood or entropy against the model (Yu, 2005; Li and Yu, 2006; Lu and Yu, 2006a; Xie and Yu, 2006a,b Xie and Yu, 2006a Xie and Yu, 2006b ; Xie and Zhang, 2012; Xie and Tang, 2012; Xie et al., 2013a,b Xie et al., 2013a Xie et al., 2013b ), recognize user click patterns (Xu et al., 2013), extract users' behavior features (Ju and Xu, 2013) for SaaS (Software as a Service), or estimate the packet loss ratios and their confidence intervals (Nguyen and Roughan, 2013).
For example, a web workload (requests/s) recorded in the peak hour is shown in Figure 1.7 (gray line). The arrival process for a given state j can be assumed as, for instance, Poisson process with one parameter . The initial value of is assumed to be proportional to its state index j, that is,
so that higher state corresponds to higher arrival rate, where M is the total number of hidden states. In considering the range of the observable values (requests/s), the total number M of hidden states is initially assumed to be 30. During the re-estimation procedure, the states that are never visited will be deleted from the state space. To characterize the second order self-similar property of the workload, the duration of state j can be assumed as, for instance, a heavy-tailed Pareto distribution with one parameter . The initial values of can be assumed equal for all states. To reduce the computational amount, the maximum duration D of the states can be assumed to be finite with sufficiently large value to cover the maximum duration of any state in the given observation sequence, where D=500 s is assumed. As a reasonable choice, the initial values of the probabilities a ij and are assumed uniform.
Given these initial assumptions for the explicit duration HSMM, the ML model parameters can be estimated using the re-estimation algorithm of Algorithm 3.1. The MAP states S t , for t=1, …, T, can be estimated using Eqn (2.15). The results showed that there were 20 hidden states modulating the arrival rate of requests, and only 41 state transitions occurring during 3600 s. The maximum duration D went up to 405 and the process stayed in the same state for a mean duration of 87.8 s. There were two classes of states among the 20 states: 5 states in the middle played a major role in modulating the arrival streams in the sense that the process spent most time in these 5 states; and the remaining 15 states having the higher and lower indices represented the rare situations that had ultra high or low arrival rates lasting very short time.
Read full chapter
URL:
https://www.sciencedirect.com/science/article/pii/B9780128027677000097
Source: https://www.sciencedirect.com/topics/computer-science/semi-markov-process
0 Response to "Continuous Semi markov Processes and Their Applications"
Post a Comment