# I. Introduction The analysis of vehicle driving styles is prominent to the field of intelligent transportation and vehicle calibration [1,2]. The term driving style can be referred as a set of dynamic activities or steps that a driver uses when driving. Hence, this type of research impacts eco-driving, road safety, and intelligent vehicles [3,4,5]. To model these driving styles, one popular approach is the use of a Hierarchical Dirichlet Process Hidden Semi-Markov Model (HDP-HSMM) [6]. This model is powerful in that it considers the sequential nature of driving kinematic signals, and estimates data segmentation, behavior state duration, and state transition probabilities. The HDP-HSMM provides semantical way for analyzing driver behaviors, and is thus popularly used for describing driving styles. Figure 1b shows an While the HDP-HSMM is powerful, literature outside of the field of transportation details how the model's use of an HDP prior can lead to redundant and inconsistent state estimations. This detail is important as it needs to be considered by researchers attempting to utilize the HDP-HSMM to describe driving styles. For example, Figure 1 clearly has redundant states as seen by the green shaded states. The redundant states can make analysis of HDP-HSMM outputs across multiple datasets difficult for researchers hoping to utilize the HDP-HSMM to model driving styles. This paper addresses this issue by presenting an algorithm that reduces redundant states to improve consistency while still aligning to the structure of a basic HDP-HSMM. The presented algorithm results a more robust HDP-HSMM (rHDP-HSMM) that is expected to output a more consistent data segmentation, behavior state duration, and state transition probabilities than a basic HDP-HSMM. This will impact the transportation field in that driving maneuver patterns can be better grouped together for classification or behavioral studies. The remainder of this paper is as follows. Section 2 will provide the background about HDP-HSMM's from a statistical perspective, and highlight the current set of approaches towards addressing the issues derived from the HDP prior. Section 3 will provide the data description and the model formulation of a basic HDP-HSMM. Section 4 discusses the details of inference for a HDP-HSMM, and how this paper's algorithm can be included within the inference to produce a more robust HDP-HSMM. Section 5 presents a simulation study, in which the rHDP-HSMM is compared to the basic HDP-HSMM based on simulated data. Section 6 presents a case study that uses realistic, naturalistic driving data to compare the rHDP-HSMM with the original HDP-HSMM method on the basis of describing driving patterns. Finally, Section 7 summarizes new contributions and major conclusions of the paper. The HDP-HSMM was designed to improve upon the structure of a discrete state-space Hidden Markov Model (HMM). HMM's are also popularly used for describing sequential data [7,8,9,10,11,12]. In particular, the HMM [13,14] utilizes a two-layer structure (Figure 2a) to represent sequential data observed at equally spaced time points. In this model, data is assumed to be generated from a set of probability distribution functions dependent on corresponding hidden states. The hidden states determine the data segmentation. Transitions among hidden states are modeled as a Markov Chain. This allows for the consideration of time sequence information during inference and further aids in the prediction of future states. One condition of using the Markov Chain is that the state duration of each hidden state is assumed to be Geometrically distributed. While the HMM is able to define data segmentation and state transitions, its definition of state duration is severely limited by the model's structure. This limitation lead to the development of the Hierarchical Dirichlet Process Hidden Semi-Markov Model (HDP-HSMM) [15] which provided two key improvements to the HMM. The first improvement was the removal of the HMM's assumption of geometrically distributed state duration. As the HDP-HSMM uses a Semi-Markovian approach to model the state transitions ?zs , this removes self-transitions from the transition matrix. As a consequence, this frees the geometric distribution restriction on the duration D s , which leads to a three-layer structure model as shown in (Figure 2b). In other words, users can choose different models for representing state duration, while allowing the segmentation of hidden states to be directly represented by z s . The second improvement was the introduction of Dirichlet Processes to the model. The Dirichlet processes is an extension to the Dirichlet distribution, as atoms can be sampled from it based on an input distribution. However, one key difference is that the Dirichlet Process assigns a probability of drawing a new atom from the input distribution and a separate probability of drawing an atom based on the atoms seen in previous samples. The resulting distribution is discrete and similar to the input distribution, but also has the possibility of having infinite discrete atoms if infinite samples were drawn. This phenomenon is interesting in the context of HMMs and HSMMs, as the Dirichlet process can be used as a prior to the state transition probability vector [16,17,15]. Doing this allows the probability vector length (i.e. models' number of states) to grow without limit during inference, which implies the Dirichlet process also acts like a prior on the number of clusters. In the HDP-HSMM, a Hierarchical Dirichlet Process (HDP) is used as a prior on the state transitions, which allows all the state transition probabilities to share a similar base distribution. This is beneficial, as all the states represented in the base distribution are shared between all the different state transition probabilities, while allowing each transition probability be dependent on the exit state. Hence, for the context of modeling of driving maneuvers, the HDP-HSMM is preferred as it allows greater flexibility in defining the relationship between the data and segmentation, state duration, and state transitions. While the Dirichlet Process's clustering properties have been seen as a tool to address the model selection for Bayesian nonparametric approaches [18,19], the Dirichlet Process is known to have inconsistency issues regarding estimation of the true number of states. posterior does not concentrate at the true number of components, and instead introduces extra clusters even if they are not needed. Under the context of HMMs, [21] showed how the Dirichlet Process also leads to the creation of redundant states, which presents an unrealistic rapid switching between states in the inferred transition matrices. Under the context of HSMM's, Figure 1 shows how this side effect occurs even in the HDP-HSMM. However, for the HDP-HSMM, the redundancy issue also affects the inference of transition probabilities and duration estimation. A few works exist that focus on solving this issue for HMM's. [22] discussed HMM's utilizing a Dirichlet prior, and the assumptions on the prior required for the consistency. [23] developed the sticky HDP-HMM (sHDP-HMM) to consider the issue of redundant states. This model adds a bias to the prior on the rows of the transition matrix which emphasizes self-transitions. This results in an increased state duration for each learnt state, which allows the sHDP-HMM to avoid redundant states with short state duration. However, this strategy cannot be applied to HDP-HSMM as the modeling structure of HMM's is inherently different from HSMM's. Outside of HMM and HSMM modeling, [24] focused on the Dirichlet Process Mixture model, and presented the Merge-Truncate-Merge algorithm, which guaranteed a consistent estimate to the number of mixture components. This post-processing procedure takes advantage of the fact that the posterior sample tends to produce a large number of atoms with small weights, and probabilistically merges atoms together. Given these approaches, this paper attempts to address the HDP's inconsistency problem by taking inspiration from both the sticky HDP-HMM and the Merge-Truncate-Merge algorithm. The idea is to apply a merging procedure during inference which promotes longer durations and the avoidance of redundant states. In doing so, this paper's contribution will include demonstrating how the HDP-HSMM becomes robust to the inconsistencies brought by the HDP prior and how this paper's method can reduce the number of redundant states to better define driving maneuvers existing in Figure 1a. A brief summary, which describes where our model fits in relation to the other models described in HMM literature, is given in Table 1. # State Duration Distribution Model Extension (not sensitive to prior) Geometric HDP-HMM [14] sticky HDP-HMM [7] Any Discrete Distribution HDP-HSMM [15] robust HDP-HSMM (This paper) In this paper, a sequential dataset consists of a series of observations collected at T chronologically ordered time points. At each time point t, y t ? R p represents the p-dimensional signal responses. The sequential data is assumed to follow multiple phases; there exists a partition [20] provided an example for Dirichlet Process Mixture Models which demonstrates how the segment, and (3) identify the probability of transitioning from one distribution to another. The challenge lies in little information being available relating to the number of states, the states' durations, and the transition probability matrix. 1 = t 1 1 ? t 1 2 ? ... ? t 1 S = T -D S , # b) Basis of HDP-HSMMs and Notations The HDP-HSMM accomplishes this objective with the following structure. The multivariate sequential data is represented by the sequence (y t ) t=1:T := {y t ? R p : t = 1, ..., T } and is assumed to transit among K different hidden states. The hidden states at each time point t are represented by the sequence (x t ) t=1:T := {x t ? {1, 2, . . . , K} : t = 1, ..., T }, and can be further divided into S segments. Within each data segment s ? {1, 2, . . . , S}, all hidden states share the same index (labeled by the super-state z s ? {1, 2, . . . , K}), and the state duration of the segment is denoted by D s . As such, the start and end times of each segment s are indexed by time stamps t 1 s and t 2 s , respectively. They can be calculated as t 1 s = s T -t|x t+1 = i)p(y t+1:T |x t+1 = i, D t+1 > T -t) B T (i) :=1, where F t = 1 denotes a new segment begins at t + 1, and D t+1 denotes the duration of the segment that begins at time t + 1 [29]. The procedure for obtaining the posterior state sequence begins by drawing a sample for the first state using the following formula: p(x 1 = k|y 1:T ) ? p(x 1 = k)B * 0 (k). Next, a sample is drawn from the posterior duration distribution by conditioning on sampled initial state x1 : p(D 1 = d|y 1:T , x 1 = x1 , F 0 = 1) = p(D 1 = d)p(y 1 : d|D 1 = d, x 1 = x1 , F 0 = 1)B d (x 1 ) B * 0 (x 1 ) . The rest of the state sequence can be sampled assuming the new initial state has distribution p(x D 1 +1 = i|x 1 = x1 ) and repeating the process, until a state is assigned for all indices t = 1, . . . , T . Step 4: Once the new state sequence is sampled, the Gibbs sampling procedure normally returns back to Step 1, increments m by 1, and repeats Steps 1 to 3 until posterior convergence. However, before doing that, this paper propose adding an additional sampling Step 4 that removes redundant states from the posterior state sequence (x t ) (m) ? h (xt) ((x t )|{? i } (m) , {? i } (m) , {? i } (m) (x t ) (m) , (y t ), H, G, ?),(2) where h (xt) (?) represents a sampling step proposed by this paper to promote robustness. The proposed Step 4 is the main contribution of this paper. This section will provide the details on how to implement Equation 2 described in Step 4 above. The procedure is described by first defining redundancy between two states: Definition 4.1. In the state sequence (x t ) t=1:T , the states i and j are identified as redundant states if D f (? i ), f (? j ) ? ? , where ? is the decision threshold and D f (? i ), f (? j ) is a measure of divergence that gets larger when the distributions f (? i ) and f (? j ) are more different from one another. Although D f (? i ), f (? j ) can be any measure of divergence satisfying Definition 4.1, the remainder of the paper will assume D f (? i ), f (? j ) = ||(? i -? j )|| 2 is the â??" 2 norm of the difference in parameters. Now that redundancy has been defined, the details of Equation 2 can be represented by Algorithm 1. In short, the procedure samples a new state sequence that contains no redundant states. [15] describes a weak-limit approximation to the Dirichlet Process prior, ?|? ? Dir(?, . . . , ?) ? j |? ? Dir(?? 1 , . . . , ?? K ), j = 1, . . . , K, b) Implementation of Step 4 as well as an augmentation that introduces auxiliary variables which are added to the ? vector to preserve conjugacy. This approximation eases the use of sampling procedures when dealing Dirichlet Processes [30]. Taking this approach, the ? vector takes no consideration of redundant states, which may negatively impact the posterior of ? j . The presence of redundant states means the posterior transition probabilities contain extra transitions to and from redundant states, which dilute the underlying transition process. To counter this, h (xt) (?) aims to adjust the ? vector in this step as to discourage transitions to redundant states in future steps, and preserve the true underlying transition process. Algorithm 1 describes h (xt) (?) entirely. The procedure begins by initializing a new vector ?, a new state sequence (x t ) (m) , and taking the input of a similarity threshold ? . Taking inspiration from [24], the states order is firstly randomized in which redundancy is checked. This is to ensure the start of the merging procedure begins at a point close to the "central mass" of the emission distribution clusters with a high probability. Going through the order, if the state exists within the new state sequence (x t ) (m) , the algorithm proceeds to find similar states based on our similarity metric and similarity threshold. Weights are then defined which will determine the probability of retaining a state from the set of redundant states. These weights are determined by the probability of other non-similar states transitioning to the state of interest and then normalized. The state by which to retain is selected randomly in accordance to the probabilistic weights, and the rest of the similar states are erased from the state sequence. Vector ? is further updated by weakening the unselected similar states values in the vector. After implementing Algorithm 1, the sampling procedure is allowed to return to Step 1. Noticeably, every time this step is implemented, the algorithm begins with the originally sampled ? and (x t ) (m) , but ends with a ? and (x t ) (m) that encourages the transition matrix in Step 1 to promote transitions to non-redundant states and allow larger sample sizes for # Algorithm 1 Sample a State Sequence Containing No Redundant States Initialize ? = ?, (x t ) (m) = (x t ) (m) , and define similarity threshold ? Reorder {? i : i ? (x t ) (m) } into new order {? I i : i ? (x t ) (m) } using random sampling without replacement where ? i corresponds to index the unique states existing in (x t ) (m) ? I i corresponds to the new index of state i in the new order I = {1, 2, 3, . . . } while I is not an empty set do Let i correspond to the first I i appearing in the new order I Calculate D f (? i ), f (? j ) for all j ? = i where j ? (x t ) (m) ? Similarity metric. Define set J = {j : D f (? i ), f (? j ) ? ? } and set J ? = {j : D f (? i ), f (? j ) > ? } for j ? J do ? j = i?J ? ? i,j ? Weights depend on transition probabilities from non-similar states. end for Sample j * from P(j * ) where P(j * = j) = ? j /( j ? j ) ? j * is the redundant state to keep. Update ?j = 0.1 * ? j for all j ? J where j ? = j * ? Influence transition prior. Update xt = j * for all {t : xt ? J} ? Influence data used for inference. Remove I j from I for all j ? (J ? i). ? Prevent merging these states in future iterations. end while Output final ? and (x t ) (m) ? These will be used in next iteration of Gibbs sampling. V. # Simulation Study In this section, simulations are used to demonstrate the advantages of the proposed rHDP-HSMM method. The robustness and modeling accuracy is compared with the existing HDP-HSMM method. The simulation is designed as follows. For each simulation, a sequence of observed data is generated with 30 total change points based on the distributions and parameters in Table 2. The emission parameters were specifically selected as they feature some small overlap between their distributions. The generated sequence begins with a state being randomly selected from the three listed in Table 2. A length of duration is sampled from the selected state's duration distribution, which determines how many samples to draw from that state's emission distribution. Once the emission samples are collected, they are stored in the sequence, and the next state is sampled according the to that state's transition probability. The process is repeated 30 times to create a simulated sequence of "observed" data. An example of a simulated dataset can be observed in the Figure 3. In each simulation, both the HDP-HSMM and the rHDP-HSMM are trained on the observed data with the same initial distributions and priors. The prior distributional forms were selected as to allow models to make use of conjugate relationships. Their parameters were selected as to ensure the true distributional parameters could be inferred with high probability. Each simulation's initial parameter values for the HDP-HSMM and rHDP-HSMM were drawn according to the selected prior. The maximum number of states for both models was set to 20. Each state's initial emission distribution was assumed Normal(µ, ? 2 ). The mean's prior distribution was set to µ ? Normal(µ 0 = 0, ? 2 0 = 4). The variance's prior distribution was set to A models had identifiability constraints implemented such as to order their states in increasing order of the posterior mean of their emission distribution. Furthermore, both models performed their respective Gibbs procedure over a maximum of 10000 iterations, or until their Gelman-Rubin statistic [31] reached less than 1.1. The burn-in period for both models was set to 100 iterations. Every 5th iteration of the sampled parameter chains was collected as to remove autocorrelation (resulting in a chain of 2000 length if convergence was not met). The rHDP-HSMM threshold for removing redundant states was set to 1.5. The posterior parameter values for each state was calculated as the mean of the most recent 20% of samples collected from the posterior parameter chains. The posterior sequence was selected to be the mode of the most recent 20% of samples collected from the posterior state sequence. The results of a single simulation are shown in Figures 4,5, and 6. Figure 4 compares the The simulation is repeated 100 times, and the results are shared in Figure 7 and Table 3. Looking at the number of estimated states between the HDP-HSMM and the rHDP-HSMM, it is clear that the rHDP-HSMM's inference procedure removes states that would be otherwise present in a standard HDP-HSMM (Figure 7). In fact, 80 of the 100 simulations resulted in the rHDP-HSMM correctly inferring the true number of states. Furthermore, Table 3 shows that the rHDP-HSMM converged on average with fewer iterations than the HDP-HSMM. This table also shows that while both models are able to correctly capture all the true change points, the standard HDP-HSMM tends to estimate many more change points than the rHDP-HSMM. This is due to the redundancy issue, which the rHDP-HSMM eliminates through its modified inference procedure. The benefit of the proposed rHDP-HSMM is demonstrated via the real-world application of modeling vehicle driving maneuver patterns. This type of modeling is useful for the development intelligent driving assistant systems and autonomous driving vehicles. The dataset analyzed in this study was collected by University of Michigan's Transportation Research Institute [32]. Several kinematic driving signals were collected from human-driven vehicles during their everyday activities. This naturalistic dataset is rich with information related to discover common driving maneuvers and behaviors. [1]. Signals are recorded on trip by trip basis, which begins when the vehicle is turned-on and ends when the vehicle is turned-off. An example of a trip can be seen in Figure 8. The kinematic signals of interest are acceleration, lane offset, and yaw rate. Acceleration and lane offset reflect a driver's intention of moving in the longitudinal and lateral directions respectively. Yaw rate captures a driver's intention of of changing the forward direction of the car. Together, they form a multivariate time-series sampled at 10 Hz which should be highly correlated with human-driving behaviors. An example of the collected signals is The colors in Figure 8 represent the labeling results after training the 0.5 threshold rHDP-HSMM. Noticeably, the rHDP-HSMM segments the road into 9 states. Looking deeper at Figure 8b, it is clear that each state is primarily dictated by changes in yaw rate. Hence this model is able to capture portions of the road where various turning maneuvers are intended by the driver (Figure 8a). Comparing Figure 8a with the HDP-HSMM segmentation shown in Figure 1a, it is clear how the rHDP-HSMM merged the HDP-HSMM's 17 states into a more clear representation of maneuvers used on the road. The rHDP-HSMM and HDP-HSMM are further compared in Figure 9 by using states obtained from the curved portion of the road marked in Figure 8a. Six other trips existed where the same driver drove on that part of the road. Hence, both the HDP-HSMM and the rHDP-HSMM are trained again on each of the other trips under the same initial parameters. The learned states from each model which occurred on the marked portion are analyzed in Figure 9. Figures 9a and9b shows the emission means and durations learned by the HDP-HSMM and the rHDP-HSMM respectively. Interestingly, Figure 9b shows how the rHDP-HSMM concentrates the emission means in various quadrants of the graph. These quadrants relay a positive yaw rate, a negative lane offset, and a positive acceleration in all the learnt means. The concentration of these means in each quadrant indicate a consistency in maneuvers among the various trips, which translates to a left turning action intended by the driver. This same conclusion is not easily recognizable in Figure 9a, as the HDP-HSMM loses this consistency in the learnt means. The difference in learning procedure between the HDP-HSMM and the rHDP-HSMM suggests that the HDP-HSMM's lack of concentrated means derives from the HDP-HSMM overestimating the number of states. As the rHDP-HSMM inference procedure merges similar states together, the emission means of each state can be inferred with a greater amount of data, providing both more consistent estimates and more consistent conclusions. The HDP-HSMM is a powerful model for discovering driving maneuver patterns from kinematic driving data. This paper details an extension to the HDP-HSMM in which this paper refers to as a robust HDP-HSMM (rHDP-HSMM). This model provides a solution to the inconsistency problem caused by the HDP prior. Looking through the lens of a weak-limit 9a shows the means from the original HDP-HSMM, while Figure 9b shows the means from the proposed rHDP-HSMM # VII. Discussion and Conclusion approximation of the HDP prior, the problem typically occurs as the Dirichlet distribution takes no consideration for redundant states, which dilutes the underlying transition process. The rHDP-HSMM solves this issue by adjusting the sample from Dirichlet distribution by checking which states can be merged together. The model then scales down the weights which encourage transitions to redundant states. As a result, the rHDP-HSMM learns fewer redundant states and estimates longer state durations when compared to the original HDP-HSMM. This change leads to improved segmentation and more accurate transition probability representation, which is useful for the application of learning driving maneuvers. Two case studies are presented to further demonstrate the ability of the proposed rHDP-HSMM over the HDP-HSMM. The first study is a simulation which utilizes 1-dimensional normal distributions for the emission function. The rHDP-HSMM demonstrates a clear improvement with regards to the posterior chains. The emission parameters converge much faster, the duration posteriors have far less variance than the HDP-HSMM's duration posterior, and finally the posterior state sequence presents far less change points than the HDP-HSMM's. Over the course of 100 simulations, the rHDP-HSMM out performs the HDP-HSMM in terms of convergence and having less extra change points relative to the truth. The second study demonstrates of the effectiveness of the model in identifying and inferring driving maneuver patterns from a naturalistic dataset of kinematic signals. It is shown how the rHDP-HSMM's merging procedure reduces the number of states to describe a trip from 17 to 9 states when compared to a regular HDP-HSMM. The states are highly interpretable and now specifically capture portions of the road where various turning maneuvers are intended by the driver. In addition to this, the study also compares the results from multiple trips occurring on a curved portion of the road. The results show how the rHDP-HSMM consistently estimates similar emission distributions from multiple trips when compared to the original HDP-HSMM estimates. In both studies, the rHDP-HSMM outperforms the HDP-HSMM in terms of estimation and consistency. This paper concludes that the rHDP-HSMM is worth applying to datasets where an HDP prior may be generating redundant states. Further inspection as to how to select the threshold may be required, however it is clear that the merging procedure within the model is still able to learn consistent and highly interpretable states for the study of driving maneuvers. # Highlights ? A robust HDP-HSMM is proposed which produces more consistent results than the HDP-HSMM ? An algorithm is described as to combat the inconsistency issues that arise from using an HDP prior ? A simulation study is performed to show the impact of the proposed robust HDP-HSMM versus the basic HDP-HSMM in terms of parameter convergence and data segmentation ? Real kinematic data is used to further compare robust HDP-HSMM and the basic HDP-HSMM in terms of learned maneuver patterns. 1![Figure 1: An Example Trip and the Kinematic Signals Belonging to it. Learned States from an HDP-HSMM are Color Coded as Labels](image-2.png "Figure 1 :") 2![Figure 2: A comparison between the structure of a Hidden Markov Model (HMM) and a Hidden Semi-Markov Model (HSMM). The variables and their descriptions are as follows: x t (hidden state at time t), y t (observed data at time t), ? x (transition probabilities of state x), f(? x ) (probability distribution of state x), z s (state of segment s), D s (state uration of segment s) (a) HMM (b) HSMM](image-3.png "Figure 2 :") 2![InvGamma(a 0 = 2, b 0 = 2). The initial duration distributions were assumed Year 2023 Global Journal of Researches in Engineering ( ) Volume Xx XIII Issue I Version I B Automatic Identification of Driving Maneuver Patterns using a Robust Hidden Semi-Markov Models](image-4.png "? 2 ?") 4![Figure 4: HDP-HSMM Versus rHDP-HSMM Emission Convergence on Simulated Data HDP-HSMM and rHDP-HSMM's emission distribution convergence. The states shown in the plots are the states appearing in the final learned state sequence for each model. Each state is indicated by a different color. The true parameters are indicated by the dashed lines. While both models' posteriors are concentrated around the true parameters, the HDP-HSMM's posterior is multimodal for many states.Figure4ashows how the many states rapidly switch which true state they want to encapture across sampling iterations. With regards to the duration, Figure5displays how both models posteriors are concentrated near the true duration. However, the variance of the HDP-HSMM's posterior samples is far larger than the variance of the rHDP-HSMM. This could be due to a large variation of samples being allocated to each state in the HDP-HSMM. To see this, Figure6shows the posterior state sequence estimated by both models. While the rHDP-HSMM distributes samples to](image-5.png "Figure 4 :") 5![Figure 5: HDP-HSMM Versus rHDP-HSMM Duration Convergence on Simulated Data](image-6.png "Figure 5 :") 6![Figure 6: HDP-HSMM Versus rHDP-HSMM Labeling of Simulated Data](image-7.png "Figure 6 :") 57![Figure 7: The Number of Estimated States from 100 Simulations Comparing both a HDP-HSMM and the rHDP-HSMM](image-8.png "5 Figure 7 :") 8![As maneuvers are expected to switch at a low frequency, the original data is down-sampled to 1 Hz by averaging every 10 data points.Both the HDP-HSMM and a 0.5 threshold rHDP-HSMM are applied to trip shown in Figure8under the following setup. A 3-dimensional multivariate Gaussian distribution is used for the emission distribution (Y ? MVN(µ, ?)). The priors to the emission mean and variance are selected asµ ? MVN([0, 0, 0], [[1, 0, 0], [0, 1, 0], [0, 0, 1]]) ? ? Inverse-Wishart(2, [[1, 0, 0], [0, 1, 0], [0, 0, 1]]).Each state's duration is assumed Poisson distributed (D ? Poisson(?)) with the prior ? ? Gamma(a = 1, b = 7). The identifiability constraints are constructed as to arrange the states in the order of smallest to largest mean and duration. The maximum number of states was limited to 20. The kinematic signals are normalized with respect to the signals observed during the trip. The learned emission means are transformed back to original space once the training is complete for analysis purposes.](image-9.png "8 .") 8![Figure 8: A Different Segmentation of the Road Shown in Figure 1 Labeled by a rHDP-HSMM under a Threshold of 0.5.](image-10.png "Figure 8 :") 9![Figure 9: Emission means corresponding to the kinematic signals from 7 different trips occurring on the curved portion of road shown in 8.Figure9ashows the means from the original HDP-HSMM, while Figure9bshows the means from the proposed rHDP-HSMM](image-11.png "Figure 9 :") ![Figure 9: Emission means corresponding to the kinematic signals from 7 different trips occurring on the curved portion of road shown in 8.Figure9ashows the means from the original HDP-HSMM, while Figure9bshows the means from the proposed rHDP-HSMM](image-12.png "Figure") 1 of Various HMM-based Models Versus Our Proposed Robust HDP-HSMM (rHDP-HSMM) III. Problem Formulation a) Data Description ( ) Volume Xx XIII Issue I V ersion I BJournal of Researches in EngineeringGlobal such that the elements within the s-th segment, denoted by y t 1 s :t 2 s , are independent and identically distributed (i.i.d.) for a state duration of D s ? 1, 2, . . . , S. 2EmissionDurationTransitionDistributionNormalPoissonN/AParameter(s) Mean VarianceRateState 1 State 2 State 3State 141600.30.7State 20160.800.2State 3-4160.40.60 3 © 2023 Global Journ als als © 2023 Global Journals * Trafficnet: An open naturalistic driving scenario library DZhao YGuo YJJia IEEE 20th International Conference on Intelligent Transportation Systems (ITSC) IEEE 2017. 2017 * Evaluation of driver car following behavior models for cooperative adaptive cruise control systems MRahman MChowdhury KDey MRIslam TKhan Transportation Research Record 2622 2017 * Stochastic mpc with learning for driver-predictive vehicle control and its application to hev energy management SDi Cairano DBernardini ABemporad IVKolmanovsky IEEE Transactions on Control Systems Technology 22 2013 * FSagberg GFSelpi JBianchi Piccinini Engstr Öm A review of research on driving styles and road safety 2015 57 * Driving style recognition for intelligent vehicle control and advanced driver assistance: A survey CMMartinez MHeucke F.-YWang BGao DCao IEEE Transactions on Intelligent Transportation Systems 19 2017 * Driving style analysis using primitive driving patterns with bayesian nonparametric approaches WWang JXi DZhao IEEE Transactions on Intelligent Transportation Systems 20 2018 * A sticky hdp-hmm with application to speaker diarization EBFox EBSudderth MIJordan ASWillsky The Annals of Applied Statistics 2011 * Multimodal Technologies for Perception of Humans CWooters MHuijbregts 2007 Springer The icsi rt07s speaker diarization system * Redd: A public data set for energy disaggregation research JZKolter MJJohnson Workshop on data mining applications in sustainability (SIGKDD) San Diego, CA 2011 25 * Unsupervised disaggregation of low frequency power measurements HKim MMarwah MArlitt GLyon JHan Proceedings of the 2011 SIAM international conference on data mining the 2011 SIAM international conference on data mining 2011 * Labeling self-tracked menstrual health records with semi-markov models LSymul SPHolmes medRxiv 2021 * Multi-centroid diastolic duration distribution based hsmm for heart sound segmentation APKamson LSharma SDandapat Biomedical signal processing and control 48 2019 * Time series analysis and its applications: with R examples RHShumway DSStoffer 2017 Springer * An introduction to hidden markov models LRabiner BJuang ieee assp magazine 3 1986 * Bayesian nonparametric hidden semi-markov models MJJohnson ASWillsky Journal of Machine Learning Research 14 2013 * MJBeal ZGhahramani CERasmussen Advances in neural information processing systems 2002 The infinite hidden markov model * Hierarchical dirichlet processes YWTeh MIJordan MJBeal DMBlei 10.1198/016214506000000302 doi:10.1198/ 016214506000000302 Journal of the American Statistical Association 101 2006 * YWTeh 2010 Dirichlet process. * Bayesian nonparametric learning: Expressive priors for intelligent systems, Heuristics, probability and causality: A tribute to MIJordan Judea Pearl 11 2010 * JWMiller MTHarrison arXiv:1301.2708 A simple example of dirichlet process mixture inconsistency for the number of components 2013 arXiv preprint * A gentle introduction to the dirichlet process, the beta process, and bayesian nonparametrics MIJordan YWTeh Dept. Statistics 2014 UC Berkeley * About the posterior distribution in hidden markov models with unknown number of states EGassiat JRousseau Bernoulli 20 2014 * An hdp-hmm for systems with state persistence EBFox EBSudderth MIJordan ASWillsky Proceedings of the 25th international conference on Machine learning the 25th international conference on Machine learning ACM 2008 * AGuha NHo XNguyen arXiv:1901.05078 On posterior contraction of parameters and interpretability in bayesian mixture modeling 2019 arXiv preprint * A constructive definition of dirichlet priors JSethuraman Statistica sinica 1994 * AGelman JBCarlin HSStern DBDunson AVehtari DBRubin Bayesian data analysis CRC press 2013 * Markov chain monte carlo methods and the label switching problem in bayesian mixture modeling AJasra CCHolmes DAStephens Statistical Science 20 2005 * Probabilistic relabelling strategies for the label switching problem in bayesian mixture models MSperrin TJaki EWit Statistics and Computing 20 2010 * KPMurphy Hidden semi-markov models (hsmms) 2002 * MJJohnson ASWillsky arXiv:1208.6537 Dirichlet posterior sampling with truncated multinomial likelihoods 2012 arXiv preprint * A single series from the gibbs sampler provides a false sense of security AGelman DBRubin Bayesian statistics 4 1992 * Integrated vehicle-based safety systems (IVBSS) light vehicle field operational test independent evaluation ENodine ALam SStevens MRazo WNajm 2011 United States. National Highway Traffic Safety Administration Technical Report