`
单眼皮大娘
  • 浏览: 113060 次
  • 性别: Icon_minigender_2
  • 来自: 上海
社区版块
存档分类
最新评论
阅读更多
一、introduction
      Often we are interested in finding patterns which appear over a space of time. These patterns occur in many areas; the pattern of commands someone uses in instructing a computer, sequences of words in sentences, the sequence of phonemes in spoken words - any area where a sequence of events occurs could produce useful patterns.
      Consider the simple example of someone trying to deduce the weather from a piece of seaweed - folklore tells us that `soggy' seaweed means wet weather, while `dry' seaweed means sun. If it is in an intermediate state (`damp'), then we cannot be sure. However, the state of the weather is not restricted to the state of the seaweed, so we may say on the basis of an examination that the weather is probably raining or sunny. A second useful clue would be the state of the weather on the preceding day (or, at least, its probable state) - by combining knowledge about what happened yesterday with the observed seaweed state, we might come to a better forecast for today.

       This is typical of the type of system we will consider in this tutorial.

       First we will introduce systems which generate probabalistic patterns in time, such as the weather fluctuating between sunny and rainy.

       We then look at systems where what we wish to predict is not what we observe - the underlying system is hidden. In the above example, the observed sequence would be the seaweed and the hidden system would be the actual weather.

       We then look at some problems that can be solved once the system has been modeled. For the above example, we may want to know

       1、What the weather was for a week given each day's seaweed observation.
       2、Given a sequence of seaweed observations, is it winter or summer? Intuitively, if the seaweed has been dry for a while it may be summer, if it has been soggy for a while it might be winter.

二、Deterministic Patterns
 
    Consider a set of traffic lights; the sequence of lights is red - red/amber - green - amber - red. The sequence can be pictured as a state machine, where the different states of the traffic lights follow each other.


 
    Notice that each state is dependent solely on the previous state, so if the lights are green, an amber light will always follow - that is, the system is deterministic. Deterministic systems are relatively easy to understand and analyse, once the transitions are fully known.

三、Non-deterministic patterns

    To make the weather example a little more realistic, introduce a third state - cloudy. Unlike the traffic light example, we cannot expect these three weather states to follow each other deterministically, but we might still hope to model the system that generates a weather pattern.

    One way to do this is to assume that the state of the model depends only upon the previous states of the model. This is called the Markov assumption and simplifies problems greatly. Obviously, this may be a gross simplification and much important information may be lost because of it.

    When considering the weather, the Markov assumption presumes that today's weather can always be predicted solely given knowledge of the weather of the past few days - factors such as wind, air pressure etc. are not considered. In this example, and many others, such assumptions are obviously unrealistic. Nevertheless, since such simplified systems can be subjected to analysis, we often accept the assumption in the knowledge that it may generate information that is not fully accurate. 



    A Markov process is a process which moves from state to state depending (only) on the previous n states. The process is called an order n model where n is the number of states affecting the choice of next state. The simplest Markov process is a first order process, where the choice of state is made purely on the basis of the previous state. Notice this is not the same as a deterministic system, since we expect the choice to be made probabalistically, not deterministically.

     The figure below shows all possible first order transitions between the states of the weather example.
    


     Notice that for a first order process with M states, there are M2 transitions between states since it is possible for any one state to follow another. Associated with each transition is a probability called the state transition probability - this is the probability of moving from one state to another. These M2 probabilities may be collected together in an obvious way into a state transition matrix. Notice that these probabilities do not vary in time - this is an important (if often unrealistic) assumption.

     The state transition matrix below shows possible transition probabilities for the weather example;



     that is, if it was sunny yesterday, there is a probability of 0.5 that it will be sunny today, and 0.375 that it will be cloudy. Notice that (because the numbers are probabilities) the sum of the entries for each row is 1.

     To initialise such a system, we need to state what the weather was (or probably was) on the day after creation; we define this in a vector of initial probabilities, called the TT vector.



- that is, we know it was sunny on day 1.
We have now defined a first order Markov process consisting of :

    states : Three states - sunny, cloudy, rainy.
    vector : Defining the probability of the system being in each of the states at time 0.
     state transition matrix : The probability of the weather given the previous day's weather.
     Any system that can be described in this manner is a Markov process.

四、Patterns generated by a hidden process

     When a Markov process may not be powerful enough

     In some cases the patterns that we wish to find are not described sufficiently by a Markov process. Returning to the weather example, a hermit may perhaps not have access to direct weather observations, but does have a piece of seaweed. Folklore tells us that the state of the seaweed is probabalistically related to the state of the weather - the weather and seaweed states are closely linked. In this case we have two sets of states, the observable states (the state of the seaweed) and the hidden states (the state of the weather). We wish to devise an algorithm for the hermit to forecast weather from the seaweed and the Markov assumption without actually ever seeing the weather.
   
     A more realistic problem is that of recognising speech; the sound that we hear is the product of the vocal chords, size of throat, position of tongue and several other things. Each of these factors interact to produce the sound of a word, and the sounds that a speech recognition system detects are the changing sound generated from the internal physical changes in the person speaking.

     Some speech recognition devices work by considering the internal speech production to be a sequence of hidden states, and the resulting sound to be a sequence of observable states generated by the speech process that at best approximates the true (hidden) states. In both examples it is important to note that the number of states in the hidden process and the number of observable states may be different. In a three state weather system (sunny, cloudy, rainy) it may be possible to observe four grades of seaweed dampness (dry, dryish, damp,soggy); pure speech may be described by (say) 80 phonemes, while a physical speech system may generate a number of distinguishable sounds that is either more or less than 80.

      In such cases the observed sequence of states is probabalistically related to the hidden process. We model such processes using a hidden Markov model where there is an underlying hidden Markov process changing over time, and a set of observable states which are related somehow to the hidden states.

五、Hidden Markov Models

    Definition of a hidden Markov model

    A hidden Markov model (HMM) is a triple (,A,B).

    1、the vector of the initial state probabilities;
2、the state transition matrix;
3、the confusion matrix;

     Each probability in the state transition matrix and in the confusion matrix is time independent - that is, the matrices do not change in time as the system evolves. In practice, this is one of the most unrealistic assumptions of Markov models about real processes.

     注:本文原地址:http://www.comp.leeds.ac.uk/roger/HiddenMarkovModels/html_dev/main.html

     想学习隐马尔可夫的朋友可以去原网址看看,应该会受益匪浅。这是摘一部分,感觉确实不是,仅供参考。

   



  • 大小: 50.2 KB
  • 大小: 2.2 KB
  • 大小: 9.8 KB
  • 大小: 5.6 KB
  • 大小: 519 Bytes
分享到:
评论

相关推荐

    HMM隐马尔可夫模型用于中文分词

    隐马尔可夫模型(Hidden Markov Model,HMM)是一种统计模型,被广泛应用于模式识别、自然语言处理等领域。HMM的核心思想是通过一个可以观察的马尔可夫过程来描述一个隐含的状态序列,其中状态不可直接观察到,但每...

    10.2 基于隐马尔可夫模型(HMM)的孤立字语音识别_隐马尔可夫模型(HMM)的孤立字语音识别_

    在本主题中,我们将深入探讨基于隐马尔可夫模型(HMM)的孤立字语音识别方法,并结合MATLAB程序实现进行讲解。 隐马尔可夫模型(Hidden Markov Model, HMM)是概率统计模型,广泛应用于自然语言处理、生物信息学...

    隐马尔可夫模型的简介以及实例介绍以及三个主要算法

    隐马尔可夫模型(Hidden Markov Model,HMM)是一种广泛应用于模式识别和序列分析的统计模型。自从1870年俄国有机化学家Vladimir V. Markovnikov提出马尔科夫模型以来,该模型在理论研究与实际应用领域都发挥了巨大...

    隐马尔可夫模型及其在自然语言处理中的应用

    隐马尔可夫模型及其在自然语言处理中的应用

    隐马尔可夫模型源代码(matlab)

    隐马尔可夫模型(Hidden Markov Model, HMM)是一种统计建模方法,常用于处理序列数据,如语音识别、自然语言处理、生物信息学等领域。在MATLAB环境中实现HMM,我们可以利用其强大的矩阵运算能力和丰富的工具箱。...

    HMM隐马尔可夫模型MATLAB实现

    隐马尔可夫模型(Hidden Markov Model,简称HMM)是概率统计领域中的一个重要模型,尤其在自然语言处理、语音识别、生物信息学等领域有着广泛的应用。在MATLAB环境中,我们可以利用其强大的数学计算能力和丰富的函数...

    一种基于隐马尔可夫模型的人脸识别方法.pdf

    本文提出了一种基于隐马尔可夫模型的人脸识别方法,该方法利用人脸隐马尔可夫模型的结构特征和Viterbi算法的特点,对特征观察序列进行分割,并使用部分序列对所有隐马尔可夫模型递进地计算最大相似度,同时排除...

    第20章-隐马尔可夫模型

    ### 隐马尔可夫模型详解 #### 一、隐马尔可夫模型概述 隐马尔可夫模型(Hidden Markov Model, HMM)作为一种重要的概率图模型,在序列预测问题中扮演着核心角色。它能够有效地处理一系列数据点间的关系,并且尤其...

    基于隐马尔可夫模型回归HMMR模型的时间序列分割处理matlab仿真+代码仿真操作视频

    2.内容:基于隐马尔可夫模型回归HMMR模型的时间序列分割处理matlab仿真+代码仿真操作视频 3.用处:用于隐马尔可夫模型回归HMMR模型的时间序列分割处理算法编程学习 4.指向人群:本硕博等教研学习使用 5.运行注意...

    连续型隐马尔可夫模型(HMM)参数迭代算法

    ### 连续型隐马尔可夫模型(HMM)参数迭代算法 #### 知识点解析 **一、隐马尔可夫模型(HMM)基础** 隐马尔可夫模型是一种统计模型,用于描述一个含有未知参数的马尔可夫过程。这种模型在自然语言处理、语音识别...

    隐马尔可夫模型c代码

    隐马尔可夫模型(Hidden Markov Model,简称HMM)是概率统计领域的一种重要模型,广泛应用于自然语言处理、语音识别、生物信息学等多个IT领域。本模型的核心思想是,尽管我们无法直接观测到系统的真实状态,但可以...

    隐马尔可夫模型(HMM)简介

    "隐马尔可夫模型(HMM)简介" 隐马尔可夫模型(Hidden Markov Model,HMM)是一种数学模型,用来描述一个系统的隐状态和观察状态之间的关系。在本文中,我们通过一个实例来了解 HMM 的基本概念和应用。 什么是 HMM?...

    隐马尔可夫模型ppt

    此ppt由专业人员编写,内容条例清晰,重点突出,结合了简单易懂的实例,深入浅出的介绍了隐马尔可夫模型。

    隐马尔可夫模型和词性标注笔记

    隐马尔可夫模型(HMM)是一种统计建模方法,尤其在自然语言处理和语音识别领域中广泛应用。它基于马尔可夫模型的概念,但增加了“隐藏”或不可观测的状态,这些状态通过一系列可观察的输出来表现。在HMM中,系统处于...

Global site tag (gtag.js) - Google Analytics