0%

Unsupervised PoS Tagging

本文主要介绍无监督贝叶斯pos tagging方法,实验结果和分析。简单概括最大似然估计(MLE),隐藏马尔可夫模型(HMM)。

原文参考 A Fully Bayesian Approach to Unsupervised Part-of-Speech Tagging (Sharon Goldwater, Thomas L. Griffiths, 2016)

还是懒癌,暂时是英文的

Intro

Inference for HMMs

  • Notation: Given that $t_{i-1} = t$, the value of $t_i$ is drawn from a multinominal dstribution with parameters $\tau^{(t)}$
  • Inference: decoding, applying the model at the test time, we need to know $\mathbf \theta$ and we can compute $P(\mathbf t, \mathbf w)$
  • Compute $P(\mathbf w)$, such as language model
  • Also $P(\mathbf t \mid \mathbf w) $, such as PoS tagger

Parameter Estimation for HMMs

  • Estimation: training the model, determing its params. A procedure to set $\mathbf \theta$ based on data.
  • Bayes Rule
  • Could use MLE, Bayesian estimation (actually no parameter estimation)

Maximum Likelihood Estimation

  • Choose the $\mathbf \theta$ that makes the data most probable, ignore the prior term. Equivalent to assuming a uniform prior.
  • supervised systems, \textit{relative frequency estimate} is equivalent to the MLE.
  • unsupervised systems, expectation maximization (EM) algorithm to estimate $\theta$
  • Process:
    • E-step: use current estimate of $\theta$ to compute expected counts of hidden events $n$.
    • M-step: recompute $\theta$ using expected counts.s
  • Examples: forward-backward algorithm for HMMs, inside-outside algorithm for PCFGs, k-meaning clustering.
  • EM Works well on : word alignment for machine translation; speech recognition….
  • Often fails:
    • probabilistic context-free grammars (PCFG): highly sensitive to initialisation; F-score reported are generally low.
    • For HMMs, even very small amounts of training data have been show to work better than EM.
    • similar picture for many other tasks

Main Model

Bayesian HMM

  • Parameter Estimation: we are not interested in the value of $\theta$
  • Bayesian integration: Integrating over $\theta$ gives us an average over all possible parameters values.

    • accounts for uncertainty as to the exact value of $\theta$
    • models the shape of the distribution over $\theta$
    • increase robustness: there may be a range of good values of $\theta$
    • we can use priors favouring sparse solutions.
  • Model:

  • intergrate out the parameters $\mathbf \theta = (\mathbf \tau, \mathbf \omega)$, we calculate probablity for each of $T$ possible tags

  • For inference, $P(\mathbf t \mid \mathbf w)$ using an estimation method called \textit{Gibbs sampling}.

Dirichlet Distribution

  • a distribution over distribution, a prior.
  • Definition
  • $\alpha$ is params. of Dirichlet Distribution, we usually use symmetric Dirichlets, where $\alpha_1 … \alpha_K$ is equal to $\beta$. Denote Dirichlet($\beta$) to mean Dirichlet($\beta,…,\beta)$
  • property:
    • With $\beta >1$, we would \textbf{prefer} uniform distributions over $\theta$
    • With $\beta = 1$, we have no preference on $\theta$, does not mean we choose uniform.
    • With $\beta < 1$, we prefer sparse (skewed) distributions

Evaluation of BHMM

  • Compare with MLHMM and CRF/CE

    • Results

      • Intergrating over parameters is useful in itself, even with uninformative prior ($\alpha = \beta =1$):
      • Better prior can help even more, though do not reach the state of the art.
    • The BHMM indentifies a sequence of tages that have high prob. over a range of parameter values, rather than choosing tags based on the single best set of paprameters (MLHMM).

    • The smaller effect of $\beta$:

      although the true output distribution tend to be sparse as well, the level of sparseness depends on the tag (consider function words vs. content words in particular). Therefore, the value of $\beta$ that accurately reflects the most probable output distributions for some tags may be a poor choice for other tags.

    • The trasition probabilities matrix is sparse: the optimal value of $\alpha$ is 0.003.
  • Sytactic clustering
    • With MLHMM: different tokens of the same word type are usually assigned to the same cluster, but types are assigned to clusters more or less at random, and all clusters have approximately the same number of types.
    • BHMM: the clusters found by BHMM tend to be more coherent and more variable in size
    • BHMM transition matrix is sparse, MLHMM is not.

Summary

  • Unsupervised PoS tagging is useful to build lexica and taggers for new language or domains;
  • maximum likelihood HMM with EM performs poorly; (see MLE)
  • Bayesian HMM with Gibbs sampling can be used instead;
    • no intereted in params, instead of parameter estimation, we do integration.
    • Gibbs Sampling (for inference, see below)
  • the Bayesian HMM improves performance by averaging out uncertainty;
    • MLE assume uniform values over $\mathbf \theta$, with $100\%$ certainty, and then pick one set of $\theta$ which maximum the liklihood.
    • Bayesian HMM does not make that assumption, and if we choose a certain hyperparameter $\alpha$ and $\beta$, we have different certainty of the set of $\theta$.
      • With $\alpha,~ \beta >1$, we would \textbf{prefer} uniform values over $\mathbf \theta$
      • With $\beta = 1$, we have no preference on $\mathbf \theta$, does not mean we choose uniform.
      • With $\beta < 1$, we prefer sparse value of $\mathbf \theta$.
  • it also allows us to use priors that favour sparse solutions as they occur in language data.
  • Other types of discrete latent variable models (e.g. for syntax or semantics) use similar methods.

References

Goldwater, S., & Griffiths, T. (2007). A fully Bayesian approach to unsupervised part-of-speech tagging. In Proceedings of the 45th annual meeting of the association of computational linguistics (pp. 744-751).

Modified 18th-May-2018 12:00