Skip to Main content Skip to Navigation
Conference papers

Domain adaptation for sequence labeling using hidden Markov models

Edouard Grave 1, 2 Guillaume Obozinski 3 Francis Bach 1, 2
2 SIERRA - Statistical Machine Learning and Parsimony
DI-ENS - Département d'informatique de l'École normale supérieure, Inria Paris-Rocquencourt, CNRS - Centre National de la Recherche Scientifique : UMR8548
Abstract : Most natural language processing systems based on machine learning are not robust to domain shift. For example, a state-of-the-art syntactic dependency parser trained on Wall Street Journal sentences has an absolute drop in performance of more than ten points when tested on textual data from the Web. An efficient solution to make these methods more robust to domain shift is to first learn a word representation using large amounts of unlabeled data from both domains, and then use this representation as features in a supervised learning algorithm. In this paper, we propose to use hidden Markov models to learn word representations for part-of-speech tagging. In particular, we study the influence of using data from the source, the target or both domains to learn the representation and the different ways to represent words using an HMM.
Complete list of metadata

Cited literature [16 references]  Display  Hide  Download

https://hal.inria.fr/hal-00918371
Contributor : Edouard Grave <>
Submitted on : Friday, December 13, 2013 - 2:11:27 PM
Last modification on : Friday, July 23, 2021 - 2:35:25 PM
Long-term archiving on: : Tuesday, March 18, 2014 - 12:35:35 PM

Files

da.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-00918371, version 1
  • ARXIV : 1312.4092

Citation

Edouard Grave, Guillaume Obozinski, Francis Bach. Domain adaptation for sequence labeling using hidden Markov models. New Directions in Transfer and Multi-Task: Learning Across Domains and Tasks (NIPS Workshop), Dec 2013, Lake Tahoe, United States. ⟨hal-00918371⟩

Share

Metrics

Record views

877

Files downloads

313