A Hybrid Approach to Word Segmentation of Vietnamese Texts - Inria - Institut national de recherche en sciences et technologies du numérique Accéder directement au contenu
Communication Dans Un Congrès Année : 2008

A Hybrid Approach to Word Segmentation of Vietnamese Texts

Résumé

We present in this article a hybrid approach to automatically tokenize Vietnamese text. The approach combines both finite-state automata technique, regular expression parsing and the maximal-matching strategy which is augmented by statistical methods to resolve ambiguities of segmentation. The Vietnamese lexicon in use is compactly represented by a minimal finite-state automaton. A text to be tokenized is first parsed into lexical phrases and other patterns using pre-defined regular expressions. The automaton is then deployed to build linear graphs corresponding to the phrases to be segmented. The application of a maximal- matching strategy on a graph results in all candidate segmentations of a phrase. It is the responsibility of an ambiguity resolver, which uses a smoothed bigram language model, to choose the most probable segmentation of the phrase. The hybrid approach is implemented to create vnTokenizer, a highly accurate tokenizer for Vietnamese texts.
Fichier principal
Vignette du fichier
LATA039.pdf (132.92 Ko) Télécharger le fichier
Origine : Fichiers éditeurs autorisés sur une archive ouverte
Loading...

Dates et versions

inria-00334761 , version 1 (27-10-2008)

Identifiants

Citer

Hong Phuong Le, Thi Minh Huyen Nguyen, Azim Roussanaly, Tuong Vinh Ho. A Hybrid Approach to Word Segmentation of Vietnamese Texts. 2nd International Conference on Language and Automata Theory and Applications - LATA 2008, Mar 2008, Tarragona, Spain. pp.240-249, ⟨10.1007/978-3-540-88282-4_23⟩. ⟨inria-00334761⟩
251 Consultations
2765 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More