IST-IS-MT1751100: Tatsuya Hiraoka
Stochastic Tokenization with Language Model for Neural Text Classification
Tatsuya Hiraoka (1751100)
(Japanese)
In unsegmented languages such as Japanese and Chinese, tokenization of a sentence has a significant impact on the performance of text classification.
Sentences are usually segmented with words or subwords by a morphological analyzer or byte pair encoding, and then encoded with word (or subword) representations for neural networks.
However, the segmentation is potentially ambiguous and it is unclear whether the segmented tokens achieve the best performance for the target task.
In this paper, we propose a method to simultaneously learn tokenization and text classification to address these problems.
Our model incorporates a language model for unsupervised tokenization into a text classifier, and then trains both models simultaneously.
To make the model robust against infrequent tokens, we sampled segmentation for each sentence stochastically during training, which resulted in improved performance of text classification.
We conducted experiments on sentiment analysis as a text classification task and show that our method achieves better performance than previous methods.