Exploring Word Representations on Time Expression Recognition
Time expression extraction has attracted longstanding interest over time, due to its great importance in many downstream Natural Language Processing (NLP) and Information Retrieval (IR) tasks. Although current approaches, either rule-based or learning-based, can achieve impressive performance in major datasets, they usually rely heavily on handcrafted rules or task-specific pre-tagging features. Recent advances in pretrained word representations motivate us to explore semi-supervised approaches for this task. We first show that simple neural architectures built on top of pre-trained word representations perform competitively and efficiently on time expression recognition. Then we further explore several design choices focusing on the need of contextualization and the training resource requirements for this type of time expression taggers.