Monash University
Browse

Weak Supervision and Active Learning for Natural Language Processing

Download (6.76 MB)
thesis
posted on 2019-07-02, 01:39 authored by MING LIU
Deep neural models in natural language processing rely on large amounts of labeled data. In the real world, annotation can be expensive and time consuming. In this project, the aim is to learn good deep neural models with a minimum amount of labeled data. We take multiple strategies to achieve this goal: we use structured information from the data, incorporate prior knowledge for the model and learn an active learning policy for annotation. Experiments on simulated and real word tasks show these strategies are useful and effective.

History

Campus location

Australia

Principal supervisor

Wray Buntine

Additional supervisor 1

Gholamreza Haffari

Year of Award

2019

Department, School or Centre

Information Technology (Monash University Clayton)

Course

Doctor of Philosophy

Degree Type

DOCTORATE

Faculty

Faculty of Information Technology