There is abundant knowledge out there carried in the form of natural language texts, such as social media posts, scientific research literature, medical records, etc., which grows at an astonishing rate. Yet this knowledge is mostly inaccessible to computers and overwhelming for human experts to absorb. Information extraction (IE) processes raw texts to produce machine understandable structured information, thus dramatically increasing the accessibility of knowledge through search engines, interactive AI agents, and medical research tools. However, traditional IE systems assume abundant human annotations for training high quality machine learning models, which is impractical when trying to deploy IE systems to a broad range of domains, settings and languages. In this talk, I will present how to use deep representation learning to leverage the distributional statistics of characters and words, the annotations for other tasks and other domains, and the linguistic and problem structures, to combat the problem of inadequate supervision, and conduct information extraction with scarce human annotations.
Nanyun (Violet) Peng
CoRE A 301
Ahmed Elgammal and Dimitri Metaxas
Faculty Candidate Talk
John Hopkins University