Skip to content Skip to navigation

Faculty Candidate Talk: Representation Learning with Joint Models for Information Extraction


There is abundant knowledge out there carried in the form of natural language texts, such as social media posts, scientific research literature, medical records, etc., which grows at an astonishing rate. Yet this knowledge is mostly inaccessible to computers and overwhelming for human experts to absorb. Information extraction (IE) processes raw texts to produce machine understandable structured information, thus dramatically increasing the accessibility of knowledge through search engines, interactive AI agents, and medical research tools. However, traditional IE systems assume abundant human annotations for training high quality machine learning models, which is impractical when trying to deploy IE systems to a broad range of domains, settings and languages. In this talk, I will present how to use deep representation learning to leverage the distributional statistics of characters and words, the annotations for other tasks and other domains, and the linguistic and problem structures, to combat the problem of inadequate supervision, and conduct information extraction with scarce human annotations.

Nanyun (Violet) Peng
CoRE A 301
Event Date: 
03/21/2017 - 10:30am
Ahmed Elgammal and Dimitri Metaxas
Event Type: 
Faculty Candidate Talk
John Hopkins University