BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//Department of Computer Science - ECPv6.15.20//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:Department of Computer Science
X-ORIGINAL-URL:https://www.cs.jhu.edu
X-WR-CALDESC:Events for Department of Computer Science
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20180311T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20181104T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20190310T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20191103T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20200308T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20201101T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20191016T090000
DTEND;TZID=America/New_York:20191016T120000
DTSTAMP:20260422T162900
CREATED:20210629T210720Z
LAST-MODIFIED:20210629T210720Z
UID:1962340-1571216400-1571227200@www.cs.jhu.edu
SUMMARY:Computer Science Student Defense: Dingquan Wang\, Johns Hopkins University – “Supervised Training on Synthetic Languages: A Novel Framework for Unsupervised Parsing”
DESCRIPTION:LocationMalone 107AbstractThis talk focuses on unsupervised dependency parsing—parsing sentences of a language into dependency trees without accessing the training data of that language. Different from most prior work that uses unsupervised learning to estimate the parsing parameters\, we estimate the parameters by supervised training on synthetic languages. Our parsing framework has three major components: Synthetic language generation gives a rich set of training languages by mix-and-match over the real languages; surface-form feature extraction maps an unparsed corpus of a language into a fixed-length vector as the syntactic signature of that language; and\, finally\, language-agnostic parsing incorporates the syntactic signature during parsing so that the decision on each word token is reliant upon the general syntax of the target language.The fundamental question we are trying to answer is whether some useful information about the syntax of a language could be inferred from its surface-form evidence (unparsed corpus). This is the same question that has been implicitly asked by previous papers on unsupervised parsing\, which only assumes an unparsed corpus to be available for the target language. We show that\, indeed\, useful features of the target language can be extracted automatically from an unparsed corpus\, which consists only of gold part-of-speech (POS) sequences. Providing these features to our neural parser enables it to parse sequences like those in the corpus. Strikingly\, our system has no supervision in the target language. Rather\, it is a multilingual system that is trained end-to-end on a variety of other languages\, so it learns a feature extractor that works well. We show experimentally across multiple languages: (1) Features computed from the unparsed corpus improve parsing accuracy. (2) Including thousands of synthetic languages in the training yields further improvement. (3) Despite being computed from unparsed corpora\, our learned task-specific features beat previous works’ interpretable typological features that require parsed corpora or expert categorization of the languageBioDingquan Wang is a Ph.D. student working with Jason Eisner since 2014. His research interest is natural language processing (NLP) for low-resource languages. He received M.S. in Computer Science from Columbia University advised by Michael Collins and Rebecca Passonneau\, and B.Eng from ACM Honored Class in Computer Science from Shanghai Jiao Tong University.HostJason Eisner
URL:https://www.cs.jhu.edu/event/computer-science-student-defense-dingquan-wang-johns-hopkins-university-supervised-training-on-synthetic-languages-a-novel-framework-for-unsupervised-parsing/
END:VEVENT
END:VCALENDAR