BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//Department of Computer Science - ECPv6.15.20//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:Department of Computer Science
X-ORIGINAL-URL:https://www.cs.jhu.edu
X-WR-CALDESC:Events for Department of Computer Science
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20200308T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20201101T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20210314T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20211107T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20220313T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20221106T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20210112T104500
DTEND;TZID=America/New_York:20210112T120000
DTSTAMP:20260422T161234
CREATED:20210629T210726Z
LAST-MODIFIED:20210629T210726Z
UID:1962550-1610448300-1610452800@www.cs.jhu.edu
SUMMARY:Machine Translation for All: Improving Machine Translation in Low Resource\, Domain Mismatch\, and Low Resource Settings
DESCRIPTION:Title: Machine Translation for All: Improving Machine Translation in Low Resource\, Domain Mismatch\, and Low Resource SettingsPresenter: Huda KhayrallahGraduate Student: Jie Ying WuOnline: https://JHUBlueJays.zoom.us/j/94119734136?pwd=aVllamlFRG1JY3FocEtPemRkZXlvdz09\, ID 941 1973 4136\, password 179445Description: Machine translation uses machine learning to automatically translate text from one language to another and has the potential to reduce language barriers. Recent improvements in machine translation have made it more widely-usable\, partly due to deep neural network approaches. However—like most deep learning algorithms—neural machine translation is sensitive to the quantity and quality of training data\, and therefore produces poor translations for some languages and styles of text. Machine translation training data typically comes in the form of parallel text—sentences translated between the two languages of interest. Limited quantities of parallel text are available for most language pairs\, leading to a low-resource problem.  Even when training data is available in the desired language pair\, it is frequently formal text—leading to a domain mismatch when models are used to translate a different type of data\, such as social media or medical text.  Neural machine translation currently performs poorly in low-resource and domain mismatch settings; my work aims to overcome these limitations\, and make machine translation a useful tool for all users.In this talk\, I will discuss a method for improving translation in low resource settings—Simulated Multiple Reference Training (SMRT; Khayrallah et al.\, 2020)—which uses a paraphraser to simulate training on all possible translations per sentence. I will also discuss work on improving domain adaptation (Khayrallah et al.\, 2018)\, and work on analyzing the effect of noisy training data (Khayrallah and Koehn\, 2018).
URL:https://www.cs.jhu.edu/event/machine-translation-for-all-improving-machine-translation-in-low-resource-domain-mismatch-and-low-resource-settings/
END:VEVENT
END:VCALENDAR