-
AAAI 20242023Toxic content detection is crucial for online services to remove inappropriate content that violates community standards. To automate the detection process, prior works have proposed varieties of machine learning (ML) approaches to train Language Models (LMs) for toxic content detection. However, both their accuracy and transferability across datasets are limited. Recently, Large Language Models (LLMs)
-
2023 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU)2023In this paper, we propose the first successful implementation of associated learning (AL) to automatic speech recognition (ASR). AL has been shown to provide better label noise robustness, faster training convergence, and flexibility on model complexity than back-propagation (BP) in classification tasks. However, extending the learning approach to autoregressive models such as ASR, where model outputs are
-
EMNLP 20232023Generating concise summaries of news events is a challenging natural language processing task. While journalists often curate timelines to highlight key sub-events, newcomers to a news event face challenges in catching up on its historical context. In this paper, we address this need by introducing the task of background news summarization, which complements each timeline update with a background summary
-
EMNLP 20232023Large Language Models (LLMs) have gained significant popularity for their impressive performance across diverse fields. However, LLMs are prone to hallucinate untruthful or nonsensical outputs that fail to meet user expectations in many real-world applications. Existing works for detecting hallucinations in LLMs either rely on external knowledge for reference retrieval or require sampling multiple responses
-
ACL Findings 20232023While impressive performance has been achieved on the task of Answer Sentence Selection (AS2) for English, the same does not hold for languages that lack large labeled datasets. In this work, we propose Cross-Lingual Knowledge Distillation (CLKD) from a strong English AS2 teacher as a method to train AS2 models for low-resource languages in the tasks without the need of labeled data for the target language
Related content
-
March 11, 2021Watch a recording of the presentation and Q&A roundtable featuring Amazon scientists and scholars.
-
March 11, 2021University teams will compete in building agents that can help customers complete complex tasks, like cooking and home improvement. Deadline for university team applications is April 16.
-
March 02, 2021The newest chapter addresses a problem that often bedevils nonparametric machine learning models.
-
March 01, 2021The Art Museum skill uses Alexa Conversations, an AI-driven dialogue management tool.
-
February 08, 2021Technique that relies on inverse reinforcement learning, or learning by example, improves task completion rate by 14% to 17% in simulations.
-
February 08, 2021Yanagisawa discusses the science behind Alexa's new bilingual Polyglot model, her career in speech research, and more.