-
ACL Findings 20232023Self-rationalizing models that also generate a free-text explanation for their predicted labels are an important tool to build trustworthy AI applications. Since generating explanations for annotated labels is a laborious and costly process, recent models rely on large pretrained language models (PLMs) as their backbone and few-shot learning. In this work we explore a self-training approach leveraging both
-
ACL 20232023Despite the popularity of Shapley Values in explaining neural text classification models, computing them is prohibitive for large pretrained models due to a large number of model evaluations. In practice, Shapley Values are often estimated with a small number of stochastic model evaluations. However, we show that the estimated Shapley Values are sensitive to random seed choices — the top-ranked features
-
ACL 20232023NLP models often degrade in performance when real world data distributions differ markedly from training data. However, existing dataset drift metrics in NLP have generally not considered specific dimensions of linguistic drift that affect model performance, and they have not been validated in their ability to predict model performance at the individual example level, where such metrics are often used in
-
ACL 20232023Off-Policy reinforcement learning has been a driving force for the state-of-the-art conversational AIs leading to more natural humanagent interactions and improving the user satisfaction for goal-oriented agents. However, in large-scale commercial settings, it is often challenging to balance between policy improvements and experience continuity on the broad spectrum of applications handled by such system
-
ACL Findings 20232023Existing efforts on text synthesis for codeswitching mostly require training on codeswitched texts in the target language pairs, limiting the deployment of the models to cases lacking code-switched data. In this work, we study the problem of synthesizing codeswitched texts for language pairs absent from the training data. We introduce GLOSS, a model built on top of a pre-trained multilingual machine translation
Related content
-
February 03, 2021Neural text-to-speech enables new multilingual model to use the same voice for Spanish and English responses.
-
January 26, 2021Sneha Rajana is an applied scientist at Amazon today, but she didn't start out that way. Learn how she made the switch, and the advice she has for others considering a similar change.
-
January 25, 2021New approach to few-shot learning improves on state of the art by combining prototypical networks with data augmentation.
-
January 21, 2021Amazon principal applied scientist Yang Liu on the frontiers of speech and dialogue.
-
January 13, 2021In experiments, multilingual models outperform monolingual models.
-
December 18, 2020Researchers propose a method to automatically generate training data for Alexa by identifying cases in which customers rephrase unsuccessful requests.