-
ACL 20232023Many machine learning-based low-code or no-code applications involve generating code that interacts with structured knowledge. For example, one of the most studied tasks in this area is generating SQL code from a natural language statement. Prior work shows that incorporating context information from the database schema, such as table and column names, is beneficial to model performance on this task. In
-
ACL 20232023Paraphrase generation is a long-standing task in natural language processing (NLP). Supervised paraphrase generation models, which rely on human-annotated paraphrase pairs, are cost-inefficient and hard to scale up. On the other hand, automatically annotated paraphrase pairs (e.g., by machine back-translation), usually suffer from the lack of syntactic diversity — the generated paraphrase sentences are
-
ACL Findings 20232023Understanding users’ intentions in e-commerce platforms requires commonsense knowledge. In this paper, we present FolkScope, an intention knowledge graph construction framework to reveal the structure of humans’ minds about purchasing items. As commonsense knowledge is usually ineffable and not expressed explicitly, it is challenging to perform information extraction. Thus, we propose a new approach that
-
ACL 20232023Identifying granular and actionable topics from customer questions (CQ) posted on e-commerce websites helps surface the missing information on the product detail page expected by customers before making a purchase. Insights on missing information on product page helps brands and sellers enrich the catalog quality to improve the overall customer experience (CX). In this paper, we propose a weakly supervised
-
ACL 20232023We present EPIC (English Perspectivist Irony Corpus), the first annotated corpus for irony analysis based on the principles of data perspectivism. The corpus contains short conversations from social media in five regional varieties of English, and it is annotated by contributors from five countries corresponding to those varieties. We analyse the resource along the perspectives induced by the diversity
Related content
-
September 10, 2021Data augmentation makes examples more realistic, while continual-learning techniques prevent “catastrophic forgetting”.
-
September 09, 2021Model using ASR hypotheses as extra inputs reduces word error rate of human transcriptions by almost 11%.
-
September 02, 2021Branching encoder networks make operation more efficient, while “neural diffing” reduces bandwidth requirements for model updates.
-
August 27, 2021Liu discusses her work in speech recognition and understanding, prosody modeling, summarization, and natural language processing.
-
August 27, 2021New voice for Alexa’s Reading Sidekick feature avoids the instabilities common to models with variable prosody.
-
August 25, 2021Katrin Kirchhoff, director of speech processing for Amazon Web Services, on the many scientific challenges her teams are tackling.