How to produce factually accurate automatic text summaries

New metric can be calculated 55 times as quickly as its state-of-the-art predecessor, making it practical for model training.

Abstractive summarization is the automatic extraction and recombination of phrases from a text in order to summarize that text. Deep-learning-based abstractive-summarization systems are usually trained to maximize the overlap between the summaries they generate and sample summaries in their training data.

The trouble with this approach is that a summary that overlaps significantly with a target summary may recombine phrases in factually inaccurate manner. In the example below, which concerns an upcoming boxing match, the summarization model correctly concludes that “has a chink in his armor” summarizes an important aspect of the input text, but it applies it to the wrong boxer:

Klitschko example.png
Conventional metrics for training abstractive-summarization models don’t account for factual accuracy.

Although abstractive-summarization models have become very good at generating fluent, syntactically correct text, their frequent factual inaccuracy has severely hampered their adoption.

In a paper we presented at this year’s meeting of the Association for Computational Linguistics (ACL), we describe a new metric for measuring the performance of abstractive-summarization models, which accounts for factual accuracy. We also describe a methodology for using our metric to train abstractive-summarization models.

Our metric adopts the same general strategy as the earlier QAGS metric, but it’s 55 times as fast to apply, which makes it more practical for model training.

QAGS-QUALS-Image.png
Our new summary-scoring metric, QUALS (bottom), uses the same strategy as the earlier QAGS (top) but has a simpler architecture, enabling it to generate a score 55 times as quickly.
Credit: Glynis Condon

Using QAGS as an evaluation metric, we compared models trained using our approach to models trained using traditional metrics and methodologies, and we found that our approach improved on the best-performing previous models by 15% on one dataset and by 2% on another.

Scoring through question answering

QAGS (which stands for question answering and generation for summarization) uses a four-step procedure to score a text summary. First, it extracts names and noun phrases from the summary; these are potential answers to potential questions about the summary. 

Second, it feeds each extracted noun, together with the text of the summary, to a trained question generation model, which produces a question whose answer is the noun. Third, it feeds each of the generated questions to a trained question-answering model, once accompanied by the summary and once accompanied by the source text. 

QAGS-Image.cropped.png
QAGS requires the sequential application of three neural models: an answer extraction model, a question-answering model, and a question generation model.
Credit: Glynis Condon

The final score assesses the similarity between the answers based on the source text and the answers based on the summary. The intuition is that if both the summary and the source text cause the question-answering model to answer the questions in the same way, the summary is factually accurate. If they cause different answers, then the summary has probably garbled some facts.

By accounting for factual accuracy, QAGS offers a better assessment of summary quality than metrics based on phrasal overlap. But it requires the sequential application of three different deep-learning networks, which is inefficient.

QUALS

Our approach, which we call QUALS (for question answering with language model score for summarization), reduces the number of models to one, which makes it 55 times as fast as QAGS.

That one model is the joint question-and-answer generation (QAGen) model that members of our group presented at last year’s ACL. It takes a text as input and generates question-and-answer pairs pertaining to it.

QUALS-Image.cropped.png
QUALS requires a single neural model, a question-and-answer generation model.
Credit: Glynis Condon

The output of the QAGen model for a given input can be thought of as a huge tree, in which the nodes are words and each edge encodes the likelihood that a particular word will be followed by another word.

For a given summary, we search the resulting tree to produce 60 high-probability question-and-answer pairs. Our search algorithm ensures that we explore diverse paths through the tree, in order to generate a variety of candidate questions and answers. Then we throw out all the question-answer pairs whose answers are not sequences of words found in the summary.

Next, we feed the source text on which the summary is based to the QAGen model. We use the resulting tree to calculate the probabilities of the same question-answer pairs we extracted for the summary. When, for the source text, the probability of generating a particular question-answer pair is small compared to the probability for the summary, the QUALS will be low. Intuitively, the discrepancy suggests that the question-answer pair was plausible for the summary but not in the source text, indicating factual inconsistency.

QUALS scoring.png
Probabilities per token (words and other standalone symbols) of two different question-answer pairs, based on a summary (blue) and an input document (orange). The large probability differences for the answer in the right-hand example give it a much lower QUALS score (-2.615) than the right-hand example (-0.054).

Training methodology

The QUALS score gives us an efficiently computable measure of a summary’s factual accuracy, but using it to train a machine learning model is not straightforward. Differences in QUALS score can’t simply be back-propagated through the QAGen model to update the summarization model.

So in our paper, we propose contrastive learning as a method for using QUALS to train a summarization model. First, we train a summarization model using the standard approach, which uses maximum-likelihood estimation (MLE) to approximate a phrasal-overlap score.

Next, we use the trained model to generate new summaries for all the source texts in the training data and create two different groups of summaries. One group, S+, contains ground truth summaries that have high QUALS scores (indicating factually accurate summaries); the other, S- contains generated summaries that have low QUALS scores (indicating factually inaccurate summaries).

Finally, we retrain the summarization model, using a loss function that encourages it to generate summaries like those in S+ and discourages it from generating summaries like those in S-.

Evaluation

Sample summaries.png
Examples from the human-evaluation study, featuring input texts and summaries produced using both MLE and the ConSeq model, which is trained using QUALS.

As baselines for the evaluation of our approach, we used two models. One was trained using MLE in the standard way, to fine-tune a BART language model. For the other, we used our contrastive-learning methodology, but instead of using QUALS to evaluate summaries, we used an ensemble of three ROUGE metrics (ROUGE 1, ROUGE 2, and ROUGE L), all of which are based on phrasal overlap.

In addition to evaluating the models’ performance using QAGS, we evaluated them according to the three ROUGE metrics and FactCC, another model-based metric that simply predicts the factual consistency of two texts. On all five metrics, models trained using QUALS outperformed the two baselines.

For validation, we also conducted a human-evaluation study, which involved 100 summaries generated using QUALS and 100 summaries generated using MLE for each of two datasets (XSUM and CNNDM). Human subjects were asked to compare the summaries on three attributes: factual consistency, informativeness and grammatical correctness.

On average, annotators found the QUALS-based summaries more factually accurate and more informative than the MLE-based summaries, for both datasets. On grammatical correctness, the two models’ performance was virtually indistinguishable.

Human-study stats.png
The results of the human-evaluation study. Subjects were asked whether summaries produced using QUALS were better than, worse than, or equal to those produced using MLE, on three axes.

Research areas

Related content

IT, Turin
Are you a MS or PhD student interested in a 2026 internship in the field of machine learning, deep learning, generative AI, large language models, speech technology, robotics, computer vision, optimization, operations research, quantum computing, automated reasoning, or formal methods? If so, we want to hear from you! We are looking for students interested in using a variety of domain expertise to invent, design and implement state-of-the-art solutions for never-before-solved problems. You can find more information about the Amazon Science community as well as our interview process via the links below; https://www.amazon.science/ https://amazon.jobs/content/en/career-programs/university/science https://amazon.jobs/content/en/how-we-hire/university-roles/applied-science Key job responsibilities As an Applied Science Intern, you will own the design and development of end-to-end systems. You’ll have the opportunity to write technical white papers, create roadmaps and drive production level projects that will support Amazon Science. You will work closely with Amazon scientists and other science interns to develop solutions and deploy them into production. You will have the opportunity to design new algorithms, models, or other technical solutions whilst experiencing Amazon’s customer focused culture. The ideal intern must have the ability to work with diverse groups of people and cross-functional teams to solve complex business problems. A day in the life At Amazon, you will grow into the high impact person you know you’re ready to be. Every day will be filled with developing new skills and achieving personal growth. How often can you say that your work changes the world? At Amazon, you’ll say it often. Join us and define tomorrow. Some more benefits of an Amazon Science internship include; • All of our internships offer a competitive stipend/salary • Interns are paired with an experienced manager and mentor(s) • Interns receive invitations to different events such as intern program initiatives or site events • Interns can build their professional and personal network with other Amazon Scientists • Interns can potentially publish work at top tier conferences each year About the team Applicants will be reviewed on a rolling basis and are assigned to teams aligned with their research interests and experience prior to interviews. Start dates are available throughout the year and durations can vary in length from 3-6 months for full time internships. This role may available across multiple locations in the EMEA region (Austria, Estonia, France, Germany, Ireland, Israel, Italy, Jordan, Luxembourg, Netherlands, Poland, Romania, Spain, South Africa, UAE, and UK). Please note these are not remote internships.
US, CA, Pasadena
The Amazon Web Services (AWS) Center for Quantum Computing (CQC) is a multi-disciplinary team of theoretical and experimental physicists, materials scientists, and hardware and software engineers on a mission to develop a fault-tolerant quantum computer. Throughout your internship journey, you'll have access to unparalleled resources, including state-of-the-art computing infrastructure, cutting-edge research papers, and mentorship from industry luminaries. This immersive experience will not only sharpen your technical skills but also cultivate your ability to think critically, communicate effectively, and thrive in a fast-paced, innovative environment where bold ideas are celebrated. Join us at the forefront of applied science, where your contributions will shape the future of Quantum Computing and propel humanity forward. Seize this extraordinary opportunity to learn, grow, and leave an indelible mark on the world of technology. Amazon has positions available for Quantum Research Science and Applied Science Internships in Santa Clara, CA and Pasadena, CA. We are particularly interested in candidates with expertise in any of the following areas: superconducting qubits, cavity/circuit QED, quantum optics, open quantum systems, superconductivity, electromagnetic simulations of superconducting circuits, microwave engineering, benchmarking, quantum error correction, etc. In this role, you will work alongside global experts to develop and implement novel, scalable solutions that advance the state-of-the-art in the areas of quantum computing. You will tackle challenging, groundbreaking research problems, work with leading edge technology, focus on highly targeted customer use-cases, and launch products that solve problems for Amazon customers. The ideal candidate should possess the ability to work collaboratively with diverse groups and cross-functional teams to solve complex business problems. A successful candidate will be a self-starter, comfortable with ambiguity, with strong attention to detail and the ability to thrive in a fast-paced, ever-changing environment. About the team Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences, inspire us to never stop embracing our uniqueness. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why we strive for flexibility as part of our working culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Hybrid Work We value innovation and recognize this sometimes requires uninterrupted time to focus on a build. We also value in-person collaboration and time spent face-to-face. Our team affords employees options to work in the office every day or in a flexible, hybrid work model near one of our U.S. Amazon offices.
US, MA, N.reading
Amazon Industrial Robotics is seeking exceptional talent to help develop the next generation of advanced robotics systems that will transform automation at Amazon's scale. We're building revolutionary robotic systems that combine cutting-edge AI, sophisticated control systems, and advanced mechanical design to create adaptable automation solutions capable of working safely alongside humans in dynamic environments. This is a unique opportunity to shape the future of robotics and automation at an unprecedented scale, working with world-class teams pushing the boundaries of what's possible in robotic dexterous manipulation, locomotion, and human-robot interaction. This role presents an opportunity to shape the future of robotics through innovative applications of deep learning and large language models. At Amazon Industrial Robotics we leverage advanced robotics, machine learning, and artificial intelligence to solve complex operational challenges at an unprecedented scale. Our fleet of robots operates across hundreds of facilities worldwide, working in sophisticated coordination to fulfill our mission of customer excellence. - We are pioneering the development of robotics dexterous hands that: - Enable unprecedented generalization across diverse tasks - Are compliant but at the same time impact resistant - Can enable power grasps with the same reliability as fine dexterity and nonprehensile manipulation - Can naturally cope with the uncertainty of the environment - Leverage mechanical intelligence, multi-modal sensor feedback and advanced control techniques. The ideal candidate will contribute to research that bridges the gap between theoretical advancement and practical implementation in robotics. You will be part of a team that's revolutionizing how robots learn, adapt, and interact with their environment. Join us in building the next generation of intelligent robotics systems that will transform the future of automation and human-robot collaboration. Key job responsibilities - Design and implement novel highly dexterous and reliable robotic dexterous hand morphologies - Develop parallel paths for rapid finger design and prototyping combining different actuation and sensing technologies as well as different finger morphologies - Develop new testing and validation strategies to support fast continuous integration and modularity - Build and test full hand prototypes to validate the performance of the solution - Create hybrid approaches combining different actuation technologies, under-actuation, active and passive compliance - Hand integration into rest of the embodiment - Partner with cross-functional teams to rapidly create new concepts and prototypes - Work with Amazon's robotics engineering and operations teams to grasp their requirements and develop tailored solutions - Document the designs, performance, and validation of the final system
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, WA, Bellevue
Are you excited about customer-facing research and reinventing the way people think about long-held assumptions? At Amazon, we are constantly inventing and re-inventing to be the most customer-centric company in the world. To get there, we need exceptionally talented, bright, and driven people. Amazon is one of the most recognizable brand names in the world and we distribute millions of products each year to our loyal customers. A day in the life The ideal candidate will be responsible for quantitative data analysis, building models and prototypes for supply chain systems, and developing state-of-the-art optimization algorithms to scale. This team plays a significant role in various stages of the innovation pipeline from identifying business needs, developing new algorithms, prototyping/simulation, to implementation by working closely with colleagues in engineering, product management, operations, retail and finance. As a senior member of the research team, you will play an integral part on our Supply Chain team with the following technical and leadership responsibilities: * Interact with engineering, operations, science and business teams to develop an understanding and domain knowledge of processes, system structures, and business requirements * Apply domain knowledge and business judgment to identify opportunities and quantify the impact aligning research direction to business requirements and make the right judgment on research project prioritization * Develop scalable mathematical models to derive optimal or near-optimal solutions to existing and new supply chain challenges * Create prototypes and simulations to test devised solutions * Advocate technical solutions to business stakeholders, engineering teams, as well as executive-level decision makers * Work closely with engineers to integrate prototypes into production system * Create policy evaluation methods to track the actual performance of devised solutions in production systems, identify areas with potential for improvement and work with internal teams to improve the solution with new features * Mentor team members for their career development and growth * Present business cases and document models, analyses, and their results in order to influence important decisions About the team Our organization leads the innovation of Amazon’s ultra-fast grocery product initiatives. Our key vision is to transform the online grocery experience and provide a wide grocery selection in order to be the primary destination to fulfill customer’s food shopping needs. We are a team of passionate tech builders who work endlessly to make life better for our customers through amazing, thoughtful, and creative new grocery shopping experiences. To succeed, we need senior technical leaders to forge a path into the future by building innovative, maintainable, and scalable systems.
LU, Luxembourg
Are you a MS student interested in a 2026 internship in the field of machine learning, deep learning, generative AI, large language models and speech technology, robotics, computer vision, optimization, operations research, quantum computing, automated reasoning, or formal methods? If so, we want to hear from you! We are looking for a customer obsessed Data Scientist Intern who can innovate in a business environment, building and deploying machine learning models to drive step-change innovation and scale it to the EU/worldwide. If this describes you, come and join our Data Science teams at Amazon for an exciting internship opportunity. If you are insatiably curious and always want to learn more, then you’ve come to the right place. You can find more information about the Amazon Science community as well as our interview process via the links below; https://www.amazon.science/ https://amazon.jobs/content/en/career-programs/university/science Key job responsibilities As a Data Science Intern, you will have following key job responsibilities: • Work closely with scientists and engineers to architect and develop new algorithms to implement scientific solutions for Amazon problems. • Work on an interdisciplinary team on customer-obsessed research • Experience Amazon's customer-focused culture • Create and Deliver Machine Learning projects that can be quickly applied starting locally and scaled to EU/worldwide • Build and deploy Machine Learning models using large data-sets and cloud technology. • Create and share with audiences of varying levels technical papers and presentations • Define metrics and design algorithms to estimate customer satisfaction and engagement A day in the life At Amazon, you will grow into the high impact person you know you’re ready to be. Every day will be filled with developing new skills and achieving personal growth. How often can you say that your work changes the world? At Amazon, you’ll say it often. Join us and define tomorrow. Some more benefits of an Amazon Science internship include; • All of our internships offer a competitive stipend/salary • Interns are paired with an experienced manager and mentor(s) • Interns receive invitations to different events such as intern program initiatives or site events • Interns can build their professional and personal network with other Amazon Scientists • Interns can potentially publish work at top tier conferences each year About the team Applicants will be reviewed on a rolling basis and are assigned to teams aligned with their research interests and experience prior to interviews. Start dates are available throughout the year and durations can vary in length from 3-6 months for full time internships. This role may available across multiple locations in the EMEA region (Austria, France, Germany, Ireland, Israel, Italy, Luxembourg, Netherlands, Poland, Romania, Spain and the UK). Please note these are not remote internships.
US, WA, Redmond
Amazon Leo is Amazon’s low Earth orbit satellite broadband network. Its mission is to deliver fast, reliable internet to customers and communities around the world, and we’ve designed the system with the capacity, flexibility, and performance to serve a wide range of customers, from individual households to schools, hospitals, businesses, government agencies, and other organizations operating in locations without reliable connectivity. Export Control Requirement: Due to applicable export control laws and regulations, candidates must be a U.S. citizen or national, U.S. permanent resident (i.e., current Green Card holder), or lawfully admitted into the U.S. as a refugee or granted asylum. We are searching for a senior manager with expertise in the spaceflight aerospace engineering domain of Flight Dynamics, including Mission Design of LEO Constellations, Trajectory, Maneuver Planning, and Navigation. This role will be responsible for the research and development of core spaceflight algorithms that enable the Amazon Leo mission. This role will manage the team responsible for designing and developing flight dynamics innovations for evolving constellation mission needs. Key job responsibilities This position requires expertise in simulation and analysis of astrodynamics models and spaceflight trajectories. This position requires demonstrated achievement in managing technology research portfolios. A strong candidate will have demonstrated achievement in managing spaceflight engineering Guidance, Navigation, and Control teams for full mission lifecycle including design, prototype development and deployment, and operations. Working with the Leo Flight Dynamics Research Science team, you will manage, guide, and direct staff to: • Implement high fidelity modeling techniques for analysis and simulation of large constellation concepts. • Develop algorithms for station-keeping and constellation maintenance. • Perform analysis in support of multi-disciplinary trades within the Amazon Leo team. • Formulate solutions to address collision avoidance and conjunction assessment challenges. • Develop the Leo ground system’s evolving Flight Dynamics System functional requirements. • Work closely with GNC engineers to manage on-orbit performance and develop flight dynamics operations processes About the team The Flight Dynamics Research Science team is staffed with subject matter experts of various areas within the Flight Dynamics domain. It also includes a growing Position, Navigation, and Timing (PNT) team.
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.