Line art of silicon chips developed by Annapurna Labs since its acquisition by Amazon in 2015.  Line art includes mentions of Graviton, Inferentia, and Trainium chips, along with AWS Nitro system.
Amazon's acquisition of Annapurna Labs in 2015 has led to, among other advancements, the development of five generations of the AWS Nitro system, three generations of Arm-based Graviton processors, as well as AWS Trainium and AWS Inferentia chips that are optimized for machine learning training and inference. These chips and systems were discussed at the AWS Silicon Innovation Day event on August 3. The event included a talk by Nafea Bshara, AWS vice president and distinguished engineer, on silicon innovation emerging from Annapurna Labs.

How silicon innovation became the ‘secret sauce’ behind AWS’s success

Nafea Bshara, AWS vice president and distinguished engineer, discusses Annapurna Lab’s path to silicon success; Annapurna co-founder was a featured speaker at AWS Silicon Innovation Day virtual event.

Nafea Bshara, Amazon Web Services vice president and distinguished engineer, and the co-founder of Annapurna Labs, an Israeli-based chipmaker that Amazon acquired in 2015, maintains a low profile, as does his friend and Annapurna co-founder, Hrvoye (Billy) Bilic.

Nafea Bshara headshot image
Nafea Bshara, AWS vice president and distinguished engineer.

Each executive’s LinkedIn profile is sparse, in fact, Bilic’s is out of date.

“We hardly do any interviews; our philosophy is to let our products do the talking,” explains Bshara.

Those products, and silicon innovations, have done a lot of talking since 2015, as the acquisition has led to, among other advancements, the development of five generations of the AWS Nitro System, three generations (1, 2, 3) of custom-designed, Arm-based Graviton processors that support data-intensive workloads, as well as AWS Trainium, and AWS Inferentia chips optimized for machine learning training and inference.

Some observers have described the silicon that emerges from Annapurna Labs in the U.S. and Israel as AWS’s “secret sauce”.

Nafea’s silicon journey began at Technion University in Israel, where he earned bachelor’s and master’s degrees in computer engineering, and where he first met Hrvoye. The two then went on to work for Israel-based Galileo, a company that made chips for networking switches, and controllers for networking routers. Galileo was acquired by U.S. semiconductor manufacturer Marvell in 2000, where Bshara and Bilic would work for a decade before deciding to venture out on their own.

“We had developed at least 50 different chips together,” Bshara explained, “so we had a track record and a first-hand understanding of customer needs, and the market dynamics. We could see that some market segments were being underserved, and with the support from our spouses, Lana and Liat, and our funding friends Avigdor [Willenz] and Manuel [Alba], we started Annapurna Labs.”

That was mid-2011, and three and half years later Amazon acquired the company. The two friends have continued their journey at Amazon, where their team’s work has spoken for itself.

Last year, industry analyst David Vellante praised AWS’s “revolution in system architecture.”

“Much in the same way that AWS defined the cloud operating model last decade, we believe it is once again leading in future systems. The secret sauce underpinning these innovations is specialized designs… We believe these moves position AWS to accommodate a diversity of workloads that span cloud, data center as well as the near and far edge.”

Annapurna’s work was highlighted during the AWS Silicon Innovation Day virtual event on August 3. In fact, Nafea was a featured speaker in the event. The Silicon Innovation Day broadcast, which highlighted AWS silicon innovations, included a keynote from David Brown, vice president, Amazon EC2; a talk about the history of AWS silicon innovation from James Hamilton, Amazon senior vice president and distinguished engineer who holds more than 200 patents in 22 countries in server and datacenter infrastructure, database, and cloud computing; and a fireside chat on the Nitro System with Anthony Liguori, AWS vice president and distinguished engineer, and Jeff Barr, AWS vice president and chief evangelist.

In advance of the silicon-innovation event, Amazon Science connected with Bshara to discuss the history of Annapurna, how the company and the industry have evolved in the past decade, and what the future portends.

  1. Q. 

    You co-founded Annapurna Labs just over 11 years ago. Why Annapurna?

    A. 

     I co-founded the company with my longtime partner, Billy, and with an amazing set of engineers and leaders who believed in the mission. We started Annapurna Labs because we looked at the way the chip industry was investing in infrastructure and data centers; it was minuscule at that time because everybody was going after the gold rush of mobile phones, smartphones, and tablets.

    We believed the industry was over indexing on investment for mobile, and under investing in the data center. The data center market was underserved. That, combined with the fact that there was increasing disappointment with the ineffective and non-productive method of developing chips, especially when compared with software development. The productivity of software developers had improved significantly in the past 25 years, while the productivity of chip developers hadn’t improved much since the ‘90s. In assessing the opportunity, we saw a data-center market that was being underserved, and an opportunity to redefine chip development with greater productivity, and with a better business model. Those factors contributed to us starting Annapurna Labs.

  2. Q. 

    How has the chip industry evolved in the past 11 years?

    A. 

    The chip industry realized, a bit late, but nevertheless realized that productivity and time to market needed to be addressed. While Annapurna has been a pioneer in advancing productivity and time to market, many others are following in our footsteps and transitioning to a building-blocks-centric development mindset, similar to how the software industry moved toward object-oriented, and service-oriented software design.

    Chip companies have now transitioned to what we refer to as an intellectual property-oriented, or IP-oriented, correct-by-design approach. Secondly, the chip industry has adopted the cloud. Cloud adoption has led to an explosion of compute power for building chips. Using the cloud, we are able to use compute in a ‘bursty’ way and in parallel. We and our chip-industry colleagues couldn’t deliver the silicon we do today without the cloud. This has led to the creation of a healthy market where chip companies have realized they don’t need to build everything in house, in much the same way software companies have realized they can buy libraries from open source or other library providers. The industry has matured to the point where now there is a healthy business model around buying building blocks, or IPs, from providers like Arm, Synopsys, Alphawave, or Cadence.

  3. Q. 

    Annapurna Labs was named after one of the tallest peaks in the Himalayas that’s regarded as one of the most dangerous mountains to climb. What's been the tallest peak you've had to climb?

    A. 

    I’m up in the cloud, I don’t need to climb anything [laughing]. Yes, Billy and I picked the name Annapurna Labs for a couple of reasons. First, Billy and I originally planned to climb Annapurna before we started the company. But then we got excited about the idea, acquired funding, and suddenly time was of the essence, so we put our climbing plans on hold and started the company. We called it Annapurna because at that time – and it’s true even today – there is a high barrier to entry in starting a chip company. The challenge is steep, and the risk is high, so it’s just like climbing Annapurna. We also believed that we wanted to reach a point above the clouds where you could see things very clearly, and without clutter. That’s always been a mantra for us as a company: Avoid the clutter, and look far into the future to understand what the customer really needs versus getting distracted by the day-to-day noise.

  4. Q. 

    What are the unique challenges you face in designing chips for ML training and inference versus more general CPU designs?

    A. 

    First, I would want to emphasize what challenge we didn’t have to worry about: with the strong foundation, methodologies, and engineering muscle we built delivering multiple generations of Nitro, we had confidence in our ability to execute on building the chips and manufacturing them at high volume, and high quality. So that was a major thing we didn’t need to worry about. Designing for machine learning is one the most challenging, but also the most rewarding tasks I've had the pleasure to participate in. There is an insatiable demand for machine learning right now, so anyone with a good product won’t have any issues finding customer demand. The demand is there, but there are a couple of challenges.

    Related content
    Two authors of Amazon Redshift research paper that will be presented at leading international forum for database researchers reflect on how far the first petabyte scale cloud data warehouse has advanced since it was announced ten years ago.

    The first is that customers want ‘just works’ solutions because they have enough challenges to work on the science side. So they are looking for a frictionless migration from the incumbent, let's say GPU-based machine learning, to AWS Trainium or AWS Inferentia. Our biggest challenge is to hide all the complexity so it’s what we refer to internally as boring to migrate. We don’t want our customers, the scientists and researchers, to have to think about moving from one piece of hardware to another. This is a challenge because the incumbent GPUs, specifically NVIDIA, have done a very good job developing broadly adopted technologies. The customer shouldn’t see or experience any of the hard work we’ve done in developing our chips; what the customer should experience is that it’s transparent and frictionless to transition to Inferentia and Trainium. That’s a hefty task and one of our internal challenges as a team.

    Trainium artwork from AWS website
    "The customer shouldn’t see or experience any of the hard work we’ve done in developing our chips; what the customer should experience is that it’s transparent and frictionless to transition to Inferentia and Trainium," says Bshara.

    The second challenge is more external; it’s the fact that science and machine learning are moving very fast. As an organization that is building hardware, our job is to predict what customers will need three, four, five years down the road because the development cycle for a chip can be two years, and then it gets deployed for three years. The lifecycle is around five years and trying to predict how the needs of scientists and the machine-learning community will evolve over that time span is difficult. Unlike CPU workloads, which aren’t evolving very quickly, machine learning workloads are, and it’s a bit of an art to keep apace. I would give ourselves a high score, not a perfect score, in being efficient in terms of execution and cost, while still being future proof. It’s the art of predicting what customers will need three years from now, while still executing on time and budget. These things only come with experience, and I’m fortunate to be part of a great team that has the experience to strike the right balance between cost, schedule, and future-proofing the product.

  5. Q. 

    At the recent re:MARS conference Rohit Prasad, Amazon senior vice president and Alexa head scientist, said the voice assistant is interacting with customers billions of times each week. Alexa is powered by EC2 Inf1 instances, which use AWS Inferentia chips. Why is it more effective for Alexa workloads to take advantage of this kind of specialized processing versus more general-purpose GPUs?

    A. 

    Alexa is one of those Amazon technologies that we want to bring to as many people as possible. It’s also a great example of the Amazon flywheel; the more people use it, the more value it delivers. One of our goals is to provide this service with as low latency as possible, and at the lowest cost possible, and over time improve the machine-learning algorithms behind Alexa. When people say improving Alexa, it really means handling much more complex machine learning, much more sophisticated models while maintaining the performance, and low latency. Using Inferentia, the chip, and Inf1, the EC2 instances that actually hosts all of these chips, Alexa is able to run much more advanced machine learning algorithms at lower costs and with lower latency than a standard general-purpose chip. It's not that the general-purpose chip couldn't do the job, it's that it would do so at higher costs and higher latency. With Inferentia we deliver lower latency and support much more sophisticated algorithms. This results in customers having a better experience with Alexa, and benefitting from a smarter Alexa.

  6. Q. 

    AI has been called the new electricity. But as ML models become increasingly large and complex as you just discussed, there also are concerns that energy consumption for AI model training and inference is damaging to the environment. At the chip level, what can be done to reduce the environmental impact of ML model training and Inference?

    A. 

    What we can do at the chip level, at the EC2 level, is actually work on three vectors, which we’re doing right now. The first is drive to lower power quickly by using more advanced silicon processes. Every time we build a chip in an advanced silicon process we're utilizing new semiconductor processes with smaller transistors that require less power for the same work. Because of our focus on efficient execution, we can deliver to EC2 customers a new chip based on a more modern, power-efficient silicon process every 18 months or so.

    The second vector is building more technologies, trying to accelerate in hardware and in algorithms, to get training and inference done faster. The faster we can handle training and inference, the less power is consumed. For example, one of the technologies we innovated in the last Trainium chip was something called stochastic rounding which, depending upon which measure you're looking at for some neural workloads, could accelerate neural network training by up to 30%. When you say 30% less time that translates into 30% less power.

    Another thing we're doing at the algorithmic level is offering different data types. For example, historically machine learning used a 32-bit floating point. Now we’re offering multiple versions of 16-bit and a few versions of 8-bit. When these different data types are used, they not only accelerate machine learning training, they significantly reduce the power for the same amount of workload. For example, doing matrix multiplication on a 16-bit float point is less than one-third the total power if we had done it with 32-bit floating point. The ability to add things like stochastic rounding or new data types at the algorithmic level provides a step-function improvement in power consumption for the same amount of workload.

    The third vector is credit to EC2 and the Nitro System, we’re offering more choice for customers. There are different chips optimized for different workloads, and the best way for customers to save energy is to follow the classic Amazon mantra – the everything store. We offer all different types of chips, including multiple generations of Nvidia GPUs, Intel Habana, and Trainium, and share with the customer the power profile and performance of each of the instances hosting these chips, so the customer can choose the right chip for the right workload, and optimize for the lowest possible power consumption at the lowest cost.

  7. Q. 

    I’ve focused primarily on machine learning. But let’s turn our attention to more general-purpose workloads running in the cloud, and your work on Graviton processors for Amazon EC2. 

    A. 

    Yes, in a way Graviton is the opposite of our work on machine learning, in the sense that the focus is on building server processors for general-purpose workloads running in EC2. The market for general-purpose chips has been there for thirty or forty years, and the workloads themselves haven’t evolved as rapidly as machine learning, so when we started designing, the target was clear to us.

    This is an image of a Graviton silicon chip with a blue background.
    AWS is three generations into its Graviton chip journey, and Bshara says the company has plans for "many more generations" to come.

    Because this segment of the industry wasn’t moving that fast, we felt our challenge was to move the industry faster, specifically in offering step function improvement in performance, and reducing costs, and power consumption. There are many times when you build plans, especially for chips, where the original plans are rosy, but as the development progresses you have to make tradeoffs, and the actual product falls short of the original promise. With first-generation Graviton, we experienced the opposite; we were pleasantly surprised that both performance and power efficiency turned out better than our original plan. That’s very rare in our industry.

    Related content
    Amazon DynamoDB was introduced 10 years ago today; one of its key contributors reflects on its origins, and discusses the 'never-ending journey' to make DynamoDB more secure, more available and more performant.

    The same has been true with Graviton2. Because of this there has been a massive movement inside Amazon for general workloads to move to Graviton2, mainly to save on power, but also on costs. For the same workloads, Graviton2 will on average consume 60% less power than same-generation competitive offerings, and we’re passing on those cost-savings to customers. Outside Amazon, at least 48 of AWS’s top 50 customers have not just tested, but have production workloads running on Graviton2.

    In May, Graviton3 processors became available, so it’s still Day 1 as we’re only three generations into this journey. We have plans for many more generations, but it’s always very satisfying and rewarding to hear how boring it is for customers to migrate to Graviton, and to hear all the customer success stories. It is incredibly satisfying to come to work every day and hear some of the success stories from the tens of thousands of customers using Graviton.

  8. Q. 

    You have more than 100 openings on your jobs page. What kind of talent are you seeking? And what are the characteristics of employees who succeed at Annapurna Labs? 

    A. 

    We are seeking individuals who like to work on cutting-edge technology, and approach challenges from a principles-first approach because most of the challenges we confront haven’t been dealt with before. While actual experience is important, we place greater value on proper thinking and a principles-first mindset, or reasoning from first principles.

    We also value individuals who enjoy working in a dynamic environment where the solution isn’t always the same hammer after the same nail. Given our principles-first approach, many of our challenges get solved at the chip level, the terminal level, and the system level, so we seek individuals who have systems understanding, and are skilled at working across disciplines. It’s difficult for an individual with a single discipline, or single domain knowledge, who isn’t willing to challenge her or himself by learning across other domains, to succeed at Annapurna. Last but not least, we look for individuals who focus on delivering, within a team environment. We recognize ideas are “cheap”, and what makes the difference is delivering on the idea all the way to production. Ideas are a commodity. Executing on those ideas is not.

  9. Q. 

    I've read that Billy and you share the belief that if you can dream it, you can do it. So what's your dream about future silicon development?

    A. 

    That’s true, and it’s the main reason Billy and I wanted to join AWS, because we had a common vision that there’s so much value we can bring to customers, and AWS leadership and Amazon in general were willing to invest in that vision for the long term. We agreed to be acquired by Amazon not only because of the funding and our common long-term vision, but also because building components for our own data centers would allow us to quickly deliver customer value. We’ve been super happy with the relationship for many reasons, but primarily because of our ability to have customer impact at global scale.

    At Amazon, we operate at such a scale and with such a diversity of customers that we are capable of doing application-specific, or domain-specific acceleration. Machine learning is one example of that. What we’ve done with Aqua (advanced query accelerator) for Amazon Redshift is another example where we’ve delivered hardware-based acceleration for analytics. Our biggest challenge these days is deciding what project to prioritize. There’s no shortage of opportunities to deliver value. The only way we’re able to take this approach is because of AWS. Developing silicon requires significant investment, and the only way to gain a good return on that investment is by having a lot of volume and cost-effective development, and we’ve been able to develop a large, and successful customer base with AWS.

    I should also add that before joining Amazon we thought we really took a long-term perspective. But once you sit in Amazon meetings, you realize what long-term strategic thinking really means. I continue to learn every day about how to master that. Suffice to say, we have a product roadmap, and a technology and investment strategy that extends to 2032. As much uncertainty as there is in the future, there are a few things we’re highly convicted in, and we’re investing in them, even though they may be ten years out. I obviously can’t disclose future product plans, but we continue to dream big on behalf of our customers.

    The AWS Annapurna Labs team has more than 100 job openings for software developers, physical design engineers, design specification engineers, and many other technical roles. The team has development centers in the U.S. and Israel.

Research areas

Related content

AU, VIC, Melbourne
We are scaling an advanced team of talented Machine Learning Scientists in Melbourne. This is your chance to join our a wider international community of ML experts changing the way our customers experience Amazon. Amazon's International Machine Learning team partners with businesses across the diverse Amazon ecosystem to drive innovation and deliver exceptional experiences for customers around the globe. Our team works on a wide variety of high-impact projects that deliver innovation at global scale, leveraging unrivalled access to the latest technology, whilst actively contributing to the research community by publishing in top machine learning conferences. As part of Amazon's Research and Development organization, you will have the opportunity to push the boundaries of applied science and deploy solutions that directly benefit millions of Amazon customers worldwide. Whether you are exploring the frontiers of generative AI, developing next-generation recommender systems, or optimizing agentic workflows, your work at Amazon has the power to truly change the world. Join us in this exciting journey as we redefine the present and the future of innovative applied science. Key job responsibilities - You will take on complex problems, work on solutions that either leverage or extend existing academic and industrial research, and utilize your own out-of-the-box pragmatic thinking. - In addition to coming up with novel solutions and building prototypes, you will deliver these to production in customer facing applications, in partnership with product and development teams. - You will publish papers internally and externally, contributing to advancing knowledge in the field of applied machine learning and generative AI. About the team Our team is composed of scientists with PhDs, with a strong publication profile and an appetite to see the impact of innovation on real-world systems at scale.
US, WA, Seattle
Join the Worldwide Sustainability (WWS) organization where we capitalize on our size, scale, and inventive culture to build a more resilient and sustainable company. WWS manages our social and environmental impacts globally, driving solutions that enable our customers, businesses, and the world around us to become more sustainable. Sustainability Science and Innovation is a multi-disciplinary team within the WW Sustainability organization that combines science, analytics, economics, statistics, machine learning, product development, and engineering expertise to identify, evaluate and/or develop new science, technologies, and innovations that aim to address long-term sustainability challenges. We are looking for a Sr. Research Scientist to help us develop and drive innovative scientific solutions that will improve the sustainability of materials in our products, packaging, operations, and infrastructure. You will be at the forefront of exploring and resolving complex sustainability issues, bringing innovative ideas to the table, and making meaningful contributions to projects across SSI’s portfolio. This role not only demands technical expertise but also a strategic mindset and the agility to adapt to evolving sustainability challenges through self-driven learning and exploration. In this role, you will leverage your breadth of expertise in AI models and methodologies and industrial research experience to build scientific tools that inform sustainability strategies related to materials and energy. The successful applicant will lead by example, pioneering science-vetted data-driven approaches, and working collaboratively to implement strategies that align with Amazon’s long-term sustainability vision. Key job responsibilities - Develop scientific models that help solve complex and ambiguous sustainability problems, and extract strategic learnings from large datasets. - Work closely with applied scientists and software engineers to implement your scientific models. - Support early-stage strategic sustainability initiatives and effectively learn from, collaborate with, and influence stakeholders to scale-up high-value initiatives. - Support research and development of cross-cutting technologies for industrial decarbonization, including building the data foundation and analytics for new AI models. - Drive innovation in key focus areas including packaging materials, building materials, and alternative fuels. About the team Diverse Experiences: World Wide Sustainability (WWS) values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Inclusive Team Culture: It’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon conferences, inspire us to never stop embracing our uniqueness. Mentorship & Career Growth: We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. Work/Life Balance: We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why flexible work hours and arrangements are part of our culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve.
GB, MLN, Edinburgh
Do you want a role with deep meaning and the ability to make a major impact? As part of Intelligent Talent Acquisition (ITA), you'll have the opportunity to reinvent the hiring process and deliver unprecedented scale, sophistication, and accuracy for Amazon Talent Acquisition operations. ITA is an industry-leading people science and technology organization made up of scientists, engineers, analysts, product professionals and more, all with the shared goal of connecting the right people to the right jobs in a way that is fair and precise. Last year we delivered over 6 million online candidate assessments, and helped Amazon deliver billions of packages around the world by making it possible to hire hundreds of thousands of workers in the right quantity, at the right location and at exactly the right time. You’ll work on state-of-the-art research, advanced software tools, new AI systems, and machine learning algorithms, leveraging Amazon's in-house tech stack to bring innovative solutions to life. Join ITA in using technologies to transform the hiring landscape and make a meaningful difference in people's lives. Together, we can solve the world's toughest hiring problems. A day in the life As a Research Scientist, you will partner on design and development of AI-powered systems to scale job analyses enterprise-wide, match potential candidates to the jobs they’ll be most successful in, and conduct validation research for top-of-funnel AI-based evaluation tools. You’ll have the opportunity to develop and implement novel research strategies using the latest technology and to build solutions while experiencing Amazon’s customer-focused culture. The ideal scientist must have the ability to work with diverse groups of people and inter-disciplinary cross-functional teams to solve complex business problems. About the team The Lead Generation & Detection Services (LEGENDS) organization is a specialized organization focused on developing AI-driven solutions to enable fair and efficient talent acquisition processes across Amazon. Our work encompasses capabilities across the entire talent acquisition lifecycle, including role creation, recruitment strategy, sourcing, candidate evaluation, and talent deployment. The focus is on utilizing state-of-the-art solutions using Deep Learning, Generative AI, and Large Language Models (LLMs) for recruitment at scale that can support immediate hiring needs as well as longer-term workforce planning for corporate roles. We maintain a portfolio of capabilities such as job-person matching, person screening, duplicate profile detection, and automated applicant evaluation, as well as a foundational competency capability used throughout Amazon to help standardize the assessment of talent interested in Amazon.
US, MA, N.reading
Amazon Industrial Robotics is seeking exceptional talent to help develop the next generation of advanced robotics systems that will transform automation at Amazon's scale. We're building revolutionary robotic systems that combine cutting-edge AI, sophisticated control systems, and advanced mechanical design to create adaptable automation solutions capable of working safely alongside humans in dynamic environments. This is a unique opportunity to shape the future of robotics and automation at an unprecedented scale, working with world-class teams pushing the boundaries of what's possible in robotic dexterous manipulation, locomotion, and human-robot interaction. This role presents an opportunity to shape the future of robotics through innovative applications of deep learning and large language models. At Amazon Industrial Robotics we leverage advanced robotics, machine learning, and artificial intelligence to solve complex operational challenges at an unprecedented scale. Our fleet of robots operates across hundreds of facilities worldwide, working in sophisticated coordination to fulfill our mission of customer excellence. - We are pioneering the development of robotics dexterous hands that: - Enable unprecedented generalization across diverse tasks - Are compliant but at the same time impact resistant - Can enable power grasps with the same reliability as fine dexterity and nonprehensile manipulation - Can naturally cope with the uncertainty of the environment - Leverage mechanical intelligence, multi-modal sensor feedback and advanced control techniques. The ideal candidate will contribute to research that bridges the gap between theoretical advancement and practical implementation in robotics. You will be part of a team that's revolutionizing how robots learn, adapt, and interact with their environment. Join us in building the next generation of intelligent robotics systems that will transform the future of automation and human-robot collaboration. Key job responsibilities - Design and implement novel sensing and actuation technologies for dexterous manipulation - Develop parallel paths for rapid finger design and prototyping combining different actuation and sensing technologies as well as different finger morphologies - Develop new testing and validation strategies to support fast continuous integration and modularity - Build and test full hand prototypes to validate the performance of the solution - Create hybrid approaches combining different actuation technologies, under-actuation, active and passive compliance - Hand integration into rest of the embodiment - Partner with cross-functional teams to rapidly create new concepts and prototypes - Work with Amazon's robotics engineering and operations teams to grasp their requirements and develop tailored solutions - Document the designs, performance, and validation of the final system
US, MA, North Reading
Amazon Industrial Robotics is seeking exceptional talent to help develop the next generation of advanced robotics systems that will transform automation at Amazon's scale. We're building revolutionary robotic systems that combine cutting-edge AI, sophisticated control systems, and advanced mechanical design to create adaptable automation solutions capable of working safely alongside humans in dynamic environments. This is a unique opportunity to shape the future of robotics and automation at an unprecedented scale, working with world-class teams pushing the boundaries of what's possible in robotic dexterous manipulation, locomotion, and human-robot interaction. This role presents an opportunity to shape the future of robotics through innovative applications of deep learning and large language models. At Amazon Industrial Robotics we leverage advanced robotics, machine learning, and artificial intelligence to solve complex operational challenges at an unprecedented scale. Our fleet of robots operates across hundreds of facilities worldwide, working in sophisticated coordination to fulfill our mission of customer excellence. We are pioneering the development of systems that: • Enables unprecedented generalization across diverse tasks • Enables contact-rich manipulation in different environments • Seamlessly integrates mobility and manipulation • Leverage mechanical intelligence, multi-modal sensor feedback and advanced control techniques. The ideal candidate will contribute to research that bridges the gap between theoretical advancement and practical implementation in robotics. You will be part of a team that's revolutionizing how robots learn, adapt, and interact with their environment. Join us in building the next generation of intelligent robotics systems that will transform the future of automation and human-robot collaboration!
US, WA, Seattle
We are a passionate team applying the latest advances in technology to solve real-world challenges. As a Data Scientist working at the intersection of machine learning and advanced analytics, you will help develop innovative products that enhance customer experiences. Our team values intellectual curiosity while maintaining sharp focus on bringing products to market. Successful candidates demonstrate responsiveness, adaptability, and thrive in our open, collaborative, entrepreneurial environment. Working at the forefront of both academic and applied research, you will join a diverse team of scientists, engineers, and product managers to solve complex business and technology problems using scientific approaches. You will collaborate closely with other teams to implement innovative solutions and drive improvements. At Amazon, we cultivate an inclusive culture through our Leadership Principles, which emphasize seeking diverse perspectives, continuous learning, and building trust. Our global community includes thirteen employee-led affinity groups with 40,000 members across 190 chapters, showcasing our commitment to embracing differences and fostering continuous learning through local, regional, and global programs. We prioritize work-life balance, recognizing it as fundamental to long-term happiness and fulfillment. Our team is committed to supporting your career development through challenging projects, mentorship opportunities, and targeted training programs that help you reach your full potential. Key job responsibilities Key job responsibilities * Deliver data analyses that optimize overall team process and guide decision-making * Deep dive to understand source of anomalies across a variety of datasets including low-level sequencing read data * Identify key metrics that are drivers to achieve team goals; work with senior stakeholders to refine your results * Use modern statistical methods to highlight insights for predictive & generative ML models and assay process * Perform correlation analysis, significance testing, and simulation on high- and low-fidelity datasets for various types of readouts * Generate reports with tables and visualization that support operational cycle analysis and one-off POC experiments * Collaborate with multi-disciplinary domain experts to support your findings and their experiments * Write well-tested scripts that can be promoted by our software teams to production pipelines * Learn about new statistical methods for our domain and adopt them in your work * Work fluently in SQL and Python. Be skilled in generating compelling visualizations. A day in the life New data has just landed and promoted to our datalake. You load the data and verify it's overall integrity by visualizing variation across target subsets. You realize we may have made progress toward our goals and begin to test the validity of your nominal results. At midday you grab lunch with new coworkers and learn about their fields or weird interests (there are many). You generate visualizations for the entire dataset and perform significance tests that reinforce specific findings. You meet with peers in the afternoon to discuss your findings and breakdown the remaining tasks to finalize your group report! About the team Innovators wanted! Are you an entrepreneur? A builder? A dreamer? This role is part of an Amazon Special Projects team that takes the company’s Think Big leadership principle to the limits. We focus on creating entirely new products and services with a goal of positively impacting the lives of our customers. No industries or subject areas are out of bounds. If you’re interested in innovating at scale to address big challenges in the world, this is the team for you.
US, NY, New York
About Sponsored Products and Brands The Sponsored Products and Brands team at Amazon Ads is re-imagining the advertising landscape through industry leading generative AI technologies, revolutionizing how millions of customers discover products and engage with brands across Amazon.com and beyond. We are at the forefront of re-inventing advertising experiences, bridging human creativity with artificial intelligence to transform every aspect of the advertising lifecycle from ad creation and optimization to performance analysis and customer insights. We are a passionate group of innovators dedicated to developing responsible and intelligent AI technologies that balance the needs of advertisers, enhance the shopping experience, and strengthen the marketplace. If you're energized by solving complex challenges and pushing the boundaries of what's possible with AI, join us in shaping the future of advertising. About our team The Search Ranking and Interleaving (R&I) team within Sponsored Products and Brands is responsible for determining which ads to show and the quality of ads shown on the search page (e.g., relevance, personalized and contextualized ranking to improve shopper experience, where to place them, and how many ads to show on the search page. This helps shoppers discover new products while helping advertisers put their products in front of the right customers, aligning shoppers’, advertisers’, and Amazon’s interests. To do this, we apply a broad range of GenAI and ML techniques to continuously explore, learn, and optimize the ranking and allocation of ads on the search page. We are an interdisciplinary team with a focus on improving the SP experience in search by gaining a deep understanding of shopper pain points and developing new innovative solutions to address them. A day in the life As an Applied Scientist on this team, you will identify big opportunities for the team to make a direct impact on customers and the search experience. You will work closely with with search and retail partner teams, software engineers and product managers to build scalable real-time GenAI and ML solutions. You will have the opportunity to design, run, and analyze A/B experiments that improve the experience of millions of Amazon shoppers while driving quantifiable revenue impact while broadening your technical skillset. Key job responsibilities - Solve challenging science and business problems that balance the interests of advertisers, shoppers, and Amazon. - Drive end-to-end GenAI & Machine Learning projects that have a high degree of ambiguity, scale, complexity. - Develop real-time machine learning algorithms to allocate billions of ads per day in advertising auctions. - Develop efficient algorithms for multi-objective optimization using deep learning methods to find operating points for the ad marketplace then evolve them - Research new and innovative machine learning approaches.
US, WA, Seattle
Innovators wanted! Are you an entrepreneur? A builder? A dreamer? This role is part of an Amazon Special Projects team that takes the company’s Think Big leadership principle to the limits. If you’re interested in innovating at scale to address big challenges in the world, this is the team for you. Our team rewards curiosity while maintaining a laser-focus in bringing products to market. Competitive candidates are responsive, flexible, and able to succeed within an open, collaborative, entrepreneurial, startup-like environment. At the forefront of both academic and applied research in this product area, you have the opportunity to work together with a diverse and talented team of scientists, engineers, and product managers and collaborate with other teams. We are seeking a Principal Applied Scientist working on machine learning applications in life sciences. This role combines scientific leadership with hands-on innovation, driving solutions from exploratory research through production-ready solutions deployment, while maintaining high scientific standards. You will work with Amazon's large-scale computing resources to accelerate advances in machine learning applications. Key job responsibilities - Lead ML for life science efforts using computational design approaches and ML-based tools. - Guide teams in applying SOTA ML methods, experimentation design, and modeling approaches. - Transform complex real world problems into scientific challenges and allocate resources effectively. - Review requirements, conduct technical architecture reviews, and make informed judgments around technical and business tradeoffs. - Provide mentorship to Applied Scientists, Research Scientists and Data Scientists while maintaining scientific rigor. - Collaborate with cross functional teams.
US, MD, Jessup
Application deadline: Applications will be accepted on an ongoing basis Are you excited to help the US Intelligence Community design, build, and implement AI algorithms, including advanced Generative AI solutions, to augment decision making while meeting the highest standards for reliability, transparency, and scalability? The Amazon Web Services (AWS) US Federal Professional Services team works directly with US Intelligence Community agencies and other public sector entities to achieve their mission goals through the adoption of Machine Learning (ML) and Generative AI methods. We build models for text, image, video, audio, and multi-modal use cases, leveraging both traditional ML approaches and state-of-the-art generative models including Large Language Models (LLMs), text-to-image generation, and other advanced AI capabilities to fit the mission. Our team collaborates across the entire AWS organization to bring access to product and service teams, to get the right solution delivered and drive feature innovation based on customer needs. At AWS, we're hiring experienced data scientists with a background in both traditional and generative AI who can help our customers understand the opportunities their data presents, and build solutions that earn the customer trust needed for deployment to production systems. In this role, you will work closely with customers to deeply understand their data challenges and requirements, and design tailored solutions that best fit their use cases. You should have broad experience building models using all kinds of data sources, and building data-intensive applications at scale. You should possess excellent business acumen and communication skills to collaborate effectively with stakeholders, develop key business questions, and translate requirements into actionable solutions. You will provide guidance and support to other engineers, sharing industry best practices and driving innovation in the field of data science and AI. This position requires that the candidate selected must currently possess and maintain an active TS/SCI Security Clearance with Polygraph. The position further requires the candidate to opt into a commensurate clearance for each government agency for which they perform AWS work. Key job responsibilities As a Data Scientist, you will: - Collaborate with AI/ML scientists and architects to research, design, develop, and evaluate AI algorithms to address real-world challenges - Interact with customers directly to understand the business problem, help and aid them in implementation of AI solutions, deliver briefing and deep dive sessions to customers and guide customer on adoption patterns and paths to production. - Create and deliver best practice recommendations, tutorials, blog posts, sample code, and presentations adapted to technical, business, and executive stakeholder - Provide customer and market feedback to Product and Engineering teams to help define product direction - This position may require up to 25% local travel. About the team Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (diversity) conferences, inspire us to never stop embracing our uniqueness. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why flexible work hours and arrangements are part of our culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional.
US, MD, Jessup
Application deadline: Applications will be accepted on an ongoing basis Are you excited to help the US Intelligence Community design, build, and implement AI algorithms, including advanced Generative AI solutions, to augment decision making while meeting the highest standards for reliability, transparency, and scalability? The Amazon Web Services (AWS) US Federal Professional Services team works directly with US Intelligence Community agencies and other public sector entities to achieve their mission goals through the adoption of Machine Learning (ML) and Generative AI methods. We build models for text, image, video, audio, and multi-modal use cases, leveraging both traditional ML approaches and state-of-the-art generative models including Large Language Models (LLMs), text-to-image generation, and other advanced AI capabilities to fit the mission. Our team collaborates across the entire AWS organization to bring access to product and service teams, to get the right solution delivered and drive feature innovation based on customer needs. At AWS, we're hiring experienced data scientists with a background in both traditional and generative AI who can help our customers understand the opportunities their data presents, and build solutions that earn the customer trust needed for deployment to production systems. In this role, you will work closely with customers to deeply understand their data challenges and requirements, and design tailored solutions that best fit their use cases. You should have broad experience building models using all kinds of data sources, and building data-intensive applications at scale. You should possess excellent business acumen and communication skills to collaborate effectively with stakeholders, develop key business questions, and translate requirements into actionable solutions. You will provide guidance and support to other engineers, sharing industry best practices and driving innovation in the field of data science and AI. This position requires that the candidate selected must currently possess and maintain an active TS/SCI Security Clearance with Polygraph. The position further requires the candidate to opt into a commensurate clearance for each government agency for which they perform AWS work. Key job responsibilities As a Data Scientist, you will: - Collaborate with AI/ML scientists and architects to research, design, develop, and evaluate AI algorithms to address real-world challenges - Interact with customers directly to understand the business problem, help and aid them in implementation of AI solutions, deliver briefing and deep dive sessions to customers and guide customer on adoption patterns and paths to production. - Create and deliver best practice recommendations, tutorials, blog posts, sample code, and presentations adapted to technical, business, and executive stakeholder - Provide customer and market feedback to Product and Engineering teams to help define product direction - This position may require up to 25% local travel. About the team Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (diversity) conferences, inspire us to never stop embracing our uniqueness. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why flexible work hours and arrangements are part of our culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional.