NeurIPS: Why causal-representation learning may be the future of AI

Francesco Locatello on the four NeurIPS papers he coauthored this year, which largely concern generalization to out-of-distribution test data.

In a conversation right before the 2021 Conference on Neural Information Processing Systems (NeurIPS), Amazon vice president and distinguished scientist Bernhard Schölkopf — according to Google Scholar, the most highly cited researcher in the field of causal inference — said that the next frontier in artificial-intelligence research was causal-representation learning.

Where existing approaches to causal inference use machine learning to discover causal relationships between variables — say, the latencies of various interrelated services on a website — causal-representation learning learns the variables themselves. “These kinds of causal representations will also go toward reasoning, which we will ultimately need if we want to move away from this pure pattern recognition view of intelligence,” Schölkopf said.

Francesco.jpg
Senior applied scientist Francesco Locatello.

Francesco Locatello, a senior applied scientist with Amazon Web Services, leads Amazon’s research on causal-representation learning, and he’s a coauthor on four papers at this year’s NeurIPS.

Assaying out-of-distribution generalization in transfer learning” concerns one of the most compelling applications of causal inference in machine learning: generalizing models trained on data with a particular probability distribution to real-world data with a different distribution.

“When you do standard machine learning, you are drawing independent samples from some probability distribution, and then you train a model that's going to generalize to the same distribution,” Locatello explains. “You're describing a physical system using a single probability distribution. Causal models are different because they model every possible state that this physical system can take as a result of an intervention. So instead of having a single probability distribution, you have a set of distributions.

Related content
Amazon Science hosts a conversation with Amazon Scholars Michael I. Jordan and Michael Kearns and Amazon distinguished scientist Bernhard Schölkopf.

“What does it mean that your test data comes from a different distribution? You have the same underlying physical system; the causal structure is the same. It's just a new intervention you have not seen. Your test distribution is different than the training, but now it's not an arbitrary distribution. It’s well posed because it's entailed by the causal structure, and it's a meaningful distribution that may happen in the real world.”

In “Assaying out-of-distribution generalization in transfer learning”, Locatello explains, “what we do is to collect a huge variety of datasets that are constructed for or adapted to this scenario where you have a very narrow data set that you can use for transfer learning, and then you have a wide variety of test data that is all out of distribution. We look at the different approaches that have been studied in the literature and compare them on fair ground.”

Although none of the approaches canvassed in the paper explicitly considers causality, Locatello says, “causal approaches should eventually be able to do better on this benchmark, and this will allow us to evaluate our progress. That's why we built it.”

Neural circuits

Today’s neural networks do representation learning as a matter of course: their inputs are usually raw data, and they learn during training which aspects of the data are most useful for the task at hand. As Schölkopf pointed out in conversation last year, causal-representation learning would simply bring causal machine learning models up to speed with conventional models.

Related content
New method goes beyond Granger causality to identify only the true causes of a target time series, given some graph constraints.

“The important thing to realize is that most machine learning applications don't come structured as a set of well-defined random variables that fully align with the underlying functioning of a physical system,” Locatello explains. “We still want to model these systems in terms of abstract variables, but nobody gives these variables to us. So you may want to learn them in order to be able to perform causal inference.”

Among his and his colleagues’ NeurIPS papers, Locatello says, the one that comes closest to the topic of causal-representation learning is “Neural attentive circuits”. Causal models typically represent causal relationships using graphs, and a neural network, too, can be thought of as an enormous graph. Locatello and his collaborators are trying to make that analogy explicit, by training a neural network to mimic the structure of a causal network.

Neural attentive circuits.png
Visualizations of graph structures learned by neural attentive circuits, from "Neural attentive circuits".

“This is a follow-up on a paper we had last year in NeurIPS,” Locatello says. “The inspiration was to design architectures that behave more similarly to causal models, where you have the noise variables — that's the data — and then you have variables that are being manipulated by functions, and they simply communicate with each other in a graph. And this graph can change dynamically when a distribution changes, for example, because of an intervention.

“In the first paper, we developed an architecture that behaves exactly like that: you have a set of neural functions that can be composed on the fly, depending on the data and the problem. The functions, the routing, and the stitching of the functions are learned. Everything is learned. But it turns out that dynamic stitching is not very scalable.

“In this new work, we essentially compiled the stitching of the functions so that for each sample it's decided beforehand — where it's going to go through the network, how the functions are going to be composed. Instead of doing it on the fly one layer at a time, you decide for the overall forward pass. And we demonstrated that these sparse learned connectivity patterns improve out-of-distribution generalization.”

Success stories

Locatello’s other NeurIPS papers are on more-conventional machine learning topics. “Self supervised amodal video object segmentation” considers the problem of reconstructing the silhouette of an occluded object, which is crucial to robotics applications, including autonomous cars.

Locatello 16_9.png
Segmentations of partially occluded objects, from "Self supervised amodal video object segmentation".

“We exploit the principle that you can build information about an object over time in a video,” Locatello explains. “Perhaps in past frames you've seen parts of the objects that are now occluded. If you can remember that you've seen this object before, and this was its segmentation mask, you can build up your segmentation over time.”

The final paper, “Are two heads the same as one? Identifying disparate treatment in fair neural networks”, considers models whose training objectives are explicitly designed to minimize bias across different types of inputs. Locatello and his colleagues find that frequently, such models — purely through training, without any human intervention — develop two “heads”: that is, they learn two different pathways through the neural network, one for inputs in the sensitive class, and one for all other inputs.

Related content
Amazon ICML paper proposes information-theoretic measurement of quantitative causal contribution.

The researchers argue that, since the network is learning two heads, anyway, it might as well be designed with a two-headed architecture: that would improve performance while meeting the same fairness standard. But this approach hasn’t been adopted, as it runs afoul of rules prohibiting disparate treatment of different groups. In this case, however, disparate treatment could be the best way to ensure fair treatment.

These last two papers are only obliquely related to causality. But, Locatello says, “causal-representation learning is a very young field. So we are trying to identify success stories, and I think these papers are going in that direction.”

“It's clear that causality will have a role in future machine learning,” he adds, “because there are a lot of open problems in machine learning that can at least be partially addressed when you start looking at causal models. And my goal really is to realize the benefits of causal models in mainstream machine learning applications. That's why some of these works are not necessarily about causality, but closer to machine learning. Because ultimately, that's our goal.”

Learn more about Amazon at NeurIPS 2022

For more on the Amazon research being presented at this year's NeurIPS, see our quick guide to Amazon's NeurIPS 2022 papers.

Related content

GB, Cambridge
We are looking for a passionate, talented, and resourceful Applied Scientist with background in Natural Language Processing (NLP), Large Language Models (LLMs), Question Answering, Information Retrieval, Reinforcement Learning, or Recommender Systems to invent and build scalable solutions for a state-of-the-art conversational assistant. The ideal candidate should have a robust foundation in machine learning and a keen interest in advancing the field. The ideal candidate would also enjoy operating in dynamic environments, have the self-motivation to take on challenging problems to deliver big customer impact, and move fast to ship solutions and then iterate on user feedback and interactions. Key job responsibilities * Work collaboratively with scientists and developers to design and implement automated, scalable NLP/ML/QA/IR models for accessing and presenting information; * Drive scalable solutions end-to-end from business requirements to prototyping, engineering, production testing to production; * Drive best practices on the team, deal with ambiguity and competing objectives, and mentor and guide junior members to achieve their career growth potential. We are open to hiring candidates to work out of one of the following locations: Cambridge, GBR
DE, BE, Berlin
We are looking for a passionate, talented, and resourceful Applied Scientist with background in Natural Language Processing (NLP), Large Language Models (LLMs), Question Answering, Information Retrieval, Reinforcement Learning, or Recommender Systems to invent and build scalable solutions for a state-of-the-art conversational assistant. The ideal candidate should have a robust foundation in machine learning and a keen interest in advancing the field. The ideal candidate would also enjoy operating in dynamic environments, have the self-motivation to take on challenging problems to deliver big customer impact, and move fast to ship solutions and then iterate on user feedback and interactions. Key job responsibilities * Work collaboratively with scientists and developers to design and implement automated, scalable NLP/ML/QA/IR models for accessing and presenting information; * Drive scalable solutions end-to-end from business requirements to prototyping, engineering, production testing to production; * Drive best practices on the team, deal with ambiguity and competing objectives, and mentor and guide junior members to achieve their career growth potential. We are open to hiring candidates to work out of one of the following locations: Berlin, BE, DEU
US, WA, Bellevue
Are you inspired by invention? Do you like the idea of seeing how your work impacts the bigger picture? Answer yes to any of these and you’ll fit right in here at Amazon Last Mile Solutions Engineering team. WW AMZL Solutions Engineering team is looking to build out our Simulation team to drive innovation across our Last Mile network. We start with the customer and work backwards in everything we do. If you’re interested in joining a rapidly growing team working to build a unique, solutions advisory group with a relentless focus on the customer, you’ve come to the right place. This is a blue-sky role that gives you a chance to roll up your sleeves and dive into big data sets in order to build simulations and experimentation systems at scale, build optimization algorithms and leverage cutting-edge technologies across Amazon. This is an opportunity to think big about how to solve a challenging problem for the customers. As a Simulation Scientist, you are an analytical problem solver who enjoys diving into data from various businesses, is excited about investigations and algorithms, can multi-task, and can credibly interface between scientists, engineers and business stakeholders. Your expertise in synthesizing and communicating insights and recommendations to audiences of varying levels of technical sophistication will enable you to answer specific business questions and innovate for the future. As a simulation scientist, you will apply cutting edge designs and methodologies for complex use cases across Last Mile network to drive innovation. In addition, you will contribute to the end state vision for simulation and experimentation of future delivery stations at Amazon. Key job responsibilities • Design, develop, and simulate engineering solutions for complex material handling challenges considering human/equipment interactions for the Last Mile network • Lead and coordinate simulation efforts for optimal solutions through equipment specification, material flow, process design, ergonomics, associate experience, operational considerations and site layout • The candidate must have the ability to work with diverse customer groups to solve business problems and provide data solutions that are organized and simple to understand. • Working with technical and non-technical customers to design experiments, simulations, and communicate results • Develop, document and update simulation standards, including standard results reports • Create basic to highly advanced models and run "what-if" scenarios to help drive to optimal solutions • Work closely with internal teams to ensure that every detail is thought through and documented using Standard Operating Procedures and/or structured change control • Work closely with vendors, suppliers and other cross functional teams to come up with innovative solutions • Simultaneously manage multiple projects and tasks while effectively influencing, negotiating, and communicating with internal and external business partners • Conduct post-mortem on simulations, after implementation of new designs, in partnering with Safety and Operations A day in the life If you are not sure that every qualification on the list above describes you exactly, we'd still love to hear from you! At Amazon, we value people with unique backgrounds, experiences, and skillsets. If you’re passionate about this role and want to make an impact on a global scale, please apply! Benefits: Amazon offers a full range of benefits that support you and eligible family members, including domestic partners and their children. Benefits can vary by location, the number of regularly scheduled hours you work, length of employment, and job status such as seasonal or temporary employment. The benefits that generally apply to regular, full-time employees include: 1. Medical, Dental, and Vision Coverage 2. Maternity and Parental Leave Options 3. Paid Time Off (PTO) 4. 401(k) Plan Learn more about our benefits here: https://amazon.jobs/en/internal/benefits/us-benefits-and-stock We are open to hiring candidates to work out of one of the following locations: Bellevue, WA, USA
BR, SP, Sao Paulo
Amazon launched the Generative AI Innovation Center in June 2023 to help AWS customers accelerate innovation and business success with Generative AI (https://press.aboutamazon.com/2023/6/aws-announces- generative -ai -innovation center). This Innovation Center provides opportunities to innovate in a fast-paced organization that contributes to breakthrough projects and technologies that are deployed across devices and the cloud. As a data scientist, you are proficient in designing and developing advanced generative AI solutions to solve diverse customer problems. You'll work with terabytes of text, images, and other types of data to solve real-world problems through Gen AI. You will work closely with account teams and ML strategists to define the use case, and with other ML scientists and engineers on the team to design experiments and find new ways to deliver customer value. The selected person will possess technical and customer-facing skills that will enable you to be part of the AWS technical team within our solution providers ecosystem/environment as well as directly to end customers. You will be able to lead discussion with customer and partner staff and senior management. A day in the life Here at AWS, we embrace our differences. We are committed to promoting our culture of inclusion. We have ten employee-led affinity groups, reaching 40,000 employees in more than 190 branches around the world. We have innovative benefit offerings and host annual and ongoing learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences. Amazon's culture of inclusion is reinforced by our 16 Leadership Principles, which remind team members to seek diverse perspectives, learn and be curious, and build trust. About the team Work/life balance Our team highly values work-life balance. It's not about how many hours you spend at home or at work; it's about the flow you establish that brings energy to both parts of your life. We believe that finding the right balance between your personal and professional life is fundamental to lifelong happiness and fulfillment. We offer flexibility in working hours and encourage you to find your own work-life balance. Mentoring and career growth Our team is dedicated to supporting new members. We have a broad mix of experience levels and mandates and are building an environment that celebrates knowledge sharing and mentorship. Our senior members enjoy one-on-one guidance and thorough but gentle code reviews. We care about your career growth and strive to assign projects based on what will help each team member become a more well-rounded engineer and enable them to take on more complex tasks in the future. We are open to hiring candidates to work out of one of the following locations: Sao Paulo, SP, BRA
MX, DIF, Mexico City
Amazon launched the Generative AI Innovation Center (GAIIC) in Jun 2023 to help AWS customers accelerate the use of Generative AI to solve business and operational problems and promote innovation in their organization (https://press.aboutamazon.com/2023/6/aws-announces-generative-ai-innovation-center). GAIIC provides opportunities to innovate in a fast-paced organization that contributes to game-changing projects and technologies that get deployed on devices and in the cloud. As a Data Science Manager in GAIIC, you'll partner with technology and business teams to build new GenAI solutions that delight our customers. You will be responsible for directing a team of data scientists, deep learning architects, and ML engineers to build generative AI models and pipelines, and deliver state-of-the-art solutions to customer’s business and mission problems. Your team will be working with terabytes of text, images, and other types of data to address real-world problems. The successful candidate will possess both technical and customer-facing skills that will allow them to be the technical “face” of AWS within our solution providers’ ecosystem/environment as well as directly to end customers. You will be able to drive discussions with senior technical and management personnel within customers and partners, as well as the technical background that enables them to interact with and give guidance to data/research/applied scientists and software developers. The ideal candidate will also have a demonstrated ability to think strategically about business, product, and technical issues. Finally, and of critical importance, the candidate will be an excellent technical team manager, someone who knows how to hire, develop, and retain high quality technical talent. AWS Sales, Marketing, and Global Services (SMGS) is responsible for driving revenue, adoption, and growth from the largest and fastest growing small- and mid-market accounts to enterprise-level customers including public sector. The AWS Global Support team interacts with leading companies and believes that world-class support is critical to customer success. AWS Support also partners with a global list of customers that are building mission-critical applications on top of AWS services. A day in the life A day in the life Here at AWS, we embrace our differences. We are committed to furthering our culture of inclusion. We have ten employee-led affinity groups, reaching 40,000 employees in over 190 chapters globally. We have innovative benefit offerings, and host annual and ongoing learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences. Amazon’s culture of inclusion is reinforced within our 16 Leadership Principles, which remind team members to seek diverse perspectives, learn and be curious, and earn trust. About the team Work/Life Balance Our team puts a high value on work-life balance. It isn’t about how many hours you spend at home or at work; it’s about the flow you establish that brings energy to both parts of your life. We believe striking the right balance between your personal and professional life is critical to life-long happiness and fulfillment. We offer flexibility in working hours and encourage you to find your own balance between your work and personal lives. Mentorship & Career Growth Our team is dedicated to supporting new members. We have a broad mix of experience levels and tenures, and we’re building an environment that celebrates knowledge sharing and mentorship. Our senior members enjoy one-on-one mentoring and thorough, but kind, code reviews. We care about your career growth and strive to assign projects based on what will help each team member develop into a better-rounded engineer and enable them to take on more complex tasks in the future. We are open to hiring candidates to work out of one of the following locations: Mexico City, DIF, MEX
US, CA, Palo Alto
The Amazon Search Mission Understanding (SMU) team is at the forefront of revolutionizing the online shopping experience through the Amazon search page. Our ambition extends beyond facilitating a seamless shopping journey; we are committed to creating the next generation of intelligent shopping assistants. Leveraging cutting-edge Large Language Models (LLMs), we aim to redefine navigation and decision-making in e-commerce by deeply understanding our users' shopping missions, preferences, and goals. By developing responsive and scalable solutions, we not only accomplish the shopping mission but also foster unparalleled trust among our customers. Through our advanced technology, we generate valuable insights, providing a guided navigation system into various search missions, ensuring a comprehensive and holistic shopping experience. Our dedication to continuous improvement through constant measurement and enhancement of the shopper experience is crucial, as we strategically navigate the balance between immediate results and long-term business growth. We are seeking an Applied Scientist who is not just adept in the theoretical aspects of Machine Learning (ML), Artificial Intelligence (AI), and Large Language Models (LLMs) but also possesses a pragmatic, hands-on approach to navigating the complexities of innovation. The ideal candidate will have a profound expertise in developing, deploying, and contributing to the next-generation shopping search engine, including but not limited to Retrieval-Augmented Generation (RAG) models, specifically tailored towards enhancing the Rufus application—an integral part of our mission to revolutionize shopping assistance. You will take the lead in conceptualizing, building, and launching groundbreaking models that significantly improve our understanding of and capabilities in enhancing the search experience. A successful applicant will display a comprehensive skill set across machine learning model development, implementation, and optimization. This includes a strong foundation in data management, software engineering best practices, and a keen awareness of the latest developments in distributed systems technology. We are looking for individuals who are determined, analytically rigorous, passionate about applied sciences, creative, and possess strong logical reasoning abilities. Join the Search Mission Understanding team, a group of pioneering ML scientists and engineers dedicated to building core ML models and developing the infrastructure for model innovation. As part of Amazon Search, you will experience the dynamic, innovative culture of a startup, backed by the extensive resources of Amazon.com (AMZN), a global leader in internet services. Our collaborative, customer-centric work environment spans across our offices in Palo Alto, CA, and Seattle, WA, offering a unique blend of opportunities for professional growth and innovation. Key job responsibilities Collaborate with cross-functional teams to identify requirements for ML model development, focusing on enhancing mission understanding through innovative AI techniques, including retrieval-Augmented Generation or LLM in general. Design and implement scalable ML models capable of processing and analyzing large datasets to improve search and shopping experiences. Must have a strong background in machine learning, AI, or computational sciences. Lead the management and experiments of ML models at scale, applying advanced ML techniques to optimize science solution. Serve as a technical lead and liaison for ML projects, facilitating collaboration across teams and addressing technical challenges. Requires strong leadership and communication skills, with a PhD in Computer Science, Machine Learning, or a related field. We are open to hiring candidates to work out of one of the following locations: Palo Alto, CA, USA | Seattle, WA, USA
US, WA, Seattle
Alexa Personality Fundamentals is chartered with infusing Alexa's trustworthy, reliable, considerate, smart, and playful personality. Come join us in creating the future of personality forward AI here at Alexa. Key job responsibilities As a Data Scientist with Alexa Personality, your work will involve machine learning, Large Language Model (LLM) and other generative technologies. You will partner with engineers, applied scientists, voice designers, and quality assurance to ensure that Alexa can sing, joke, and delight our customers in every interaction. You will take a central role in defining our experimental roadmap, sourcing training data, authoring annotation criteria and building automated benchmarks to track the improvement of our Alexa's personality. We are open to hiring candidates to work out of one of the following locations: Bellevue, WA, USA | Seattle, WA, USA
US, CA, Palo Alto
The Amazon Search Mission Understanding (SMU) team is at the forefront of revolutionizing the online shopping experience through the Amazon search page. Our ambition extends beyond facilitating a seamless shopping journey; we are committed to creating the next generation of intelligent shopping assistants. Leveraging cutting-edge Large Language Models (LLMs), we aim to redefine navigation and decision-making in e-commerce by deeply understanding our users' shopping missions, preferences, and goals. By developing responsive and scalable solutions, we not only accomplish the shopping mission but also foster unparalleled trust among our customers. Through our advanced technology, we generate valuable insights, providing a guided navigation system into various search missions, ensuring a comprehensive and holistic shopping experience. Our dedication to continuous improvement through constant measurement and enhancement of the shopper experience is crucial, as we strategically navigate the balance between immediate results and long-term business growth. We are seeking an Applied Scientist who is not just adept in the theoretical aspects of Machine Learning (ML), Artificial Intelligence (AI), and Large Language Models (LLMs) but also possesses a pragmatic, hands-on approach to navigating the complexities of innovation. The ideal candidate will have a profound expertise in developing, deploying, and contributing to the next-generation shopping search engine, including but not limited to Retrieval-Augmented Generation (RAG) models, specifically tailored towards enhancing the Rufus application—an integral part of our mission to revolutionize shopping assistance. You will take the lead in conceptualizing, building, and launching groundbreaking models that significantly improve our understanding of and capabilities in enhancing the search experience. A successful applicant will display a comprehensive skill set across machine learning model development, implementation, and optimization. This includes a strong foundation in data management, software engineering best practices, and a keen awareness of the latest developments in distributed systems technology. We are looking for individuals who are determined, analytically rigorous, passionate about applied sciences, creative, and possess strong logical reasoning abilities. Join the Search Mission Understanding team, a group of pioneering ML scientists and engineers dedicated to building core ML models and developing the infrastructure for model innovation. As part of Amazon Search, you will experience the dynamic, innovative culture of a startup, backed by the extensive resources of Amazon.com (AMZN), a global leader in internet services. Our collaborative, customer-centric work environment spans across our offices in Palo Alto, CA, and Seattle, WA, offering a unique blend of opportunities for professional growth and innovation. Key job responsibilities Collaborate with cross-functional teams to identify requirements for ML model development, focusing on enhancing mission understanding through innovative AI techniques, including retrieval-Augmented Generation or LLM in general. Design and implement scalable ML models capable of processing and analyzing large datasets to improve search and shopping experiences. Must have a strong background in machine learning, AI, or computational sciences. Lead the management and experiments of ML models at scale, applying advanced ML techniques to optimize science solution. Serve as a technical lead and liaison for ML projects, facilitating collaboration across teams and addressing technical challenges. Requires strong leadership and communication skills, with a PhD in Computer Science, Machine Learning, or a related field. We are open to hiring candidates to work out of one of the following locations: Palo Alto, CA, USA | Seattle, WA, USA
US, WA, Bellevue
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Applied Science Manager with a strong deep learning background, to lead the development of industry-leading technology with multimodal systems. Key job responsibilities As an Applied Science Manager with the AGI team, you will lead the development of novel algorithms and modeling techniques to advance the state of the art with multimodal systems. Your work will directly impact our customers in the form of products and services that make use of vision and language technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate development with multimodal Large Language Models (LLMs) and Generative Artificial Intelligence (GenAI) in Computer Vision. About the team The AGI team has a mission to push the envelope with multimodal LLMs and GenAI in Computer Vision, in order to provide the best-possible experience for our customers. We are open to hiring candidates to work out of one of the following locations: Bellevue, WA, USA | Seattle, WA, USA | Sunnyvale, CA, USA
US, WA, Seattle
Amazon is looking for a passionate, talented, and inventive Senior Applied Scientist with a strong machine learning background to help build industry-leading language technology. Our mission is to provide a delightful experience to Amazon’s customers by pushing the envelope in Natural Language Processing (NLP), Generative AI, Large Language Model (LLM), Natural Language Understanding (NLU), Machine Learning (ML), Retrieval-Augmented Generation, Responsible AI, Agent, Evaluation, and Model Adaptation. As part of our AI team in Amazon AWS, you will work alongside internationally recognized experts to develop novel algorithms and modeling techniques to advance the state-of-the-art in human language technology. Your work will directly impact millions of our customers in the form of products and services, as well as contributing to the wider research community. You will gain hands on experience with Amazon’s heterogeneous text and structured data sources, and large-scale computing resources to accelerate advances in language understanding. The Science team at AWS Bedrock builds science foundations of Bedrock, which is a fully managed service that makes high-performing foundation models available for use through a unified API. We are adamant about continuously learning state-of-the-art NLP/ML/LLM technology and exploring creative ways to delight our customers. In our daily job we are exposed to large scale NLP needs and we apply rigorous research methods to respond to them with efficient and scalable innovative solutions. At AWS Bedrock, you’ll experience the benefits of working in a dynamic, entrepreneurial environment, while leveraging AWS resources, one of the world’s leading cloud companies and you’ll be able to publish your work in top tier conferences and journals. We are building a brand new team to help develop a new NLP service for AWS. You will have the opportunity to conduct novel research and influence the science roadmap and direction of the team. Come join this greenfield opportunity! Amazon Bedrock team is part of Utility Computing (UC) About the team AWS Utility Computing (UC) provides product innovations — from foundational services such as Amazon’s Simple Storage Service (S3) and Amazon Elastic Compute Cloud (EC2), to consistently released new product innovations that continue to set AWS’s services and features apart in the industry. As a member of the UC organization, you’ll support the development and management of Compute, Database, Storage, Internet of Things (Iot), Platform, and Productivity Apps services in AWS, including support for customers who require specialized security solutions for their cloud services. Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences, inspire us to never stop embracing our uniqueness. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why we strive for flexibility as part of our working culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Hybrid Work We value innovation and recognize this sometimes requires uninterrupted time to focus on a build. We also value in-person collaboration and time spent face-to-face. Our team affords employees options to work in the office every day or in a flexible, hybrid work model near one of our U.S. Amazon offices. We are open to hiring candidates to work out of one of the following locations: Seattle, WA, USA