Jesse Levinson, co-founder and CTO of Zoox
Jesse Levinson, co-founder and CTO of Zoox, completed his PhD and postdoc under Sebastian Thrun at Stanford. He developed algorithms for Stanford’s entry in the 2007 DARPA Urban Challenge and went on to lead the self-driving team’s research and development efforts.
Zoox

The future of mobility-as-a-service

Jesse Levinson, co-founder and CTO of Zoox, answers 3 questions about the challenges of developing autonomous vehicles and why he’s excited about Zoox’s robotaxi fleet.

In June 2020, Amazon acquired Zoox, a then six-year-old California-based startup focused on “creating autonomous mobility from the ground up.”

Six months later, Zoox, now an independent Amazon subsidiary, shared publicly for the first time a look at its electric, autonomous vehicle created for dense, urban environments. The vehicle reveal marked a key milestone toward the organization’s vision of creating an autonomous robotaxi fleet and ride-hailing service designed with passengers in mind.

At its unveiling in December 2020, Zoox CEO Aicha Evans said her team is transforming the rider experience to provide superior “mobility-as-a-service” for customers. Moreover, she added, given the current data related to carbon emissions and traffic accidents, “It’s more important than ever that we build a sustainable, safe solution that allows riders to get from point A to point B.”

See how a Zoox robotaxi traverses city streets.

Jesse Levinson, co-founder and chief technology officer of Zoox, guides the company’s technology roadmap and execution to turn its mobility-as-a-service vision into reality. After graduating summa cum laude from Princeton, he completed his PhD and postdoc under Sebastian Thrun at Stanford. There, he developed algorithms for Stanford’s successful entry in the 2007 DARPA Urban Challenge and went on to lead the self-driving team’s research and development efforts.

Amazon Science asked Levinson about the challenges of developing self-driving vehicles and why he’s excited about Zoox’s approach.

Q. You were one of the authors on the 2008 paper, Junior: The Stanford Entry in the Urban Challenge. That race was a closed-course competition, and not quite representative of real-world challenges. But what key observations did you take away from that experience?

Probably the most important realization after the race was the dichotomy of how much there was still left to solve and the fact that it was actually all going to be solvable. It’s quite easy to get enchanted with one or the other of those observations; either that the problem is practically impossible because of all the things that still aren’t perfect, or that it must be almost solved because of some super cool demo or milestone that seems incredibly impressive. The reality is in between, and for whatever reason, it’s surprisingly hard for people to maintain a nuanced appreciation of that balance.

Achieving a world with ubiquitous autonomous vehicles will be an incremental process that advances every year — and remember, the alternative is the bar of human performance that stays nearly stagnant.
Jesse Levinson

In 2004, DARPA held its first Grand Challenge:  a 125-mile race in the desert. Of the 20 teams that entered, none completed the race, and the best vehicle only completed about six miles. The industry (and the media) widely regarded the outcome as an abysmal failure of AI. Yet it was not a failure, but an incredible feat of engineering. If an autonomous vehicle can drive six miles in the desert all by itself, then it doesn’t take an incredible imagination to foresee it driving 125 miles.

Lo and behold, the very next year, six vehicles finished the full 125-mile course. It was a promising step towards the future, and a year later, in 2006, DARPA announced the Urban Challenge, which several teams completed successfully. Our entry at Stanford came in second place. Excited by the results, many people made overly optimistic predictions on the mass-adoption of self-driving cars, which were subsequently deflated by various challenges we’ve seen in the industry since that time.

It has been eye-opening to watch the public's reaction to self-driving cars over time. I have always tried my best to be upfront, honest, and realistic about where the technology is — and while I’ve certainly not nailed all of my predictions, I do think I’ve managed to be fairly balanced overall. As technologists, when we are overly optimistic or pessimistic, we do a disservice to ourselves, the industry, and our technology. Achieving a world with ubiquitous autonomous vehicles will be an incremental process that advances every year — and remember, the alternative is the bar of human performance that stays nearly stagnant. It’s the opportunity of a lifetime to participate in the journey of making autonomous driving technology relentlessly better. Soon, it will reach a crossover point where the public begins to adopt it at scale, which will be a transformative win for society at large.

Q. Following up on your answer, what did you learn from that experience that you apply to your current role at Zoox? Has your approach changed since that challenge or remained largely the same?

So much! I’m grateful for that experience because it was formative in the early approach of Zoox. Here’s some of the lessons I took away from it:

Zoox Autonomous Vehicle - Single Side - Coit Tower SF.png
Zoox notes is "the first in the industry to showcase a driving, purpose-built robotaxi capable of operating up to 75 miles per hour."
Zoox

First, teaching cars to drive will not take as long as we thought. In the early 2000s, we all thought it would be many, many decades before self-driving cars would be a reality. The DARPA challenge changed that. To build a vehicle that could navigate many realistic traffic scenarios only took about a year for a small team. Of course, there’s a huge difference between that and what’s required to operate an autonomous vehicle on public roads. But it was an important milestone that highlighted that autonomous driving technology could be a reality within a couple of decades.

Second, system integration and wide-scale testing is critical. No amount of knowledge about artificial intelligence, or anything else for that matter, will lead a mythical genius to intellectually divine a perfect solution. We need to combine and integrate many different complex systems and then see what works and fails through simulations, then closed courses, then public roads (with safety drivers). We have to test and experiment and iterate with massive data and scale, as opposed to trying to reason our way to a perfect solution.

On the other hand, blindly searching for progress without having any vision or architectural insights is also a bad idea; that’s one of the reasons why we identified the benefits of 270-degree sensing on all four corners of our ground-up vehicle at Zoox way back in 2014, a few years before we could drive autonomously in cities — because we knew from first principles that it was the right way to perceive the world.

Zoox Autonomous Vehicle - Reveal Sensor Detail.png
The Zoox vehicles utilize a unique sensor (some of which are seen here) architecture of cameras, radar, and LIDAR to obtain a 270-degree field of view on all four corners of the vehicle.
Zoox

Last, we have to test the various software and hardware components collectively to see how they respond to errors and uncertainty. By building a robust system that handles a cascading series of errors and ambiguities, you can explicitly track uncertainty and represent the state of the world more thoroughly. The proper representation of the world is not a singular, perfect model, but rather a distribution of probabilities and uncertainties. If you can design your system to be robust to imperfect sensor data, unpredictable agents, and unusual environments, you have a real shot at solving the problem in a world that’s not always the way you want it to be. It’s actually what humans do really well all the time, even though we’re rarely conscious that we’re doing it.

Q. You’ve said that safety is the foundation of everything Zoox does, and that the experience of building Zoox’s robotaxi has given you the opportunity to reimagine passenger safety. Can you give us insight into some of the systems you’ve developed for passenger safety, particularly the AI stack that underpins these efforts?

Yes, that’s right: safety is absolutely fundamental to the Zoox mission. With apologies for using an overused phrase, autonomous mobility allows for a paradigm shift (sorry!) in safety — from reactive to proactive. It’s an important point: automotive safety has always been reactive, focused on protecting vehicle occupants in crashes, which are seen to be inevitable. By building an autonomous vehicle from the ground-up, we can add a layer of proactive crash prevention that simply does not exist in today’s human-driven cars, and a focus on preventing crashes from occurring in the first place. We have more than a hundred safety innovations that do not exist in conventional cars today.

Zoox Autonomous Vehicle - Interior day.png
The vehicle features a four-seat, face-to-face symmetrical seating configuration that eliminates the steering wheel and bench seating seen in conventional car designs.
Zoox

We are also developing the AI, vehicle, and service all together. Integrating the software, sensor, and vehicle subsystems is a complex challenge that requires tight, cross-functional collaboration. It would be difficult to create this level of system integration across multiple companies with divergent commercial interests. Building a ground-up vehicle has allowed us to design and choose our own sensor suite to best solve self-driving. We’ve outfitted our Toyota Highlander fleet with this same sensor architecture as our ground-up vehicle so that we can gather large amounts of data and test in environments like San Francisco and Las Vegas while our in-house vehicle is still under development.

Our software stack includes mapping, localization, sensor calibration, perception, prediction, path planning, vehicle control, infrastructure, firmware, diagnostics/messaging/monitoring/logging, and simulation. All of this software is continuously improving, with additions of new features and iterative software updates that are put through rigorous offline validations and on-vehicle structured testing.

Our vehicles also use a variety of advanced sensors, including LIDAR, cameras, and radar, to see objects on all sides of the vehicle. And because of the geometrical configuration of these sensors, we can almost always see around and behind the objects nearest to us, which is particularly helpful in dense urban environments. Our software then uses a combination of machine learning and geometric reasoning to understand the sensor data, make sense of the scene unfolding around the vehicle, and effectively navigate the roads.

We’re excited to launch our first commercial driverless service, but we won’t do so until we’re ready to operate on public roads at safety levels that meaningfully surpass that of humans.
Jesse Levinson

For example, in a busy downtown intersection, our vehicle might be identifying a construction zone based on road cones and signs, while also detecting, tracking, and predicting the motion of hundreds of other agents (vehicles, pedestrians, bicyclists, etc.) around it. Once the perception system understands the environment and can predict how surrounding agents will move, the planner uses that information and context to adapt its driving behavior to the dynamic road conditions. The planner normally tries to maintain a certain lateral distance between itself and other vehicles, but it could decide to slightly reduce that distance in order to avoid a cone in the road ahead.

By combining both the hardware and software design, we are able to reimagine passenger safety. We are confident in our sensors’ abilities to detect activity in the environment around the vehicle, but that has to be validated in a wide range of scenarios. And our vehicle has performed extremely well in crash testing, which is still important, because no matter how sophisticated the AI is, we can’t guarantee that nothing will ever hit us. We’re excited to launch our first commercial driverless service, but we won’t do so until we’re ready to operate on public roads at safety levels that meaningfully surpass that of humans.

Research areas

Related content

US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, San Francisco
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Member of Technical Staff with a strong deep learning background, to build industry-leading Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As a Member of Technical Staff with the AGI team, you will lead the development of algorithms and modeling techniques, to advance the state of the art with LLMs. You will lead the foundational model development in an applied research role, including model training, dataset design, and pre- and post-training optimization. Your work will directly impact our customers in the form of products and services that make use of GenAI technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in LLMs. About the team The AGI team has a mission to push the envelope in GenAI with LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, Sunnyvale
Prime Video is a first-stop entertainment destination offering customers a vast collection of premium programming in one app available across thousands of devices. Prime members can customize their viewing experience and find their favorite movies, series, documentaries, and live sports – including Amazon MGM Studios-produced series and movies; licensed fan favorites; and programming from Prime Video add-on subscriptions such as Apple TV+, Max, Crunchyroll and MGM+. All customers, regardless of whether they have a Prime membership or not, can rent or buy titles via the Prime Video Store, and can enjoy even more content for free with ads. Are you interested in shaping the future of entertainment? Prime Video's technology teams are creating best-in-class digital video experience. As a Prime Video technologist, you’ll have end-to-end ownership of the product, user experience, design, and technology required to deliver state-of-the-art experiences for our customers. You’ll get to work on projects that are fast-paced, challenging, and varied. You’ll also be able to experiment with new possibilities, take risks, and collaborate with remarkable people. We’ll look for you to bring your diverse perspectives, ideas, and skill-sets to make Prime Video even better for our customers. With global opportunities for talented technologists, you can decide where a career Prime Video Tech takes you! We are looking for a self-motivated, passionate and resourceful Sr. Applied Scientists with Recommender System or Search Ranking or Ads Ranking experience to bring diverse perspectives, ideas, and skill-sets to make Prime Video even better for our customers. You will spend your time as a hands-on machine learning practitioner and a research leader. You will play a key role on the team, building and guiding machine learning models from the ground up. At the end of the day, you will have the reward of seeing your contributions benefit millions of Amazon.com customers worldwide. Key job responsibilities - Develop AI solutions for various Prime Video Recommendation/Search systems using Deep learning, GenAI, Reinforcement Learning, and optimization methods; - Work closely with engineers and product managers to design, implement and launch AI solutions end-to-end; - Design and conduct offline and online (A/B) experiments to evaluate proposed solutions based on in-depth data analyses; - Effectively communicate technical and non-technical ideas with teammates and stakeholders; - Stay up-to-date with advancements and the latest modeling techniques in the field; - Publish your research findings in top conferences and journals. About the team Prime Video Recommendation/Search Science team owns science solution to power search experience on various devices, from sourcing, relevance, ranking, to name a few. We work closely with the engineering teams to launch our solutions in production.
US, WA, Seattle
Amazon's Price Perception and Evaluation team is seeking a driven Principal Applied Scientist to harness planet scale multi-modal datasets, and navigate a continuously evolving competitor landscape, in order to build and scale an advanced self-learning scientific price estimation and product understanding system, regularly generating fresh customer-relevant prices on billions of Amazon and Third Party Seller products worldwide. We are looking for a talented, organized, and customer-focused technical leader with a charter to derive deep neural product relationships, quantify substitution and complementarity effects, and publish trust-preserving probabilistic price ranges on all products listed on Amazon. This role requires an individual with excellent scientific modeling and system design skills, bar-raising business acumen, and an entrepreneurial spirit. We are looking for an experienced leader who is a self-starter comfortable with ambiguity, demonstrates strong attention to detail, and has the ability to work in a fast-paced and ever-changing environment. Key job responsibilities - Develop the team. Mentor a highly talented group of applied machine learning scientists & researchers. - See the big picture. Shape long term vision for Amazon's science-based competitive, perception-preserving pricing techniques - Build strong collaborations. Partner with product, engineering, and science teams within Pricing & Promotions to deploy machine learning price estimation and error correction solutions at Amazon scale - Stay informed. Establish mechanisms to stay up to date on latest scientific advancements in machine learning, neural networks, natural language processing, probabilistic forecasting, and multi-objective optimization techniques. Identify opportunities to apply them to relevant Pricing & Promotions business problems - Keep innovating for our customers. Foster an environment that promotes rapid experimentation, continuous learning, and incremental value delivery. - Deliver Impact. Develop, Deploy, and Scale Amazon's next generation foundational price estimation and understanding system
US, WA, Seattle
Here at Amazon, we embrace our differences. We are committed to furthering our culture of diversity and inclusion of our teams within the organization. How do you get items to customers quickly, cost-effectively, and—most importantly—safely, in less than an hour? And how do you do it in a way that can scale? Our teams of hundreds of scientists, engineers, aerospace professionals, and futurists have been working hard to do just that! We are delivering to customers, and are excited for what’s to come. Check out more information about Prime Air on the About Amazon blog (https://www.aboutamazon.com/news/transportation/amazon-prime-air-delivery-drone-reveal-photos). If you are seeking an iterative environment where you can drive innovation, apply state-of-the-art technologies to solve real world delivery challenges, and provide benefits to customers, Prime Air is the place for you. Come work on the Amazon Prime Air Team! We are seeking a highly skilled Navigation Scientist to help develop advanced algorithms and software for our Prime Air delivery drone program. In this role, you will conduct comprehensive navigation analysis to support cross-functional decision-making, define system architecture and requirements, contribute to the development of flight algorithms, and actively identify innovative technological opportunities that will drive significant enhancements to meet our customers' evolving demands. Export Control License: This position may require a deemed export control license for compliance with applicable laws and regulations. Placement is contingent on Amazon’s ability to apply for and obtain an export control license on your behalf.
IN, KA, Bengaluru
Alexa+ is Amazon’s next-generation, AI-powered virtual assistant. Building on the original Alexa, it uses generative AI to deliver a more conversational, personalized, and effective experience. As an Applied Scientist II on the Alexa Sensitive Content Intelligence (ASCI) team, you'll be part of an elite group developing industry-leading technologies in attribute extraction and sensitive content detection that work seamlessly across all languages and countries. In this role, you'll join a team of exceptional scientists pushing the boundaries of Natural Language Processing. Working in our dynamic, fast-paced environment, you'll develop novel algorithms and modeling techniques that advance the state of the art in NLP. Your innovations will directly shape how millions of customers interact with Amazon Echo, Echo Dot, Echo Show, and Fire TV devices every day. What makes this role exciting is the unique blend of scientific innovation and real-world impact. You'll be at the intersection of theoretical research and practical application, working alongside talented engineers and product managers to transform breakthrough ideas into customer-facing experiences. Your work will be crucial in ensuring Alexa remains at the forefront of AI technology while maintaining the highest standards of trust and safety. We're looking for a passionate innovator who combines strong technical expertise with creative problem-solving skills. Your deep understanding of NLP models (including LSTM and transformer-based architectures) will be essential in tackling complex challenges and identifying novel solutions. You'll leverage your exceptional technical knowledge, strong Computer Science fundamentals, and experience with large-scale distributed systems to create reliable, scalable, and high-performance products that delight our customers. Key job responsibilities In this dynamic role, you'll design and implement GenAI solutions that define the future of AI interaction. You'll pioneer novel algorithms, conduct ground breaking experiments, and optimize user experiences through innovative approaches to sensitive content detection and mitigation. Working alongside exceptional engineers and scientists, you'll transform theoretical breakthroughs into practical, scalable solutions that strengthen user trust in Alexa globally. You'll also have the opportunity to mentor rising talent, contributing to Amazon's culture of scientific excellence while helping build high-performing teams that deliver swift, impactful results. A day in the life Imagine starting your day collaborating with brilliant minds on advancing state-of-the-art NLP algorithms, then moving on to analyze experiment results that could reshape how Alexa understands and responds to users. You'll partner with cross-functional teams - from engineers to product managers - to ensure data quality, refine policies, and enhance model performance. Your expertise will guide technical discussions, shape roadmaps, and influence key platform features that require cross-team leadership. About the team The mission of the Alexa Sensitive Content Intelligence (ASCI) team is to (1) minimize negative surprises to customers caused by sensitive content, (2) detect and prevent potential brand-damaging interactions, and (3) build customer trust through appropriate interactions on sensitive topics. The term “sensitive content” includes within its scope a wide range of categories of content such as offensive content (e.g., hate speech, racist speech), profanity, content that is suitable only for certain age groups, politically polarizing content, and religiously polarizing content. The term “content” refers to any material that is exposed to customers by Alexa (including both 1P and 3P experiences) and includes text, speech, audio, and video.
IN, KA, Bengaluru
Alexa+ is Amazon’s next-generation, AI-powered virtual assistant. Building on the original Alexa, it uses generative AI to deliver a more conversational, personalized, and effective experience. As an Applied Scientist II on the Alexa Sensitive Content Intelligence (ASCI) team, you'll be part of an elite group developing industry-leading technologies in attribute extraction and sensitive content detection that work seamlessly across all languages and countries. In this role, you'll join a team of exceptional scientists pushing the boundaries of Natural Language Processing. Working in our dynamic, fast-paced environment, you'll develop novel algorithms and modeling techniques that advance the state of the art in NLP. Your innovations will directly shape how millions of customers interact with Amazon Echo, Echo Dot, Echo Show, and Fire TV devices every day. What makes this role exciting is the unique blend of scientific innovation and real-world impact. You'll be at the intersection of theoretical research and practical application, working alongside talented engineers and product managers to transform breakthrough ideas into customer-facing experiences. Your work will be crucial in ensuring Alexa remains at the forefront of AI technology while maintaining the highest standards of trust and safety. We're looking for a passionate innovator who combines strong technical expertise with creative problem-solving skills. Your deep understanding of NLP models (including LSTM and transformer-based architectures) will be essential in tackling complex challenges and identifying novel solutions. You'll leverage your exceptional technical knowledge, strong Computer Science fundamentals, and experience with large-scale distributed systems to create reliable, scalable, and high-performance products that delight our customers. Key job responsibilities In this dynamic role, you'll design and implement GenAI solutions that define the future of AI interaction. You'll pioneer novel algorithms, conduct ground breaking experiments, and optimize user experiences through innovative approaches to sensitive content detection and mitigation. Working alongside exceptional engineers and scientists, you'll transform theoretical breakthroughs into practical, scalable solutions that strengthen user trust in Alexa globally. You'll also have the opportunity to mentor rising talent, contributing to Amazon's culture of scientific excellence while helping build high-performing teams that deliver swift, impactful results. A day in the life Imagine starting your day collaborating with brilliant minds on advancing state-of-the-art NLP algorithms, then moving on to analyze experiment results that could reshape how Alexa understands and responds to users. You'll partner with cross-functional teams - from engineers to product managers - to ensure data quality, refine policies, and enhance model performance. Your expertise will guide technical discussions, shape roadmaps, and influence key platform features that require cross-team leadership. About the team The mission of the Alexa Sensitive Content Intelligence (ASCI) team is to (1) minimize negative surprises to customers caused by sensitive content, (2) detect and prevent potential brand-damaging interactions, and (3) build customer trust through appropriate interactions on sensitive topics. The term “sensitive content” includes within its scope a wide range of categories of content such as offensive content (e.g., hate speech, racist speech), profanity, content that is suitable only for certain age groups, politically polarizing content, and religiously polarizing content. The term “content” refers to any material that is exposed to customers by Alexa (including both 1P and 3P experiences) and includes text, speech, audio, and video.