Human Trafficking
The International Labor Organization estimates that today, more than 24.9 million people around the world are victims of human trafficking.
Credit: kgtoh

How Marinus Analytics uses knowledge graphs powered by Amazon Neptune to combat human trafficking

Traffic Jam leverages machine learning technologies from Amazon Web Services to find patterns in ads posted by sexual traffickers on the internet every day.

The International Labor Organization estimates that today, more than 24.9 million people around the world are victims of human trafficking. Nearly 20% of these victims are sexually exploited.

According to the U.S. State Department 2019 Trafficking in Persons Report, 7,481 traffickers were convicted worldwide in 2018. These numbers may appear low, but they represent an increase of 68% from 2014.

Organizations like Marinus Analytics that leverage the power of machine learning to analyze patterns in the advertisements offering sexual services on the internet are helping increase the number of convictions by providing actionable insights to law enforcement organizations.

Emily Kennedy started working on the idea that would eventually become Marinus Analytics when she was an undergraduate student at Pittsburgh’s Carnegie Mellon University (CMU). Kennedy decided to fight the scourge of human trafficking after a trip to Eastern Europe as a teenager, where she came across orphans believed to be controlled by the Russian mafia begging on the streets.

Marinus Analytics Leaders
Emily Kennedy (l), and Cara Jones are the co-founders of Marinus Analytics. The company focuses on how AI can turn big data online into actionable intelligence.
Credit: Marinus Analytics

Kennedy wanted to leverage the power of big data to help rescue victims of human trafficking. She pitched her idea to researchers at CMU’s machine learning- focused Auton Lab, who were intrigued by Kennedy’s vision. At the Auton Lab, Kennedy connected with researcher and engineer Cara Jones to make the then nascent Traffic Jam product operational.

Traffic Jam leverages machine learning technologies from Amazon Web Services to find patterns in the 300,000 plus ads, many of which are posted by sexual traffickers on the internet every day. Viswanathan’s team at AWS conducted a deep dive exploration of Traffic Jam’s data to arrive at the optimal for storage of crawled ad networks’ data in Amazon Neptune. The team also developed a knowledge graph to capture the information found in online classifieds websites, uncover underlying patterns, surface insights to investigators, and bring criminals to justice.

Today, law enforcement officials use Traffic Jam to find victims of human trafficking and dismantle organized crime rings. In 2019 alone, Traffic Jam was used to identify and rescue an estimated 3,800 victims of sex trafficking.

Small needles in especially large haystacks

Prem Viswanathan is a data scientist with AWS Professional Services, and also an adjunct professor at CMU. At CMU he had met Emily Kennedy, during one of her guest lectures, when she was working on Traffic Jam. Today, at AWS Professional Services, Viswanathan is helping organizations like Marinus Analytics harness the power of machine learning to meet their objectives.

“Identifying an ad posted by an organized crime network is challenging,” Viswanathan says. “First, most of the ads posted on the Internet don’t have structured data. To analyze information effectively, it is necessary to sift through the text of every ad to pull out relevant information like the location, date of posting, images, social media handles and other pertinent information.”

To complicate matters even more, there are millions of ads offering sexual services posted on the internet every day. A majority of these ads are placed by people who are offering these services on their own accord. Traffic Jam is focused on finding victims of human trafficking who are forced into the trade against their will.

Traffic Jam uses knowledge graphs to accomplish this objective. Knowledge graphs comprise entities or nodes. Nodes are distinct entities that hold a piece of information. For example, in Traffic Jam, each ad is represented as a distinct node, as are other criteria such as the ad location, phone number, and the month in which the ad was posted.

Traffic Jam know
Traffic Jam utilizes knowledge graphs to help find human traffickers. The knowledge graph for human trafficking contains more than 1 billion edges connecting ads, phone numbers, images, and other entities.
Credit: Marinus Analytics

Knowledge graphs also store the relationships among these different nodes. They do this in the form of edges. With the rapidly growing number of ads added to the internet every day, the knowledge graph utilized by Traffic Jam contains more than a billion edges connecting ads, phone numbers, images and other entities.

“Traffic Jam sifts through the information contained in these large number of nodes to uncover suspicious patterns,” says Viswanathan. “Consider an example of two ads that have different images, and posted from different locations, but share the same phone number. If you combine text indicators of potential human trafficking to these signals, you arrive at a movement pattern that analysts might identify as problematic, and surface to law enforcement for further review.”

AWS also developed a custom user interface using ReactJS and D3. The user interface enables investigators to visualize the patterns. The knowledge graph-based setup also enables investigators to query up to four times more information than previously feasible, while performing their analysis. This allows them to find prior ads more easily, where a member of a human trafficking network might have used a real phone number or revealed other identifying information.

Deep Graph Learning – an area ripe for innovation

George Karypis is a professor within the Department of Computer Science & Engineering at the University of Minnesota. In the course of his career, Karypis has focused on a variety of areas related to big data including data mining, recommender systems, and high-performance computing. Karypis joined Amazon in 2019 as an Amazon Scholar—a select group of academic professionals that work on large-scale technical challenges while continuing to teach and conduct research at their universities. "The opportunity to help organizations like Marinus Analytics to harness the power of big data, and have a real-world impact is deeply meaningful to me," Karypis said.

George Karypis
Amazon Scholar George Karypis is a professor at the University of Minnesota.

At Amazon, Karypis’ team is focused on unlocking innovations that drive efficient and scalable deep learning on knowledge graphs. The team has been responsible for developing the Deep Graph Library (DGL), an easy-to-use, high performance and scalable Python package for deep learning on graphs. DGL is a framework that allows developers to program a class of machine learning models called graph neural networks (GNN). DGL supplements existing tensor-based frameworks such as Tensorflow, PyTorch, and MXNet to support the growing area of deep graph learning.

The adoption of GNNs has exploded in recent years, as data scientists move beyond developing deep learning models for 2D signals (such as images) and 3D signals (such as video) to learning from structured, related data embedded in graphs.

Today, GNNs are used in a number of fields. For example, they play an increasingly important role in social networks, where graphs show connections among related people. At Amazon, they are used to develop recommender systems, build mechanisms for fraud and abuse detection and develop Alexa chatbots among other applications.

Organizations like Marinus Analytics use GNNs to contrast information between different nodes, and surface interesting insights, such as whether a particular ad has characteristics common with ads posted by organized crime rings.

For Karypis, GNNs represent one of the most exciting areas in the world of machine learning. More specifically, he believes there are three areas in the world of deep graph learning that are particularly ripe for innovation.

“At the most basic level, there are multiple experiments that are trying to determine the best way to express machine learning models in deep graph learning,” says Karypis. “What are the right models? What are the most appropriate abstractions?”

The integration with Amazon Neptune has been a game changer for Traffic Jam
Cara Jones, CEO, Marinus Analytics

The second challenge pertains to the training of these models. GNN training requires irregular memory accesses. In addition, the training involves fewer operations for each word of memory that it accesses and is computationally demanding. Moreover, knowledge graphs such as the one used by Traffic Jam have billions of data points. “In order to realize the benefits afforded by GNNs, it is critical to develop efficient and scalable distributed GNN training approaches for large graphs,” says Karypis.

Finally, Karypis and his team are intrigued by the most effective ways to compute knowledge graph embeddings. This involves embedding both the entities of a graph and underlying relations in a vector form in a d-dimensional space. For Traffic Jam, representing nodes and their relations in a vector form is what enables the comparison of different ad networks, each of which is represented as a sub-graph.

“Language modelling is a very well understood problem, as are various facets related to computer vision,” he says. “However, it’s still early days when it comes to GNNs, and I’m excited to be at AWS where a lot of the innovation is happening.”

Traffic Jam’s new offerings that use Amazon Neptune and advanced ML techniques to track different ad networks and analyze their likelihood of belonging to an existing crime group is currently in beta. The new features are expected to be made generally available to users soon.

“The integration with Amazon Neptune has been a game changer for Traffic Jam,” says Cara Jones, CEO and co-founder of Marinus Analytics. “Using the knowledge graph and associated sub-graphs, we are now able to capture four times as much information as previously possible. More importantly, we are able to analyze data and identify potential crime groups in real-time, even as new information comes in.”

Research areas

Related content

US, WA, Seattle
Here at Amazon, we embrace our differences. We are committed to furthering our culture of diversity and inclusion of our teams within the organization. How do you get items to customers quickly, cost-effectively, and—most importantly—safely, in less than an hour? And how do you do it in a way that can scale? Our teams of hundreds of scientists, engineers, aerospace professionals, and futurists have been working hard to do just that! We are delivering to customers, and are excited for what’s to come. Check out more information about Prime Air on the About Amazon blog (https://www.aboutamazon.com/news/transportation/amazon-prime-air-delivery-drone-reveal-photos). If you are seeking an iterative environment where you can drive innovation, apply state-of-the-art technologies to solve real world delivery challenges, and provide benefits to customers, Prime Air is the place for you. Come work on the Amazon Prime Air Team! Prime Air is seeking an experienced Applied Science Manager to help develop our advanced Navigation algorithms and flight software applications. In this role, you will lead a team of scientists and engineers to conduct analyses, support cross-functional decision-making, define system architectures and requirements, contribute to the development of flight algorithms, and actively identify innovative technological opportunities that will drive significant enhancements to meet our customers' evolving demands. This person must be comfortable working with a team of top-notch software developers and collaborating with our science teams. We’re looking for someone who innovates, and loves solving hard problems. You will work hard, have fun, and make history! Export Control License: This position may require a deemed export control license for compliance with applicable laws and regulations. Placement is contingent on Amazon’s ability to apply for and obtain an export control license on your behalf.
US, VA, Herndon
Application deadline: Applications will be accepted on an ongoing basis Are you excited to help the US Intelligence Community design, build, and implement AI algorithms, including advanced Generative AI solutions, to augment decision making while meeting the highest standards for reliability, transparency, and scalability? The Amazon Web Services (AWS) US Federal Professional Services team works directly with US Intelligence Community agencies and other public sector entities to achieve their mission goals through the adoption of Machine Learning (ML) and Generative AI methods. We build models for text, image, video, audio, and multi-modal use cases, leveraging both traditional ML approaches and state-of-the-art generative models including Large Language Models (LLMs), text-to-image generation, and other advanced AI capabilities to fit the mission. Our team collaborates across the entire AWS organization to bring access to product and service teams, to get the right solution delivered and drive feature innovation based on customer needs. At AWS, we're hiring experienced data scientists with a background in both traditional and generative AI who can help our customers understand the opportunities their data presents, and build solutions that earn the customer trust needed for deployment to production systems. In this role, you will work closely with customers to deeply understand their data challenges and requirements, and design tailored solutions that best fit their use cases. You should have broad experience building models using all kinds of data sources, and building data-intensive applications at scale. You should possess excellent business acumen and communication skills to collaborate effectively with stakeholders, develop key business questions, and translate requirements into actionable solutions. You will provide guidance and support to other engineers, sharing industry best practices and driving innovation in the field of data science and AI. This position requires that the candidate selected must currently possess and maintain an active TS/SCI Security Clearance with Polygraph. The position further requires the candidate to opt into a commensurate clearance for each government agency for which they perform AWS work. Key job responsibilities As an Data Scientist, you will: - Collaborate with AI/ML scientists and architects to research, design, develop, and evaluate AI algorithms to address real-world challenges - Interact with customers directly to understand the business problem, help and aid them in implementation of AI solutions, deliver briefing and deep dive sessions to customers and guide customer on adoption patterns and paths to production. - Create and deliver best practice recommendations, tutorials, blog posts, sample code, and presentations adapted to technical, business, and executive stakeholder - Provide customer and market feedback to Product and Engineering teams to help define product direction - This position may require up to 25% local travel. About the team About AWS Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences and inspire us to never stop embracing our uniqueness. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why flexible work hours and arrangements are part of our culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud.
US, TX, Austin
Our team is involved with pre-silicon design verification for custom IP. A critical requirement of the verification flow is the requirement of legal and realistic stimulus of a custom Machine Learning Accelerator Chip. Content creation is built using formal methods that model legal behavior of the design and then solving the problem to create the specific assembly tests. The entire frame work for creating these custom tests is developed using a SMT solver and custom software code to guide the solution space into templated scenarios. This highly visible and innovative role requires the design of this solving framework and collaborating with design verification engineers, hardware architects and designers to ensure that interesting content can be created for the projects needs. Key job responsibilities Develop an understanding for a custom machine learning instruction set architecture. Model correctness of instruction streams using first order logic. Create custom API's to allow control over scheduling and randomness. Deploy algorithms to ensure concurrent code is safely constructed. Create coverage metrics to ensure solution space coverage. Use novel methods like machine learning to automate content creation. About the team Utility Computing (UC) AWS Utility Computing (UC) provides product innovations — from foundational services such as Amazon’s Simple Storage Service (S3) and Amazon Elastic Compute Cloud (EC2), to consistently released new product innovations that continue to set AWS’s services and features apart in the industry. As a member of the UC organization, you’ll support the development and management of Compute, Database, Storage, Internet of Things (Iot), Platform, and Productivity Apps services in AWS, including support for customers who require specialized security solutions for customers who require specialized security solutions for their cloud services. Annapurna Labs (our organization within AWS UC) designs silicon and software that accelerates innovation. Customers choose us to create cloud solutions that solve challenges that were unimaginable a short time ago—even yesterday. Our custom chips, accelerators, and software stacks enable us to take on technical challenges that have never been seen before, and deliver results that help our customers change the world. About AWS Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why we strive for flexibility as part of our working culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon conferences, inspire us to never stop embracing our uniqueness. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional.
CN, 11, Beijing
职位:Applied scientist 应用科学家实习生 毕业时间:2026年10月 - 2027年7月之间毕业的应届毕业生 · 入职日期:2026年6月及之前 · 实习时间:保证一周实习4-5天全职实习,至少持续3个月 · 工作地点:北京朝阳区 投递须知: 1 填写简历申请时,请把必填和非必填项都填写完整。提交简历之后就无法修改了哦! 2 学校的英文全称请准确填写。中英文对应表请查这里(无法浏览请登录后浏览)https://docs.qq.com/sheet/DVmdaa1BCV0RBbnlR?tab=BB08J2 如果您正在攻读计算机,AI,ML或搜索领域专业的博士或硕士研究生,而且对应用科学家的实习工作感兴趣。如果您也喜爱深入研究棘手的技术问题并提出解决方案,用成功的产品显著地改善人们的生活。 那么,我们诚挚邀请您加入亚马逊的International Technology搜索团队改善Amazon的产品搜索服务。我们的目标是帮助亚马逊的客户找到他们所需的产品,并发现他们感兴趣的新产品。 这会是一份收获满满的工作。您每天的工作都与全球数百万亚马逊客户的体验紧密相关。您将提出和探索创新,基于TB级别的产品和流量数据设计机器学习模型。您将集成这些模型到搜索引擎中为客户提供服务,通过数据,建模和客户反馈来完成闭环。您对模型的选择需要能够平衡业务指标和响应时间的需求。
CN, 44, Shenzhen
职位:Applied scientist 应用科学家实习生 毕业时间:2026年10月 - 2027年7月之间毕业的应届毕业生 · 入职日期:2026年6月及之前 · 实习时间:保证一周实习4-5天全职实习,至少持续3个月 · 工作地点:深圳福田区 投递须知: 1 填写简历申请时,请把必填和非必填项都填写完整。提交简历之后就无法修改了哦! 2 学校的英文全称请准确填写。中英文对应表请查这里(无法浏览请登录后浏览)https://docs.qq.com/sheet/DVmdaa1BCV0RBbnlR?tab=BB08J2 如果您正在攻读计算机,AI,ML领域专业的博士或硕士研究生,而且对应用科学家的实习工作感兴趣。如果您也喜爱深入研究棘手的技术问题并提出解决方案,用成功的产品显著地改善人们的生活。 那么,我们诚挚邀请您加入亚马逊。这会是一份收获满满的工作。您每天的工作都与全球数百万亚马逊客户的体验紧密相关。您将提出和探索创新,基于TB级别的产品和流量数据设计机器学习模型。您将集成这些为客户提供服务,通过数据,建模和客户反馈来完成闭环。您对模型的选择需要能够平衡业务指标和响应时间的需求。
LU, Luxembourg
Join our team as an Applied Scientist II where you'll develop innovative machine learning solutions that directly impact millions of customers. You'll work on ambiguous problems where neither the problem nor solution is well-defined, inventing novel scientific approaches to address customer needs at the project level. This role combines deep scientific expertise with hands-on implementation to deliver production-ready solutions that drive measurable business outcomes. Key job responsibilities Invent: - Design and develop novel machine learning models and algorithms to solve ambiguous customer problems where textbook solutions don't exist - Extend state-of-the-art scientific techniques and invent new approaches driven by customer needs at the project level - Produce internal research reports with the rigor of top-tier publications, documenting scientific findings and methodologies - Stay current with academic literature and research trends, applying latest techniques when appropriate Implement: - Write production-quality code that meets or exceeds SDE I standards, ensuring solutions are testable, maintainable, and scalable - Deploy components directly into production systems supporting large-scale applications and services - Optimize algorithm and model performance through rigorous testing and iterative improvements - Document design decisions and implementation details to enable reproducibility and knowledge transfer - Contribute to operational excellence by analyzing performance gaps and proposing solutions Influence: - Collaborate with cross-functional teams to translate business goals into scientific problems and metrics - Mentor junior scientists and help new teammates understand customer needs and technical solutions - Present findings and recommendations to both technical and non-technical stakeholders - Contribute to team roadmaps, priorities, and strategic planning discussions - Participate in hiring and interviewing to build world-class science teams
US, CA, Sunnyvale
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Applied Scientist with a strong deep learning background, to build Generative Artificial Intelligence (GenAI) technology with Large Language Models (LLMs) and multimodal systems. Key job responsibilities As an Applied Scientist with the AGI team, you will work with talented peers to support the development of GenAI algorithms and modeling techniques, to advance the state of the art with LLMs. Your work will directly impact our customers in the form of products and services that make use of speech and language technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate advances in GenAI. About the team The AGI team has a mission to push the envelope with GenAI in LLMs and multimodal systems, in order to provide the best-possible experience for our customers.
US, CA, East Palo Alto
Amazon Aurora DSQL is a serverless, distributed SQL database with virtually unlimited scale, highest availability, and zero infrastructure management. Aurora DSQL provides active-active high availability, providing strong data consistency designed for 99.99% single-Region and 99.999% multi-Region availability. Aurora DSQL automatically manages and scales system resources, so you don't have to worry about maintenance downtime and provisioning, patching, or upgrading infrastructure. As a Senior Applied Scientist, you will be expected to lead research and development in advanced query optimization techniques for distributed sql services. You will innovate in the query planning and execution layer to help Aurora DSQL succeed at delivering high performance for complex OLTP workloads. You will develop novel approaches to stats collection, query planning, execution and optimization. You will drive industry leading research, publish your research and help convert your research into implementations to make Aurora DSQL the fastest sql database for OLTP workloads. AWS Utility Computing (UC) provides product innovations — from foundational services such as Amazon’s Simple Storage Service (S3) and Amazon Elastic Compute Cloud (EC2), to consistently released new product innovations that continue to set AWS’s services and features apart in the industry. As a member of the UC organization, you’ll support the development and management of Compute, Database, Storage, Internet of Things (Iot), Platform, and Productivity Apps services in AWS, including support for customers who require specialized security solutions for their cloud services. Key job responsibilities Our engineers collaborate across diverse teams, projects, and environments to have a firsthand impact on our global customer base. You’ll bring a passion for innovation, data, search, analytics, and distributed systems. You’ll also: Solve challenging technical problems, often ones not solved before, at every layer of the stack. Design, implement, test, deploy and maintain innovative software solutions to transform service performance, durability, cost, and security. Build high-quality, highly available, always-on products. Research implementations that deliver the best possible experiences for customers. A day in the life As you design and code solutions to help our team drive efficiencies in software architecture, you’ll create metrics, implement automation and other improvements, and resolve the root cause of software defects. You’ll also: Build high-impact solutions to deliver to our large customer base. Participate in design discussions, code review, and communicate with internal and external stakeholders. Work cross-functionally to help drive business decisions with your technical input. Work in a startup-like development environment, where you’re always working on the most important stuff. About the team Our team is dedicated to supporting new members. We have a broad mix of experience levels and tenures, and we’re building an environment that celebrates knowledge-sharing and mentorship. Our senior members enjoy one-on-one mentoring and thorough, but kind, code reviews. We care about your career growth and strive to assign projects that help our team members develop your engineering expertise so you feel empowered to take on more complex tasks in the future. Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. About AWS Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon conferences, inspire us to never stop embracing our uniqueness. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why we strive for flexibility as part of our working culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional.
US, CA, Sunnyvale
The Region Flexibility Engineering (RFE) team builds and leverages foundational infrastructure capabilities, tools, and datasets needed to support the rapid global expansion of Amazon's SOA infrastructure. Our team focuses on robust and scalable architecture patterns and engineering best practices, driving adoption of ever-evolving and AWS technologies. RFE is looking for a passionate, results-oriented, inventive Data Scientist to refine and execute experiments towards our grand vision, influence and implement technical solutions for regional placement automation, cross-region libraries, and tooling useful for teams across Amazon. As a Data Scientist in Region Flexibility, you will work to enable Amazon businesses to leverage new AWS regions and improve the efficiency and scale of our business. Our project spans across all of Amazon Stores, Digital and Others (SDO) Businesses and we work closely with AWS teams to advise them on SDO requirements. As innovators who embrace new technology, you will be empowered to choose the right highly scalable and available technology to solve complex problems and will directly influence product design. The end-state architecture will enable services to break region coupling while retaining the ability to keep critical business functions within a region. This architecture will improve customer latency through local affinity to compute resources and reduce the blast radius in case of region failures. We leverage off the sciences of data, information processing, machine learning, and generative AI to improve user experience, automation, service resilience, and operational efficiency. Key job responsibilities As an RFE Data Scientist, you will work closely with product and technical leaders throughout Amazon and will be responsible for influencing technical decisions and building data-driven automation capabilities in areas of development/modeling that you identify as critical future region flexibility offerings. You will identify both enablers and blockers of adoption for region flex, and build models to raise the bar in terms of understanding questions related to data set and service relationships and predict the impact of region changes and provide offerings to mitigate that impact. About the team The Regional Flexibility Engineering (RFE) organization supports the rapid global expansion of Amazon's infrastructure. Our projects support Amazon businesses like Stores, Alexa, Kindle, and Prime Video. We drive adoption of ever-evolving and AWS and non-AWS technologies, and work closely with AWS teams to improve AWS public offerings. Our organization focuses on robust and scalable solutions, simple to use, and delivered with engineering best practices. We leverage and build foundational infrastructure capabilities, tools, and datasets that enable Amazon teams to delight our customers. With millions of people using Amazon’s products every day, we appreciate the importance of making our solutions “just work”.
US, WA, Seattle
Amazon Prime is looking for an ambitious Economist to help create econometric insights for world-wide Prime. Prime is Amazon's premiere membership program, with over 200M members world-wide. This role is at the center of many major company decisions that impact Amazon's customers. These decisions span a variety of industries, each reflecting the diversity of Prime benefits. These range from fast-free e-commerce shipping, digital content (e.g., exclusive streaming video, music, gaming, photos), and grocery offerings. Prime Science creates insights that power these decisions. As an economist in this role, you will create statistical tools that embed causal interpretations. You will utilize massive data, state-of-the-art scientific computing, econometrics (causal, counterfactual/structural, time-series forecasting, experimentation), and machine-learning, to do so. Some of the science you create will be publishable in internal or external scientific journals and conferences. You will work closely with a team of economists, applied scientists, data professionals (business analysts, business intelligence engineers), product managers, and software engineers. You will create insights from descriptive statistics, as well as from novel statistical and econometric models. You will create internal-to-Amazon-facing automated scientific data products to power company decisions. You will write strategic documents explaining how senior company leaders should utilize these insights to create sustainable value for customers. These leaders will often include the senior-most leaders at Amazon. The team is unique in its exposure to company-wide strategies as well as senior leadership. It operates at the research frontier of utilizing data, econometrics, artificial intelligence, and machine-learning to form business strategies. A successful candidate will have demonstrated a capacity for building, estimating, and defending statistical models (e.g., causal, counterfactual, time-series, machine-learning) using software such as R, Python, or STATA. They will have a willingness to learn and apply a broad set of statistical and computational techniques to supplement deep-training in one area of econometrics. For example, many applications on the team use structural econometrics, machine-learning, and time-series forecasting. They rely on building scalable production software, which involves a broad set of world-class software-building skills often learned on-the-job. As a consequence, already-obtained knowledge of SQL, machine learning, and large-scale scientific computing using distributed computing infrastructures such as Spark-Scala or PySpark would be a plus. Additionally, this candidate will show a track-record of delivering projects well and on-time, preferably in collaboration with other team members (e.g. co-authors). Candidates must have very strong writing and emotional intelligence skills (for collaborative teamwork, often with colleagues in different functional roles), a growth mindset, and a capacity for dealing with a high-level of ambiguity. Endowed with these traits and on-the-job-growth, the role will provide the opportunity to have a large strategic, world-wide impact on the customer experiences of Prime members.