The science behind Amazon’s spatial audio-processing technology

Combining psychoacoustics, signal processing, and speaker beamforming enhances stereo audio and delivers an immersive sound experience for customers.

With every new Echo device and upgrade, we challenge ourselves to bring the best audio experience to our customers at an affordable price. This year, we’re introducing Amazon’s own custom-built spatial audio-processing technology, designed to enhance stereo sound on compatible Echo devices.

The version of the technology on Echo Studio, for instance, is customized to the specific acoustic design of the speakers and employs digital-processing methods — such as upmixing and virtualization — so stereo audio, television shows, and movie soundtracks feel closer to the listener, with greater width, clarity, and presence. It turns the Echo Studio into a hi-fi audio system that mirrors that of a stereo reference arrangement. Vocal performances are more present in the center soundstage, and stereo panned instruments are better defined on the sides, thereby creating a more immersive sound experience that reproduces the artist's intent.

In this blog post, we break down how we built this spatial audio-processing technology with an emphasis on the way humans perceive sound — or psychoacoustics — by using a combination of crosstalk cancellation, speaker beamforming, and upmixing to create a room-filling, spatial audio experience.

Psychoacoustics: Width, depth, and listening zones

Throughout development, we characterize the stereo image by its psychoacoustic qualities, including width, depth, and listening zones. We then investigate how sound waves interact with listeners in various room shapes and sizes and how signal-processing methods affect the listener’s experience.

Stereo angle.png
Echo Studio virtualizes the stereo sound field at the listener’s location in the far field.

Width

Width: The angular extent (wide vs. narrow) of localizable elements in the stereo image along the horizontal — or azimuth — plane.

When determining the width of a sound field, we first consider localizable elements such as a point-source that would induce time and level differences in the acoustic responses at the listener’s two ears. To model this phenomenon, it is helpful to compare the listening experiences on headphones vs. a loudspeaker in terms of the separation of left and right ear responses.

Unlike loudspeaker listening, headphone listening lacks a crosstalk path, as illustrated in the image below. In order to make headphone listening realistic, we can model crosstalk from the point-source to the two ears using an all-pass signal-processing filter for one ear and a delayed low-pass filter for the other ear. The two filters approximate and parameterize the listener’s ear responses with respect to their relative head-related transfer functions (HRTFs), which contain important cues that the human ear uses to localize sound. Moreover, the filter design ensures that there’s minimal modification to the signal spectra — or tonal balance — and therefore preserves the original playback content.

Crosstalk simulation.png
All-pass and delayed low-pass filters approximate the angle-dependent relative ipsilateral (same side of the body) and contralateral (opposite side of the body) head-related transfer functions (HRTFs).

However, unlike headphones, an external speaker can create its own crosstalk for the listener, depending on its placement. For example, the left and right speaker transducers, or drivers, on the Echo Studio are narrowly spaced within the device, whereas the speakers in a standard stereo pair are 60 degrees apart relative to the listener.

With the spatial audio-processing technology on Echo Studio, we decouple the crosstalk of the driver pair by modeling and then inverting the system of equations between each driver and the listener’s ears, via crosstalk cancellation (CTC) methods. If we have more than two drivers, then the more general formulation is called null-steering, where filters are designed for all the drivers so that their acoustic responses cancel at one ear.

In both cases, we can normalize the filter design to satisfy a target cancellation gain curve defined by the power ratio of the acoustic energy at the ipsilateral (same side of the body) and contralateral (opposite side of the body) ears across frequencies. This prevents overfitting the cancellation to an exact location, since a listener may be at varying distances or not perfectly centered to the device.

Once the driver’s CTC filters are designed for stereo inputs, they can be combined with the approximated HRTF filters that introduce the amount of crosstalk consistent with a stereo reference system.

CTC filters.png
Stereo virtualization for external speaker playback specifies an additional pair of crosstalk cancellation (CTC) filters for nulling the contralateral acoustic response. The relative transfer function (RTF) filter realizes the ratio of the two CTC filter responses.

Depth

Depth: The distance (frontal vs. recessed) of the perceived sound field from the listener.

The distance at which sound elements in an audio track localize correlates with the relationship — or coherence — of the two signals between the sound source and the listener’s ears. For example, a simple left or right signal from a speaker is easy to understand, but if the audio mixes with the room’s reverberation, the audio clarity decreases, and the audio sounds recessed.

In speaker playback, however, we contend with the speaker directivity and its interaction with the room environment. For example, a direct acoustic path between a speaker and a listener preserves the desired clarity of the original content. But when the acoustic signal reflects off of walls, the loss in coherence recesses the perceived sound field and causes elements to smear spatially. This is why tracks heard anechoically or on headphones appear closer — or even inside the listener’s head — and clearer than tracks heard over external speakers in a reverberant room. In the first case, the acoustic response is direct from the driver to the listener’s ears, while external speakers must contend with the effects of the room environment.

Beamformer impact.png
Strong room reflections and reverberation mask the binaural cues and reduce the perceived distance of the soundstage. Speaker beamforming pushes the soundstage forward by attenuating the indirect sound energy, increasing the critical distance and coherence.

As part of our custom-built spatial audio technology, we can control the speaker directivity via careful beamforming. The speaker drivers can be filtered to produce a sound field with a directivity that sums coherently on-axis and cancels off-axis. That is, the acoustic response is greatest when the listener is lined up in front of the speaker and, conversely, weakest when the listener is to the side at +/- 90 degrees.

Therefore, one way to design with such directivity is to place two nulls at +/- 90-degree angles and either control for the cancellation gain between on-/off-axis power responses or the shape of the nulls as a function of azimuth. The resulting beam pattern is one with a main lobe that is wide enough for the direct path to be strong, at up to a +/- 45-degree azimuth listening window, before quickly tapering off to minimize the acoustic energy further off-axis, which would reflect off the walls.

This has the intended effect of making stereo audio feel closer to the listener, with greater clarity than is typical in an acoustically untreated listening environment like a living room. The effect is similar to how theaters reproduce a frontal soundstage over different seating areas, despite the speakers’ being far away.

Beamforming.png
The speaker beamformer increases directivity after placing two off-axis nulls in the midrange frequencies. The acoustic responses over frequency and azimuth contrast that of simple matrix mixing with the beamformer realized in relative-transfer-function (RTF) form.

Listening zones

Listening zone: The mapping between the listening area and the stereo soundstage.

A listening “sweet spot” — the stereo image in a hi-fi audio system reference stereo pair — is best reproduced when the listener’s location forms an equilateral triangle with the stereo speaker pair. If the listener angle exceeds +/- 30 degrees, then a hole is created in the listener’s phantom center due to the loss of inter-speaker-to-ear coherence as room reflections grow stronger. Important elements of the audio mix, such as vocals, lose their presence. If the listener angle falls below +/- 30 degrees, then the stereo image narrows, as audio elements collapse toward the center. If the listener’s location is off-axis, then the stereo image biases towards one side or the other.

Phantom center.png
The stereo field relies on a “phantom center”, where important lead vocals and instruments are mixed. The center content can be separated from the original stereo left and right input after the mid-/side decomposition.

To combat this, our spatial audio technology aims to reproduce the stereo image over the largest listening area. In practice, the intended listening area of CTC-filtered playback conflicts with that of beamforming designs that control for speaker directivity. We can achieve a compromise by performing stereo upmixing and then applying different beamforming filters to each channel. For example, we can upmix into left, right, and center (LRC), where the center is minimally correlated with left-minus-right in the mid-/side decomposition.

The upmixed left channel is processed through the CTC filter that nulls the right ear after virtualization, the upmixed right channel nulls the left ear, and the center channel is beamformed with a wide main lobe. This means that vocal performances are more present in the center, while the stereo panned instruments are better defined on the side, creating a more immersive sound experience for the listener.

Signal flow.png
After upmixing, the virtualization and the crosstalk cancellation (CTC) widens the left and right channels, and the midrange beamformer pushes the center content forward. Subsequent delay blocks phase-align the faster of the two paths.

We’re continuing to iterate and refine technology across the Echo portfolio to bring the best audio experience to our customers. If you’d like to learn more about beamforming and speaker directivity in room acoustics, read papers published by our engineering team: “Fast source-room-receiver modeling”, in EUSIPCO 2020, and “Spherical harmonic beamformer designs", in EURASIP 2021.

Research areas

Related content

US, WA, Seattle
Amazon is looking for a passionate, talented, and inventive Senior Applied Scientist with a strong machine learning background to help build industry-leading language technology. Our mission is to provide a delightful experience to Amazon’s customers by pushing the envelope in Natural Language Processing (NLP), Generative AI, Large Language Model (LLM), Natural Language Understanding (NLU), Machine Learning (ML), Retrieval-Augmented Generation, Responsible AI, Agent, Evaluation, and Model Adaptation. As part of our AI team in Amazon AWS, you will work alongside internationally recognized experts to develop novel algorithms and modeling techniques to advance the state-of-the-art in human language technology. Your work will directly impact millions of our customers in the form of products and services, as well as contributing to the wider research community. You will gain hands on experience with Amazon’s heterogeneous text and structured data sources, and large-scale computing resources to accelerate advances in language understanding. The Science team at AWS Bedrock builds science foundations of Bedrock, which is a fully managed service that makes high-performing foundation models available for use through a unified API. We are adamant about continuously learning state-of-the-art NLP/ML/LLM technology and exploring creative ways to delight our customers. In our daily job we are exposed to large scale NLP needs and we apply rigorous research methods to respond to them with efficient and scalable innovative solutions. At AWS Bedrock, you’ll experience the benefits of working in a dynamic, entrepreneurial environment, while leveraging AWS resources, one of the world’s leading cloud companies and you’ll be able to publish your work in top tier conferences and journals. We are building a brand new team to help develop a new NLP service for AWS. You will have the opportunity to conduct novel research and influence the science roadmap and direction of the team. Come join this greenfield opportunity! Amazon Bedrock team is part of Utility Computing (UC) About the team AWS Utility Computing (UC) provides product innovations — from foundational services such as Amazon’s Simple Storage Service (S3) and Amazon Elastic Compute Cloud (EC2), to consistently released new product innovations that continue to set AWS’s services and features apart in the industry. As a member of the UC organization, you’ll support the development and management of Compute, Database, Storage, Internet of Things (Iot), Platform, and Productivity Apps services in AWS, including support for customers who require specialized security solutions for their cloud services. Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences, inspire us to never stop embracing our uniqueness. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why we strive for flexibility as part of our working culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Hybrid Work We value innovation and recognize this sometimes requires uninterrupted time to focus on a build. We also value in-person collaboration and time spent face-to-face. Our team affords employees options to work in the office every day or in a flexible, hybrid work model near one of our U.S. Amazon offices. We are open to hiring candidates to work out of one of the following locations: Seattle, WA, USA
US, WA, Seattle
Alexa Personality Fundamentals is chartered with infusing Alexa's trustworthy, reliable, considerate, smart, and playful personality. Come join us in creating the future of personality forward AI here at Alexa. Key job responsibilities As a Data Scientist with Alexa Personality, your work will involve machine learning, Large Language Model (LLM) and other generative technologies. You will partner with engineers, applied scientists, voice designers, and quality assurance to ensure that Alexa can sing, joke, and delight our customers in every interaction. You will take a central role in defining our experimental roadmap, sourcing training data, authoring annotation criteria and building automated benchmarks to track the improvement of our Alexa's personality. We are open to hiring candidates to work out of one of the following locations: Bellevue, WA, USA | Seattle, WA, USA
US, CA, Palo Alto
The Amazon Search Mission Understanding (SMU) team is at the forefront of revolutionizing the online shopping experience through the Amazon search page. Our ambition extends beyond facilitating a seamless shopping journey; we are committed to creating the next generation of intelligent shopping assistants. Leveraging cutting-edge Large Language Models (LLMs), we aim to redefine navigation and decision-making in e-commerce by deeply understanding our users' shopping missions, preferences, and goals. By developing responsive and scalable solutions, we not only accomplish the shopping mission but also foster unparalleled trust among our customers. Through our advanced technology, we generate valuable insights, providing a guided navigation system into various search missions, ensuring a comprehensive and holistic shopping experience. Our dedication to continuous improvement through constant measurement and enhancement of the shopper experience is crucial, as we strategically navigate the balance between immediate results and long-term business growth. We are seeking an Applied Scientist who is not just adept in the theoretical aspects of Machine Learning (ML), Artificial Intelligence (AI), and Large Language Models (LLMs) but also possesses a pragmatic, hands-on approach to navigating the complexities of innovation. The ideal candidate will have a profound expertise in developing, deploying, and contributing to the next-generation shopping search engine, including but not limited to Retrieval-Augmented Generation (RAG) models, specifically tailored towards enhancing the Rufus application—an integral part of our mission to revolutionize shopping assistance. You will take the lead in conceptualizing, building, and launching groundbreaking models that significantly improve our understanding of and capabilities in enhancing the search experience. A successful applicant will display a comprehensive skill set across machine learning model development, implementation, and optimization. This includes a strong foundation in data management, software engineering best practices, and a keen awareness of the latest developments in distributed systems technology. We are looking for individuals who are determined, analytically rigorous, passionate about applied sciences, creative, and possess strong logical reasoning abilities. Join the Search Mission Understanding team, a group of pioneering ML scientists and engineers dedicated to building core ML models and developing the infrastructure for model innovation. As part of Amazon Search, you will experience the dynamic, innovative culture of a startup, backed by the extensive resources of Amazon.com (AMZN), a global leader in internet services. Our collaborative, customer-centric work environment spans across our offices in Palo Alto, CA, and Seattle, WA, offering a unique blend of opportunities for professional growth and innovation. Key job responsibilities Collaborate with cross-functional teams to identify requirements for ML model development, focusing on enhancing mission understanding through innovative AI techniques, including retrieval-Augmented Generation or LLM in general. Design and implement scalable ML models capable of processing and analyzing large datasets to improve search and shopping experiences. Must have a strong background in machine learning, AI, or computational sciences. Lead the management and experiments of ML models at scale, applying advanced ML techniques to optimize science solution. Serve as a technical lead and liaison for ML projects, facilitating collaboration across teams and addressing technical challenges. Requires strong leadership and communication skills, with a PhD in Computer Science, Machine Learning, or a related field. We are open to hiring candidates to work out of one of the following locations: Palo Alto, CA, USA | Seattle, WA, USA
US, WA, Bellevue
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Applied Science Manager with a strong deep learning background, to lead the development of industry-leading technology with multimodal systems. Key job responsibilities As an Applied Science Manager with the AGI team, you will lead the development of novel algorithms and modeling techniques to advance the state of the art with multimodal systems. Your work will directly impact our customers in the form of products and services that make use of vision and language technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate development with multimodal Large Language Models (LLMs) and Generative Artificial Intelligence (GenAI) in Computer Vision. About the team The AGI team has a mission to push the envelope with multimodal LLMs and GenAI in Computer Vision, in order to provide the best-possible experience for our customers. We are open to hiring candidates to work out of one of the following locations: Bellevue, WA, USA | Seattle, WA, USA | Sunnyvale, CA, USA
US, MA, Boston
The Artificial General Intelligence (AGI) - Automations team is developing AI technologies to automate workflows, processes for browser automation, developers and ops teams. As part of this, we are developing services and inference engine for these automation agents, and techniques for reasoning, planning, and modeling workflows. If you are interested in a startup mode team in Amazon to build the next level of agents then come join us. Scientists in AGI - Automations will develop cutting edge multimodal LLMs to observe, model and derive insights from manual workflows to automate them. You will get to work in a joint scrum with engineers for rapid invention, develop cutting edge automation agent systems, and take them to launch for millions of customers. Key job responsibilities - Build automation agents by developing novel multimodal LLMs. A day in the life An Applied Scientist with the AGI team will support the science solution design, run experiments, research new algorithms, and find new ways of optimizing the customer experience.; while setting examples for the team on good science practice and standards. Besides theoretical analysis and innovation, an Applied Scientist will also work closely with talented engineers and scientists to put algorithms and models into practice. We are open to hiring candidates to work out of one of the following locations: Boston, MA, USA
US, MA, Boston
The Artificial General Intelligence (AGI) - Automations team is developing AI technologies to automate workflows, processes for browser automation, developers and ops teams. As part of this, we are developing services and inference engine for these automation agents, and techniques for reasoning, planning, and modeling workflows. If you are interested in a startup mode team in Amazon to build the next level of agents then come join us. Scientists in AGI - Automations will develop cutting edge multimodal LLMs to observe, model and derive insights from manual workflows to automate them. You will get to work in a joint scrum with engineers for rapid invention, develop cutting edge automation agent systems, and take them to launch for millions of customers. Key job responsibilities - Build automation agents by developing novel multimodal LLMs. A day in the life An Applied Scientist with the AGI team will support the science solution design, run experiments, research new algorithms, and find new ways of optimizing the customer experience.; while setting examples for the team on good science practice and standards. Besides theoretical analysis and innovation, an Applied Scientist will also work closely with talented engineers and scientists to put algorithms and models into practice. We are open to hiring candidates to work out of one of the following locations: Boston, MA, USA
US, CA, Palo Alto
The Amazon Search Mission Understanding (SMU) team is at the forefront of revolutionizing the online shopping experience through the Amazon search page. Our ambition extends beyond facilitating a seamless shopping journey; we are committed to creating the next generation of intelligent shopping assistants. Leveraging cutting-edge Large Language Models (LLMs), we aim to redefine navigation and decision-making in e-commerce by deeply understanding our users' shopping missions, preferences, and goals. By developing responsive and scalable solutions, we not only accomplish the shopping mission but also foster unparalleled trust among our customers. Through our advanced technology, we generate valuable insights, providing a guided navigation system into various search missions, ensuring a comprehensive and holistic shopping experience. Our dedication to continuous improvement through constant measurement and enhancement of the shopper experience is crucial, as we strategically navigate the balance between immediate results and long-term business growth. We are seeking an Applied Scientist who is not just adept in the theoretical aspects of Machine Learning (ML), Artificial Intelligence (AI), and Large Language Models (LLMs) but also possesses a pragmatic, hands-on approach to navigating the complexities of innovation. The ideal candidate will have a profound expertise in developing, deploying, and contributing to the next-generation shopping search engine, including but not limited to Retrieval-Augmented Generation (RAG) models, specifically tailored towards enhancing the Rufus application—an integral part of our mission to revolutionize shopping assistance. You will take the lead in conceptualizing, building, and launching groundbreaking models that significantly improve our understanding of and capabilities in enhancing the search experience. A successful applicant will display a comprehensive skill set across machine learning model development, implementation, and optimization. This includes a strong foundation in data management, software engineering best practices, and a keen awareness of the latest developments in distributed systems technology. We are looking for individuals who are determined, analytically rigorous, passionate about applied sciences, creative, and possess strong logical reasoning abilities. Join the Search Mission Understanding team, a group of pioneering ML scientists and engineers dedicated to building core ML models and developing the infrastructure for model innovation. As part of Amazon Search, you will experience the dynamic, innovative culture of a startup, backed by the extensive resources of Amazon.com (AMZN), a global leader in internet services. Our collaborative, customer-centric work environment spans across our offices in Palo Alto, CA, and Seattle, WA, offering a unique blend of opportunities for professional growth and innovation. Key job responsibilities Collaborate with cross-functional teams to identify requirements for ML model development, focusing on enhancing mission understanding through innovative AI techniques, including retrieval-Augmented Generation or LLM in general. Design and implement scalable ML models capable of processing and analyzing large datasets to improve search and shopping experiences. Must have a strong background in machine learning, AI, or computational sciences. Lead the management and experiments of ML models at scale, applying advanced ML techniques to optimize science solution. Serve as a technical lead and liaison for ML projects, facilitating collaboration across teams and addressing technical challenges. Requires strong leadership and communication skills, with a PhD in Computer Science, Machine Learning, or a related field. We are open to hiring candidates to work out of one of the following locations: Palo Alto, CA, USA | Seattle, WA, USA
US, WA, Seattle
The Artificial General Intelligence (AGI) team is looking for a passionate, talented, and inventive Senior Applied Scientist with a strong deep learning background, to lead the development of industry-leading technology with multimodal systems. Key job responsibilities As a Senior Applied Scientist with the AGI team, you will lead the development of novel algorithms and modeling techniques to advance the state of the art with multimodal systems. Your work will directly impact our customers in the form of products and services that make use of vision and language technology. You will leverage Amazon’s heterogeneous data sources and large-scale computing resources to accelerate development with multimodal Large Language Models (LLMs) and Generative Artificial Intelligence (GenAI) in Computer Vision. About the team The AGI team has a mission to push the envelope with multimodal LLMs and GenAI in Computer Vision, in order to provide the best-possible experience for our customers. We are open to hiring candidates to work out of one of the following locations: Cambridge, MA, USA | New York, NY, USA | Seattle, WA, USA | Sunnyvale, CA, USA
US, WA, Bellevue
The Artificial General Intelligent team (AGI) seeks an Applied Scientist with a strong background in machine learning and production level software engineering to spearhead the advancement and deployment of cutting-edge ML systems. As part of this team, you will collaborate with talented peers to create scalable solutions for an innovative conversational assistant, aiming to revolutionize user experiences for millions of Alexa customers. The ideal candidate possesses a solid understanding of machine learning fundamentals and has experience writing high quality software in production setting. The candidate is self-motivated, thrives in ambiguous and fast-paced environments, possess the drive to tackle complex challenges, and excel at swiftly delivering impactful solutions while iterating based on user feedback. Join us in our mission to redefine industry standards and provide unparalleled experiences for our customers. Key job responsibilities You will be expected to: · Analyze, understand, and model customer behavior and the customer experience based on large scale data · Build and measure novel online & offline metrics for personal digital assistants and customer scenarios, on diverse devices and endpoints · Create, innovate and deliver deep learning, policy-based learning, and/or machine learning based algorithms to deliver customer-impacting results · Build and deploy automated model training and evaluation pipelines · Perform model/data analysis and monitor metrics through online A/B testing · Research and implement novel machine learning and deep learning algorithms and models. We are open to hiring candidates to work out of one of the following locations: Bellevue, WA, USA | Boston, MA, USA
ZA, Cape Town
We are a new team in AWS' Kumo organisation - a combination of software engineers and AI/ML experts. Kumo is the software engineering organization that scales AWS’ support capabilities. Amazon’s mission is to be earth’s most customer-centric company and this also applies when it comes to helping our own Amazon employees with their everyday IT Support needs. Our team is innovating for the Amazonian, making the interaction with IT Support as smooth as possible. We achieve this through multiple mechanisms which eliminate root causes altogether, automate issue resolution or point customers towards the optimal troubleshooting steps for their situation. We deliver the support solutions plus the end-user content with instructions to help them self-serve. We employ machine learning solutions on multiple ends to understand our customer's behavior, predict customer's intent, deliver personalized content and automate issue resolution through chatbots. As an applied scientist on our team, you will help to build the next generation of case routing using artificial intelligence to optimize business metric targets addressing the business challenge of ensuring that the right case gets worked by the right agent within the right time limit whilst meeting the target business success metric. You will develop machine learning models and pipelines, harness and explain rich data at Amazon scale, and provide automated insights to improve case routing that impact millions of customers every day. You will be a pragmatic technical leader comfortable with ambiguity, capable of summarizing complex data and models through clear visual and written explanations. About AWS Diverse Experiences AWS values diverse experiences. Even if you do not meet all of the qualifications and skills listed in the job description, we encourage candidates to apply. If your career is just starting, hasn’t followed a traditional path, or includes alternative experiences, don’t let it stop you from applying. Why AWS? Amazon Web Services (AWS) is the world’s most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating — that’s why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses. Inclusive Team Culture Here at AWS, it’s in our nature to learn and be curious. Our employee-led affinity groups foster a culture of inclusion that empower us to be proud of our differences. Ongoing events and learning experiences, including our Conversations on Race and Ethnicity (CORE) and AmazeCon (gender diversity) conferences, inspire us to never stop embracing our uniqueness. Mentorship & Career Growth We’re continuously raising our performance bar as we strive to become Earth’s Best Employer. That’s why you’ll find endless knowledge-sharing, mentorship and other career-advancing resources here to help you develop into a better-rounded professional. Work/Life Balance We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home, which is why flexible work hours and arrangements are part of our culture. When we feel supported in the workplace and at home, there’s nothing we can’t achieve in the cloud. Sales, Marketing and Global Services (SMGS) AWS Sales, Marketing, and Global Services (SMGS) is responsible for driving revenue, adoption, and growth from the largest and fastest growing small- and mid-market accounts to enterprise-level customers including public sector. Amazon knows that a diverse, inclusive culture empowers us all to deliver the best results for our customers. We celebrate diversity in our workforce and in the ways we work. As part of our inclusive culture, we offer accommodations during the interview and onboarding process. If you’d like to discuss your accommodation options, please contact your recruiter, who will partner you with the Applicant-Candidate Accommodation Team (ACAT). You may also contact ACAT directly by emailing acat-africa@amazon.com. We want all Amazonians to have the best possible Day 1 experience. If you’ve already completed the interview process, you can contact ACAT for accommodation support before you start to ensure all your needs are met Day 1. Key job responsibilities Deliver real world production systems at AWS scale. Work closely with the business to understand the problem space, identify the opportunities and formulate the problems. Use machine learning, data mining, statistical techniques, Generative AI and others to create actionable, meaningful, and scalable solutions for the business problems. Establish scalable, efficient, automated processes for large scale data analyses, model development, model validation and model implementation Analyze complex support case datasets and metrics to drive insight Design, build, and deploy effective and innovative ML solutions to optimize case routing Evaluate the proposed solutions via offline benchmark tests as well as online A/B tests in production. Drive collaborative research and creative problem solving across science and software engineering team Propose and validate hypothesis to deliver and direct our product road map Work with engineers to deliver low latency model predictions to production We are open to hiring candidates to work out of one of the following locations: Cape Town, ZAF