Job Description
The Product: AWS Machine Learning accelerators are at the forefront of AWS innovation. The Inferentia chip delivers best-in-class ML inference performance at the lowest cost in cloud. Trainium will deliver the best-in-class ML training performance with the most teraflops (TFLOPS) of compute power for ML in the cloud. This is all enabled by cutting edge software stack, the AWS Neuron Software Development Kit (SDK), which includes an ML compiler, runtime and natively integrates into popular ML frameworks, such as PyTorch, TensorFlow and JAX. AWS Neuron and Inferentia are used at scale with customers like Snap, Autodesk, Amazon Alexa, Amazon Rekognition and more customers in various other segments.
The Team: The Amazon Annapurna Labs team is a responsible for building innovation in silicon and software for AWS customers. We are at the forefront of innovation by combining cloud scale with the world’s most talented engineers. Our team covers multiple disciplines including silicon engineering, hardware design and verification, software and operations. Because of our teams breadth of talent, we have been able to improve AWS cloud infrastructure in networking and security with products such as AWS Nitro, Enhanced Network Adapter (ENA), and Elastic Fabric Adapter (EFA), in compute with AWS Graviton and the EC2 F1 FPGA instances, in storage with scalable NVMe, and now in AI and Machine Learning with AWS Neuron SDK, Inferentia and Trainium ML accelerators.
You: In this customer-facing role, you will work closely with our Neuron software development team and strategic customers on cutting edge accelerated Machine Learning solutions. You will bring your hands-on experience developing and deploying Deep Learning models and integrate it with our ML accelerator products, into large-scalable production applications.
You will need to be technically capable and credible in your own right, to become a trusted advisor for customers developing, deploying and scaling Deep Learning applications on AWS ML accelerators. You’ll succeed in this position if you enjoy capturing and sharing best practices and insights, and help shape how AWS ML accelerator technology gets used. You will be a hands-on partner to AWS services teams, technical field communities, sales, marketing, business development, and professional services, to drive adoption. You’ll leverage your communications skills, and be very technical when doing so, to help amplify the thought-leadership around AWS Neuron technology stack to the broader AWS field community, as well as our customers.
Key job responsibilities
- Design architectures and own Proof of Concept (PoC) solutions for strategic customers, leveraging AWS ML accelerators technologies and the broader set of AWS features and services.
- Drive adoption by taking ownership of technical engagements with eco-system partners and strategic customers, assisting with the definition and implementation of technical roadmaps and enabling them to successfully deploy on AWS ML Accelerator.
- Develop strong partnership with engineering organizations, serving as the customer advocate, to help drive product roadmap working backwards from customers feedback.
- Drive thought leadership by crafting and delivering compelling audience-specific messaging artifacts (product videos, demos, workshops, how to guides etc.) presenting AWS ML accelerator technology through AWS Blogs, reference architectures and solutions, and public-speaking events.
- Capture, implement and share best-practices knowledge among the AWS technical community regarding AWS ML Accelerators.
The Team: The Amazon Annapurna Labs team is a responsible for building innovation in silicon and software for AWS customers. We are at the forefront of innovation by combining cloud scale with the world’s most talented engineers. Our team covers multiple disciplines including silicon engineering, hardware design and verification, software and operations. Because of our teams breadth of talent, we have been able to improve AWS cloud infrastructure in networking and security with products such as AWS Nitro, Enhanced Network Adapter (ENA), and Elastic Fabric Adapter (EFA), in compute with AWS Graviton and the EC2 F1 FPGA instances, in storage with scalable NVMe, and now in AI and Machine Learning with AWS Neuron SDK, Inferentia and Trainium ML accelerators.
You: In this customer-facing role, you will work closely with our Neuron software development team and strategic customers on cutting edge accelerated Machine Learning solutions. You will bring your hands-on experience developing and deploying Deep Learning models and integrate it with our ML accelerator products, into large-scalable production applications.
You will need to be technically capable and credible in your own right, to become a trusted advisor for customers developing, deploying and scaling Deep Learning applications on AWS ML accelerators. You’ll succeed in this position if you enjoy capturing and sharing best practices and insights, and help shape how AWS ML accelerator technology gets used. You will be a hands-on partner to AWS services teams, technical field communities, sales, marketing, business development, and professional services, to drive adoption. You’ll leverage your communications skills, and be very technical when doing so, to help amplify the thought-leadership around AWS Neuron technology stack to the broader AWS field community, as well as our customers.
Key job responsibilities
- Design architectures and own Proof of Concept (PoC) solutions for strategic customers, leveraging AWS ML accelerators technologies and the broader set of AWS features and services.
- Drive adoption by taking ownership of technical engagements with eco-system partners and strategic customers, assisting with the definition and implementation of technical roadmaps and enabling them to successfully deploy on AWS ML Accelerator.
- Develop strong partnership with engineering organizations, serving as the customer advocate, to help drive product roadmap working backwards from customers feedback.
- Drive thought leadership by crafting and delivering compelling audience-specific messaging artifacts (product videos, demos, workshops, how to guides etc.) presenting AWS ML accelerator technology through AWS Blogs, reference architectures and solutions, and public-speaking events.
- Capture, implement and share best-practices knowledge among the AWS technical community regarding AWS ML Accelerators.
Job Alerts
Get the latest AI jobs delivered to your inbox every Wednesday.
Free, weekly digest. No spam.
Amazon
127 jobs posted
About the job
Posted on
Feb 16, 2026
Apply before
Mar 18, 2026
Job typeFull-time
CategoryOther AI jobs
Location
US, NC
Skills
TensorflowPytorchDeep Learning
Job Alerts
Get the latest AI jobs delivered to your inbox every Wednesday.
Free, weekly digest. No spam.
Similar Jobs

Mistral
20 days agoAI Solution Architect - Montreal
View detailsAmazon
10 days agoGenAI Solution Architect
ThailandView detailsAmazon
10 days agoGenAI Solution Architect
PhilippinesView details
Mistral
20 days agoAI Solution Architect - Palo Alto
View detailsDell Technologies
13 days agoIntel AI Solution Architect-CSC Consultant
Round Rock, Texas$171K - $221K/yrView detailsDigitalOcean
9 hours agoSenior Solutions Architect (AI/ML)
Seattle, WA$117K - $147K/yrView detailsDigitalOcean
9 hours agoSenior Solutions Architect (AI/ML)
Denver$117K - $147K/yrView detailsDigitalOcean
9 hours agoSenior Solutions Architect (AI/ML)
Boston$117K - $147K/yrView detailsDigitalOcean
9 hours agoSenior Solutions Architect (AI/ML)
Austin$117K - $147K/yrView detailsDigitalOcean
9 hours agoSenior Solutions Architect (AI/ML)
San Francisco, CA$117K - $147K/yrView details
Looking for something different?
Browse all AI jobs