AMD
Company
Senior AI/ML Infrastructure Engineer
Austin, Texas
Job Description
WHAT YOU DO AT AMD CHANGES EVERYTHING At AMD, our mission is to build great products that accelerate next-generation computing experiences—from AI and data centers, to PCs, gaming and embedded systems. Grounded in a culture of innovation and collaboration, we believe real progress comes from bold ideas, human ingenuity and a shared passion to create something extraordinary. When you join AMD, you’ll discover the real differentiator is our culture. We push the limits of innovation to solve the world’s most important challenges—striving for execution excellence, while being direct, humble, collaborative, and inclusive of diverse perspectives. Join us as we shape the future of AI and beyond. Together, we advance your career. THE ROLE: AMD is looking for a specialized software engineer who is passionate about improving the performance of key applications and benchmarks. You will be a member of a core team of incredibly talented industry specialists and will work with the very latest hardware and software technology. THE PERSON: The ideal candidate should be passionate about software engineering and possess leadership skills to drive sophisticated issues to resolution. Able to communicate effectively and work optimally with different teams across AMD. KEY RESPONSIBILITIES: Architect and scale end-to-end infrastructure for training, fine-tuning, and deploying machine learning and large language models (LLMs) across distributed, hybrid cloud, and on-prem environments. Design and maintain MLOps/LLMOps pipelines, ensuring seamless integration between data ingestion, model training, evaluation, and production deployment. Collaborate closely with AI researchers, data scientists, and platform engineers to enable efficient experimentation, versioning, and continuous improvement of models. Implement and manage robust CI/CD and CI/CT (Continuous Training) pipelines, automating testing, validation, and deployment workflows for ML/LLM systems. Develop tools and frameworks for model serving, latency optimization, and inference performance tuning, leveraging GPU acceleration and distributed systems best practices. Monitor and improve production ML systems - track drift, performance degradation, and operational efficiency; implement alerting and rollback strategies. Drive software engineering excellence, emphasizing code quality, testability, and scalability for all ML infrastructure components. Optimize compute and storage resources across cloud and on-prem clusters, ensuring cost efficiency and high utilization. Stay current with the latest advancements in ML orchestration, LLM optimization (quantization, distillation, caching), and deployment frameworks. Document, communicate, and mentor - promote best practices in reproducibility, version control, and model governance. PREFERRED EXPERIENCE: MLOps / LLMOps Expertise: Proven experience building and maintaining production ML pipelines, model registries, and deployment systems for large-scale AI workloads. Software Development: Strong background in systems engineering and backend development (Python preferred), with deep understanding of distributed systems, APIs, and infrastructure reliability. Deep Learning Frameworks: Proficiency with PyTorch, TensorFlow, and LLM toolchains (Hugging Face, DeepSpeed, vLLM, etc.). Infrastructure & Cloud: Familiarity with Kubernetes, Docker and experience with on-prem GPU clusters is a plus. CI/CD & Automation: Strong understanding of GitOps, continuous integration/testing/deployment systems (e.g., Jenkins, GitHub Actions, Dagster, Airflow). Optimization & Performance: Hands-on experience with model and system-level optimizations, including quantization, batching, caching, and hardware utilization tuning. Problem-Solving: Demonstrated ability to debug complex distributed systems, resolve performance bottlenecks, and drive root-cause analysis in production ML environments. Collaboration: Excellent communication skills and the ability to work cross-functionally with research, data, and product teams. Continuous Learning: Track record of quickly adapting to new ML tools, frameworks, and technologies in a fast-evolving AI ecosystem. ACADEMIC CREDENTIALS: Bachelor’s or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, or equivalent #LI-JG1 Benefits offered are described: AMD benefits at a glance. AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process.
Benefits offered are described: AMD benefits at a glance. AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process.
THE ROLE: AMD is looking for a specialized software engineer who is passionate about improving the performance of key applications and benchmarks. You will be a member of a core team of incredibly talented industry specialists and will work with the very latest hardware and software technology. THE PERSON: The ideal candidate should be passionate about software engineering and possess leadership skills to drive sophisticated issues to resolution. Able to communicate effectively and work optimally with different teams across AMD. KEY RESPONSIBILITIES: Architect and scale end-to-end infrastructure for training, fine-tuning, and deploying machine learning and large language models (LLMs) across distributed, hybrid cloud, and on-prem environments. Design and maintain MLOps/LLMOps pipelines, ensuring seamless integration between data ingestion, model training, evaluation, and production deployment. Collaborate closely with AI researchers, data scientists, and platform engineers to enable efficient experimentation, versioning, and continuous improvement of models. Implement and manage robust CI/CD and CI/CT (Continuous Training) pipelines, automating testing, validation, and deployment workflows for ML/LLM systems. Develop tools and frameworks for model serving, latency optimization, and inference performance tuning, leveraging GPU acceleration and distributed systems best practices. Monitor and improve production ML systems - track drift, performance degradation, and operational efficiency; implement alerting and rollback strategies. Drive software engineering excellence, emphasizing code quality, testability, and scalability for all ML infrastructure components. Optimize compute and storage resources across cloud and on-prem clusters, ensuring cost efficiency and high utilization. Stay current with the latest advancements in ML orchestration, LLM optimization (quantization, distillation, caching), and deployment frameworks. Document, communicate, and mentor - promote best practices in reproducibility, version control, and model governance. PREFERRED EXPERIENCE: MLOps / LLMOps Expertise: Proven experience building and maintaining production ML pipelines, model registries, and deployment systems for large-scale AI workloads. Software Development: Strong background in systems engineering and backend development (Python preferred), with deep understanding of distributed systems, APIs, and infrastructure reliability. Deep Learning Frameworks: Proficiency with PyTorch, TensorFlow, and LLM toolchains (Hugging Face, DeepSpeed, vLLM, etc.). Infrastructure & Cloud: Familiarity with Kubernetes, Docker and experience with on-prem GPU clusters is a plus. CI/CD & Automation: Strong understanding of GitOps, continuous integration/testing/deployment systems (e.g., Jenkins, GitHub Actions, Dagster, Airflow). Optimization & Performance: Hands-on experience with model and system-level optimizations, including quantization, batching, caching, and hardware utilization tuning. Problem-Solving: Demonstrated ability to debug complex distributed systems, resolve performance bottlenecks, and drive root-cause analysis in production ML environments. Collaboration: Excellent communication skills and the ability to work cross-functionally with research, data, and product teams. Continuous Learning: Track record of quickly adapting to new ML tools, frameworks, and technologies in a fast-evolving AI ecosystem. ACADEMIC CREDENTIALS: Bachelor’s or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, or equivalent #LI-JG1
Benefits offered are described: AMD benefits at a glance. AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process.
THE ROLE: AMD is looking for a specialized software engineer who is passionate about improving the performance of key applications and benchmarks. You will be a member of a core team of incredibly talented industry specialists and will work with the very latest hardware and software technology. THE PERSON: The ideal candidate should be passionate about software engineering and possess leadership skills to drive sophisticated issues to resolution. Able to communicate effectively and work optimally with different teams across AMD. KEY RESPONSIBILITIES: Architect and scale end-to-end infrastructure for training, fine-tuning, and deploying machine learning and large language models (LLMs) across distributed, hybrid cloud, and on-prem environments. Design and maintain MLOps/LLMOps pipelines, ensuring seamless integration between data ingestion, model training, evaluation, and production deployment. Collaborate closely with AI researchers, data scientists, and platform engineers to enable efficient experimentation, versioning, and continuous improvement of models. Implement and manage robust CI/CD and CI/CT (Continuous Training) pipelines, automating testing, validation, and deployment workflows for ML/LLM systems. Develop tools and frameworks for model serving, latency optimization, and inference performance tuning, leveraging GPU acceleration and distributed systems best practices. Monitor and improve production ML systems - track drift, performance degradation, and operational efficiency; implement alerting and rollback strategies. Drive software engineering excellence, emphasizing code quality, testability, and scalability for all ML infrastructure components. Optimize compute and storage resources across cloud and on-prem clusters, ensuring cost efficiency and high utilization. Stay current with the latest advancements in ML orchestration, LLM optimization (quantization, distillation, caching), and deployment frameworks. Document, communicate, and mentor - promote best practices in reproducibility, version control, and model governance. PREFERRED EXPERIENCE: MLOps / LLMOps Expertise: Proven experience building and maintaining production ML pipelines, model registries, and deployment systems for large-scale AI workloads. Software Development: Strong background in systems engineering and backend development (Python preferred), with deep understanding of distributed systems, APIs, and infrastructure reliability. Deep Learning Frameworks: Proficiency with PyTorch, TensorFlow, and LLM toolchains (Hugging Face, DeepSpeed, vLLM, etc.). Infrastructure & Cloud: Familiarity with Kubernetes, Docker and experience with on-prem GPU clusters is a plus. CI/CD & Automation: Strong understanding of GitOps, continuous integration/testing/deployment systems (e.g., Jenkins, GitHub Actions, Dagster, Airflow). Optimization & Performance: Hands-on experience with model and system-level optimizations, including quantization, batching, caching, and hardware utilization tuning. Problem-Solving: Demonstrated ability to debug complex distributed systems, resolve performance bottlenecks, and drive root-cause analysis in production ML environments. Collaboration: Excellent communication skills and the ability to work cross-functionally with research, data, and product teams. Continuous Learning: Track record of quickly adapting to new ML tools, frameworks, and technologies in a fast-evolving AI ecosystem. ACADEMIC CREDENTIALS: Bachelor’s or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, or equivalent #LI-JG1
AMD
110 jobs posted
About the job
Similar Jobs
Discover more opportunities that match your interests
- 30 days ago
Senior Staff AI Infrastructure Engineer
AMD
Santa Clara, CaliforniaView details - 23 days ago
Senior Specialist – AI-ML Engineer
Nasdaq
India - Bangalore - KarnatakaView details - 22 days ago
Senior Staff ML Infrastructure Engineer
Coupang
Mountain View, USAView details
16 days agoNetwork Engineer, AI/ML Infrastructure
Boson AI
TorontoView details
16 days agoHPC Engineer, AI/ML Infrastructure
Boson AI
TorontoView details
17 days agoNetwork Engineer, AI/ML Infrastructure
Boson AI
Santa Clara HQView details
17 days agoHPC Engineer, AI/ML Infrastructure
Boson AI
Santa Clara HQView details
1 day agoNetwork Engineer, AI/ML Infrastructure
Boson AI
TorontoView details
1 day agoNetwork Engineer, AI/ML Infrastructure
Boson AI
Santa Clara HQView details- 28 days ago
AI Infrastructure Engineer
AMD
San Jose, CaliforniaView details
View all ML Engineer jobs
Looking for something different?
Browse all AI jobs