AMD logo

Technical Marketing Engineer – AI Training Workloads & Performance

Posted 9 hours ago

Job Description

WHAT YOU DO AT AMD CHANGES EVERYTHING At AMD, our mission is to build great products that accelerate next-generation computing experiences—from AI and data centers, to PCs, gaming and embedded systems. Grounded in a culture of innovation and collaboration, we believe real progress comes from bold ideas, human ingenuity and a shared passion to create something extraordinary. When you join AMD, you’ll discover the real differentiator is our culture. We push the limits of innovation to solve the world’s most important challenges—striving for execution excellence, while being direct, humble, collaborative, and inclusive of diverse perspectives. Join us as we shape the future of AI and beyond. Together, we advance your career. THE ROLE: AMD is committed to delivering exceptional AI training performance across the full model lifecycle. Ease of use, clarity, and strong technical enablement are core to how our customers succeed with AMD technologies. As a Technical Marketing Engineer (TME) within the Software Product Management organization for AMD’s Data Center GPU Business Unit, you will play a critical role in helping customers achieve optimal performance and efficiency when training and adapting models on AMD Instinct GPUs. In this role, you will bridge deep technical expertise with clear communication, helping customers, partners, and internal teams understand how AMD GPUs unlock performance and value across AI training workloads and industry benchmarks. Your work will ensure customers can get productive quickly, supported by high-quality documentation, performance insights, and hands-on examples that remove friction and accelerate adoption. THE PERSON: Are you an engineering expert with a talent for pushing AI training workloads to the limits? Are you passionate about optimizing large-scale model training, understanding distributed systems, and analyzing performance bottlenecks across complex environments? We seek someone who thrives at the intersection of performance engineering and technical storytelling—someone who can deeply understand AI training workloads and communicate practical strategies to improve performance, scalability, and efficiency. KEY RESPONSIBLITIES: Partner with AMD’s AI software engineering team to develop performance-focused technical content for AI training workloads, including optimization guides, benchmarking results, scaling studies, and tuning methodologies. Serve as a subject matter expert on AI training performance across the full model lifecycle, including: Large-scale pre-training (foundation models) Fine-tuning and parameter-efficient methods (e.g., LoRA, PEFT) Reinforcement learning workflows (e.g., RLHF, RLAIF) Distillation and model compression techniques Quantization-aware training (QAT) Develop and publish deep technical content for training workloads, including: Performance analysis and bottleneck breakdowns Scaling studies (single-node and multi-node) Optimization guides for both pre-training and post-training workflows Distributed training best practices (data/model/pipeline parallelism) Workload-specific tuning strategies and competitive positioning insights Analyze and optimize training performance across key system dimensions, including compute utilization, memory efficiency, communication overhead, and scaling behavior in distributed environments. Engage with internal and external experts to validate performance claims against real-world scenarios and large-scale training runs. Validate and analyze training performance results from internal benchmarks and customer proof-of-concept (POC) engagements, ensuring accuracy, reproducibility, and credibility. Partner with engineering to translate low-level optimizations (kernels, communication patterns, memory usage) into actionable guidance and influence product improvements based on real-world workload feedback. Stay current on industry training frameworks, distributed training strategies, and emerging techniques, identifying where AMD platforms deliver differentiated performance. Act as a bridge between engineering, field teams (FAE/FDE), and business stakeholders, ensuring alignment on training performance capabilities, best practices, and messaging. Design and deliver technical enablement (e.g., workshops, demos, technical sessions) to help internal teams and customers optimize AI training workloads. Build deep empathy for the challenges faced by solutions architects and engineers deploying AMD GPUs, and create documentation that proactively removes barriers to adoption. Actively gather and incorporate feedback from customers, field teams, and internal technical experts to continuously improve content quality and relevance. Lead by example in documentation excellence, mentoring and influencing peers to raise the overall quality, consistency, and effectiveness of technical materials. Communicate with clarity, integrity, and transparency across written, verbal, and visual formats PREFERRED EXPERIENCE: 5+ years optimizing AI training workloads on GPUs or accelerators at scale. Hands-on experience with distributed training frameworks (e.g., PyTorch, DeepSpeed, Megatron-LM). Strong programming skills in Python and/or C/C++. Experience with ROCm, CUDA, or similar GPU compute stacks. Experience working with ISVs, hyperscalers, or large-scale AI deployments. Background in solution validation, benchmarking, or certification frameworks is highly desirable. Proven ability to create high-quality technical documentation and performance guides. Proven experience delivering technical presentations, workshops, or participating in industry conferences, meetups, or hackathons. Expertise in modern documentation and publishing workflows, including Markdown, Read the Docs, and Jupyter Notebooks. Proficiency with standard productivity and publishing tools (e.g., Microsoft Office, Adobe Acrobat) used for technical documentation and training delivery. Confidence presenting complex technical material in customer‑facing settings within a fast‑paced, large‑scale organization. ACADEMIC CREDENTIALS BS in Computer Science, Computer Engineering, Electrical Engineering, or a related field required, MS or PhD is a plus. LOCATION Santa Clara, CA This role is not eligible for visa sponsorship. Benefits offered are described: AMD benefits at a glance. AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process. AMD may use Artificial Intelligence to help screen, assess or select applicants for this position. AMD’s “Responsible AI Policy” is available here. This posting is for an existing vacancy.

Benefits offered are described: AMD benefits at a glance. AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process. AMD may use Artificial Intelligence to help screen, assess or select applicants for this position. AMD’s “Responsible AI Policy” is available here. This posting is for an existing vacancy.

THE ROLE: AMD is committed to delivering exceptional AI training performance across the full model lifecycle. Ease of use, clarity, and strong technical enablement are core to how our customers succeed with AMD technologies. As a Technical Marketing Engineer (TME) within the Software Product Management organization for AMD’s Data Center GPU Business Unit, you will play a critical role in helping customers achieve optimal performance and efficiency when training and adapting models on AMD Instinct GPUs. In this role, you will bridge deep technical expertise with clear communication, helping customers, partners, and internal teams understand how AMD GPUs unlock performance and value across AI training workloads and industry benchmarks. Your work will ensure customers can get productive quickly, supported by high-quality documentation, performance insights, and hands-on examples that remove friction and accelerate adoption. THE PERSON: Are you an engineering expert with a talent for pushing AI training workloads to the limits? Are you passionate about optimizing large-scale model training, understanding distributed systems, and analyzing performance bottlenecks across complex environments? We seek someone who thrives at the intersection of performance engineering and technical storytelling—someone who can deeply understand AI training workloads and communicate practical strategies to improve performance, scalability, and efficiency. KEY RESPONSIBLITIES: Partner with AMD’s AI software engineering team to develop performance-focused technical content for AI training workloads, including optimization guides, benchmarking results, scaling studies, and tuning methodologies. Serve as a subject matter expert on AI training performance across the full model lifecycle, including: Large-scale pre-training (foundation models) Fine-tuning and parameter-efficient methods (e.g., LoRA, PEFT) Reinforcement learning workflows (e.g., RLHF, RLAIF) Distillation and model compression techniques Quantization-aware training (QAT) Develop and publish deep technical content for training workloads, including: Performance analysis and bottleneck breakdowns Scaling studies (single-node and multi-node) Optimization guides for both pre-training and post-training workflows Distributed training best practices (data/model/pipeline parallelism) Workload-specific tuning strategies and competitive positioning insights Analyze and optimize training performance across key system dimensions, including compute utilization, memory efficiency, communication overhead, and scaling behavior in distributed environments. Engage with internal and external experts to validate performance claims against real-world scenarios and large-scale training runs. Validate and analyze training performance results from internal benchmarks and customer proof-of-concept (POC) engagements, ensuring accuracy, reproducibility, and credibility. Partner with engineering to translate low-level optimizations (kernels, communication patterns, memory usage) into actionable guidance and influence product improvements based on real-world workload feedback. Stay current on industry training frameworks, distributed training strategies, and emerging techniques, identifying where AMD platforms deliver differentiated performance. Act as a bridge between engineering, field teams (FAE/FDE), and business stakeholders, ensuring alignment on training performance capabilities, best practices, and messaging. Design and deliver technical enablement (e.g., workshops, demos, technical sessions) to help internal teams and customers optimize AI training workloads. Build deep empathy for the challenges faced by solutions architects and engineers deploying AMD GPUs, and create documentation that proactively removes barriers to adoption. Actively gather and incorporate feedback from customers, field teams, and internal technical experts to continuously improve content quality and relevance. Lead by example in documentation excellence, mentoring and influencing peers to raise the overall quality, consistency, and effectiveness of technical materials. Communicate with clarity, integrity, and transparency across written, verbal, and visual formats PREFERRED EXPERIENCE: 5+ years optimizing AI training workloads on GPUs or accelerators at scale. Hands-on experience with distributed training frameworks (e.g., PyTorch, DeepSpeed, Megatron-LM). Strong programming skills in Python and/or C/C++. Experience with ROCm, CUDA, or similar GPU compute stacks. Experience working with ISVs, hyperscalers, or large-scale AI deployments. Background in solution validation, benchmarking, or certification frameworks is highly desirable. Proven ability to create high-quality technical documentation and performance guides. Proven experience delivering technical presentations, workshops, or participating in industry conferences, meetups, or hackathons. Expertise in modern documentation and publishing workflows, including Markdown, Read the Docs, and Jupyter Notebooks. Proficiency with standard productivity and publishing tools (e.g., Microsoft Office, Adobe Acrobat) used for technical documentation and training delivery. Confidence presenting complex technical material in customer‑facing settings within a fast‑paced, large‑scale organization. ACADEMIC CREDENTIALS BS in Computer Science, Computer Engineering, Electrical Engineering, or a related field required, MS or PhD is a plus. LOCATION Santa Clara, CA This role is not eligible for visa sponsorship.

Please mention that you found this job on MoAIJobs, this helps us grow. Thank you!

Share this job

AMD logo

AMD

78 jobs posted

View all AMD jobs

About the job

Posted on

May 12, 2026

Apply before

Jun 11, 2026

Job typeFull-time

Similar Jobs

AI Training jobs

AI jobs in your inbox

Get the latest AI jobs delivered to your inbox every week. Free, no spam.