Advanced Micro Devices, Inc.
AI/ML Compiler Development Engineer
Advanced Micro Devices, Inc., San Jose, California, United States, 95199
WHAT YOU DO AT AMD CHANGES EVERYTHING
At AMD, our mission is to build great products that accelerate next‑generation computing experiences—from AI and data centers, to PCs, gaming and embedded systems. Grounded in a culture of innovation and collaboration, we believe real progress comes from bold ideas, human ingenuity and a shared passion to create something extraordinary. When you join AMD, you’ll discover the real differentiator is our culture. We push the limits of innovation to solve the world’s most important challenges—striving for execution excellence, while being direct, humble, collaborative, and inclusive of diverse perspectives. Join us as we shape the future of AI and beyond.
Together, we advance your career.
THE ROLE We are looking for a dynamic, energetic candidate to join our growing team in AI Group. In this role, you will architect and define AI workload models, dataflow, block‑level and system‑level performance of Neural Processing Unit (NPU), NPU network performance modeling, and performance bottleneck analysis on pre/post silicon platforms. As a member of our dynamic team, you will shape the future of AI model development.
THE PERSON
You possess strong engineering skills to tackle complex challenges on AI model development work.
You have experience in optimizing and accelerating CNN/Generative AI models and excellent cross‑team collaboration skills.
You have strong experience in developing an ML compiler for efficient network mapping on NPU.
You work with cross‑functional teams to optimize various parts of the SW stack – AI compiler, AI frameworks, device drivers and firmware.
You bring up emerging ML models based on CNN, transformers and characterize performance.
Work on quantization, sparsity, and architecture search methods to optimize and enhance the performance, efficiency, and accuracy of Generative AI models.
You collaborate closely with software engineers, data scientists and researchers to integrate AI models into software applications and platforms.
KEY RESPONSIBILITIES
Research, design and implement novel methods for efficient CNN, Gen AI models.
Model optimization methods including quantization, sparsity, NAS, etc.
Collaborate with team members and teams.
Collaborate with compiler team to develop optimization strategies for the compiler.
PREFERRED EXPERIENCE
Experience with deep learning framework, e.g., PyTorch/ONNX/TensorFlow.
Experience on model compression, quantization and end‑to‑end inference optimization.
Strong coding skills in C/C++ and Python.
Experience with LLMs, stable diffusion, NeRF or text‑to‑video generation is a plus.
Solid knowledge of AI and ML concepts and techniques; practical experience applying these concepts to solve real‑world problems.
Understanding performance implications on AI acceleration of different compute, memory and communication configurations and hardware and software implementation choices.
Developing and optimizing code for VLIW processors. Analyzing code for high-performance CONV, GEMM and non‑linear operators. Deep understanding of AI frameworks, preferably ONNX.
ACADEMIC CREDENTIALS
Minimum of a BS degree, MS or above preferred.
LOCATION San Jose, CA (Hybrid)
REWARDS AND BENEFITS At AMD, your base pay is one part of your total rewards package… you may be eligible for incentives… You’ll also be eligible for competitive benefits described in more detail here.
EEO STATEMENT AMD does not accept unsolicited resumes from headhunters, recruitment agencies or fee‑based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third‑party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants' needs under the respective laws throughout all stages of the recruitment and selection process.
#J-18808-Ljbffr
Together, we advance your career.
THE ROLE We are looking for a dynamic, energetic candidate to join our growing team in AI Group. In this role, you will architect and define AI workload models, dataflow, block‑level and system‑level performance of Neural Processing Unit (NPU), NPU network performance modeling, and performance bottleneck analysis on pre/post silicon platforms. As a member of our dynamic team, you will shape the future of AI model development.
THE PERSON
You possess strong engineering skills to tackle complex challenges on AI model development work.
You have experience in optimizing and accelerating CNN/Generative AI models and excellent cross‑team collaboration skills.
You have strong experience in developing an ML compiler for efficient network mapping on NPU.
You work with cross‑functional teams to optimize various parts of the SW stack – AI compiler, AI frameworks, device drivers and firmware.
You bring up emerging ML models based on CNN, transformers and characterize performance.
Work on quantization, sparsity, and architecture search methods to optimize and enhance the performance, efficiency, and accuracy of Generative AI models.
You collaborate closely with software engineers, data scientists and researchers to integrate AI models into software applications and platforms.
KEY RESPONSIBILITIES
Research, design and implement novel methods for efficient CNN, Gen AI models.
Model optimization methods including quantization, sparsity, NAS, etc.
Collaborate with team members and teams.
Collaborate with compiler team to develop optimization strategies for the compiler.
PREFERRED EXPERIENCE
Experience with deep learning framework, e.g., PyTorch/ONNX/TensorFlow.
Experience on model compression, quantization and end‑to‑end inference optimization.
Strong coding skills in C/C++ and Python.
Experience with LLMs, stable diffusion, NeRF or text‑to‑video generation is a plus.
Solid knowledge of AI and ML concepts and techniques; practical experience applying these concepts to solve real‑world problems.
Understanding performance implications on AI acceleration of different compute, memory and communication configurations and hardware and software implementation choices.
Developing and optimizing code for VLIW processors. Analyzing code for high-performance CONV, GEMM and non‑linear operators. Deep understanding of AI frameworks, preferably ONNX.
ACADEMIC CREDENTIALS
Minimum of a BS degree, MS or above preferred.
LOCATION San Jose, CA (Hybrid)
REWARDS AND BENEFITS At AMD, your base pay is one part of your total rewards package… you may be eligible for incentives… You’ll also be eligible for competitive benefits described in more detail here.
EEO STATEMENT AMD does not accept unsolicited resumes from headhunters, recruitment agencies or fee‑based recruitment services. AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and/or third‑party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law. We encourage applications from all qualified candidates and will accommodate applicants' needs under the respective laws throughout all stages of the recruitment and selection process.
#J-18808-Ljbffr