Amazon Web Services (AWS)
Machine Learning - Compiler Engineer , AWS Neuron, Annapurna Labs
Amazon Web Services (AWS), Cupertino, California, United States, 95014
5 days ago Be among the first 25 applicants
Description Do you want to be part of the AI revolution? At AWS our vision is to make deep learning pervasive for everyday developers and to democratize access to AI hardware and software infrastructure. In order to deliver on that vision, we’ve created innovative software and hardware solutions that make it possible. AWS Neuron is the SDK that optimizes the performance of complex ML models executed on AWS Inferentia and Trainium, our custom chips designed to accelerate deep‑learning workloads.
This role is for a software engineer in the Compiler team for AWS Neuron. As part of this role, you will be responsible for building the next‑generation Neuron compiler which transforms ML models written in popular frameworks (e.g., PyTorch, TensorFlow, and JAX) to be deployed on AWS Inferentia and Trainium based servers in the Amazon cloud. You will solve hard compiler optimization problems to achieve optimum performance for a wide variety of ML model families, including massive‑scale large language models such as Llama and Deepseek, as well as stable diffusion, vision transformers, and multi‑model solutions. You will need an in‑depth understanding of how these models work inside‑out to make informed decisions that coax the compiler into generating optimal instruction implementations. You will leverage technical communications skills to partner with internal and external customers/stakeholders and will be involved in pre‑silicon design, bringing new products/features to market, ultimately making the Neuron compiler highly performant and easy‑to‑use.
Key job responsibilities
Design, implement, test, deploy, and maintain innovative software solutions to improve Neuron compiler performance, stability, and user interface.
Collaborate with chip architects, runtime/OS engineers, scientists, and ML Apps teams to seamlessly deploy state‑of‑the‑art ML models on AWS accelerators with optimal cost/performance benefits.
Work with open‑source software (StableHLO, OpenXLA, MLIR) to pioneer optimizations for advanced ML workloads on AWS hardware.
Create compiler optimization and verification passes, build feature surface areas for AWS accelerators, implement tools to analyze numerical errors, and resolve the root cause of compiler defects.
Participate in design discussions, code reviews, and communicate with internal and external stakeholders, including open‑source communities.
Day in the life As you design and code solutions to help our team drive efficiencies in compiler architecture, you’ll create optimization passes, build feature surface areas for AWS accelerators, implement analysis tools, and resolve compiler defects. You’ll also engage in design discussions, code reviews, and maintain effective communication with internal teams and external communities in a startup‑like environment.
About the team Our team is dedicated to supporting new members. We have a broad mix of experience levels and celebrate knowledge‑sharing and mentorship. Senior members enjoy one‑on‑one mentoring and thorough, but kind, code reviews. We care about career growth and strive to assign projects that help team members develop engineering expertise.
Basic Qualifications
3+ years of non‑internship professional software development experience.
2+ years of non‑internship design or architecture (design patterns, reliability, and scaling) of new and existing systems.
Experience programming with at least one software programming language.
Preferred Qualifications
Master’s degree or PhD in Computer Science, or a related technical field.
3+ years of experience writing production‑grade code in object‑oriented languages such as C++ or Java.
Experience in compiler design for CPU/GPU/Vector engines/ML‑accelerators.
Experience with OpenSource compiler toolsets like LLVM/MLIR.
Experience with PyTorch, OpenXLA, StableHLO, JAX, TVM, deep‑learning models, and algorithms.
Experience with modern build systems such as Bazel or CMake.
Amazon is an equal‑opportunity employer and does not discriminate on the basis of protected veteran status, disability, or other legally protected status.
Los Angeles County applicants: Job duties for this position include: working safely and cooperatively with other employees, supervisors, and staff; adhering to standards of excellence despite stressful conditions; communicating effectively and respectfully with employees, supervisors, and staff to ensure exceptional customer service; and following all federal, state, and local laws and Company policies. Criminal history may have a direct, adverse, and negative relationship with some of the material job duties of this position. Pursuant to the Los Angeles County Fair Chance Ordinance, we will consider qualified applicants with arrest and conviction records.
Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information.
Compensation reflects the cost of labor across several US geographic markets. Base pay ranges from $129,300/year in our lowest geographic market up to $223,600/year in our highest geographic market. Pay is based on multiple factors including location, experience, and skills. For more information, please visit https://www.aboutamazon.com/workplace/employee-benefits. This position will remain posted until filled. Applicants should apply via our internal or external career site.
#J-18808-Ljbffr
Description Do you want to be part of the AI revolution? At AWS our vision is to make deep learning pervasive for everyday developers and to democratize access to AI hardware and software infrastructure. In order to deliver on that vision, we’ve created innovative software and hardware solutions that make it possible. AWS Neuron is the SDK that optimizes the performance of complex ML models executed on AWS Inferentia and Trainium, our custom chips designed to accelerate deep‑learning workloads.
This role is for a software engineer in the Compiler team for AWS Neuron. As part of this role, you will be responsible for building the next‑generation Neuron compiler which transforms ML models written in popular frameworks (e.g., PyTorch, TensorFlow, and JAX) to be deployed on AWS Inferentia and Trainium based servers in the Amazon cloud. You will solve hard compiler optimization problems to achieve optimum performance for a wide variety of ML model families, including massive‑scale large language models such as Llama and Deepseek, as well as stable diffusion, vision transformers, and multi‑model solutions. You will need an in‑depth understanding of how these models work inside‑out to make informed decisions that coax the compiler into generating optimal instruction implementations. You will leverage technical communications skills to partner with internal and external customers/stakeholders and will be involved in pre‑silicon design, bringing new products/features to market, ultimately making the Neuron compiler highly performant and easy‑to‑use.
Key job responsibilities
Design, implement, test, deploy, and maintain innovative software solutions to improve Neuron compiler performance, stability, and user interface.
Collaborate with chip architects, runtime/OS engineers, scientists, and ML Apps teams to seamlessly deploy state‑of‑the‑art ML models on AWS accelerators with optimal cost/performance benefits.
Work with open‑source software (StableHLO, OpenXLA, MLIR) to pioneer optimizations for advanced ML workloads on AWS hardware.
Create compiler optimization and verification passes, build feature surface areas for AWS accelerators, implement tools to analyze numerical errors, and resolve the root cause of compiler defects.
Participate in design discussions, code reviews, and communicate with internal and external stakeholders, including open‑source communities.
Day in the life As you design and code solutions to help our team drive efficiencies in compiler architecture, you’ll create optimization passes, build feature surface areas for AWS accelerators, implement analysis tools, and resolve compiler defects. You’ll also engage in design discussions, code reviews, and maintain effective communication with internal teams and external communities in a startup‑like environment.
About the team Our team is dedicated to supporting new members. We have a broad mix of experience levels and celebrate knowledge‑sharing and mentorship. Senior members enjoy one‑on‑one mentoring and thorough, but kind, code reviews. We care about career growth and strive to assign projects that help team members develop engineering expertise.
Basic Qualifications
3+ years of non‑internship professional software development experience.
2+ years of non‑internship design or architecture (design patterns, reliability, and scaling) of new and existing systems.
Experience programming with at least one software programming language.
Preferred Qualifications
Master’s degree or PhD in Computer Science, or a related technical field.
3+ years of experience writing production‑grade code in object‑oriented languages such as C++ or Java.
Experience in compiler design for CPU/GPU/Vector engines/ML‑accelerators.
Experience with OpenSource compiler toolsets like LLVM/MLIR.
Experience with PyTorch, OpenXLA, StableHLO, JAX, TVM, deep‑learning models, and algorithms.
Experience with modern build systems such as Bazel or CMake.
Amazon is an equal‑opportunity employer and does not discriminate on the basis of protected veteran status, disability, or other legally protected status.
Los Angeles County applicants: Job duties for this position include: working safely and cooperatively with other employees, supervisors, and staff; adhering to standards of excellence despite stressful conditions; communicating effectively and respectfully with employees, supervisors, and staff to ensure exceptional customer service; and following all federal, state, and local laws and Company policies. Criminal history may have a direct, adverse, and negative relationship with some of the material job duties of this position. Pursuant to the Los Angeles County Fair Chance Ordinance, we will consider qualified applicants with arrest and conviction records.
Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information.
Compensation reflects the cost of labor across several US geographic markets. Base pay ranges from $129,300/year in our lowest geographic market up to $223,600/year in our highest geographic market. Pay is based on multiple factors including location, experience, and skills. For more information, please visit https://www.aboutamazon.com/workplace/employee-benefits. This position will remain posted until filled. Applicants should apply via our internal or external career site.
#J-18808-Ljbffr