Amazon
Software Development Engineer - Generative AI, AGIF | Inference Engine
Amazon, New York, New York, us, 10261
Software Development Engineer – Generative AI, AGIF | Inference Engine
Description
Are you interested in advancing Amazon's Generative AI capabilities? Come work with a talented team of engineers and scientists in a highly collaborative and friendly team. We are building state‑of‑the‑art Generative AI technology that will benefit all Amazon businesses and customers.
Key job responsibilities
As a Software Development Engineer, you will be responsible for designing, developing, testing, and deploying high‑performance model inference capabilities, including but not limited to multimodality, SOTA model architectures, latency, throughput, and cost. You will collaborate closely with a team of engineers and scientists to influence our overall strategy, define the team’s roadmap, drive system architecture, spearhead best practices, and mentor junior engineers.
A day in the life
You will consult with scientists to get inspiration of emerging techniques and blend those into our roadmap; you will design and experiment with new algorithms from public and internal papers, benchmark the latency and accuracy of your implementations; most importantly you will implement production‑grade solutions, see them through deployments swiftly; you may need to collaborate with other science and engineering teams to get things done properly; you will hold the highest bar in operational excellence and support production systems, constantly creating solutions to minimize the ops load.
About The Team
Our mission is to build best‑in‑class, fast, accurate, and cost‑efficient frontier model inference solutions and infrastructure that will enable Amazon businesses to deliver more value to their customers.
Basic Qualifications
3+ years of non‑internship professional software development experience
Experience with software performance optimization
Knowledge of Deep Learning and Transformer architectures
Preferred Qualifications
3+ years of full software development life cycle experience, including coding standards, code reviews, source control management, build processes, testing, and operations experience
Bachelor's degree in computer science or equivalent
Experience with Large Language Model Inference
Experience with GPU programming (TensorRT‑LLM)
Experience with Python, PyTorch, and C++ programming and performance optimization
Experience with Trainium and Inferentia Development
Amazon is an equal opportunity employer and does not discriminate on the basis of protected veteran status, disability, or other legally protected status. Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, including support for the interview or onboarding process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information.
#J-18808-Ljbffr
Are you interested in advancing Amazon's Generative AI capabilities? Come work with a talented team of engineers and scientists in a highly collaborative and friendly team. We are building state‑of‑the‑art Generative AI technology that will benefit all Amazon businesses and customers.
Key job responsibilities
As a Software Development Engineer, you will be responsible for designing, developing, testing, and deploying high‑performance model inference capabilities, including but not limited to multimodality, SOTA model architectures, latency, throughput, and cost. You will collaborate closely with a team of engineers and scientists to influence our overall strategy, define the team’s roadmap, drive system architecture, spearhead best practices, and mentor junior engineers.
A day in the life
You will consult with scientists to get inspiration of emerging techniques and blend those into our roadmap; you will design and experiment with new algorithms from public and internal papers, benchmark the latency and accuracy of your implementations; most importantly you will implement production‑grade solutions, see them through deployments swiftly; you may need to collaborate with other science and engineering teams to get things done properly; you will hold the highest bar in operational excellence and support production systems, constantly creating solutions to minimize the ops load.
About The Team
Our mission is to build best‑in‑class, fast, accurate, and cost‑efficient frontier model inference solutions and infrastructure that will enable Amazon businesses to deliver more value to their customers.
Basic Qualifications
3+ years of non‑internship professional software development experience
Experience with software performance optimization
Knowledge of Deep Learning and Transformer architectures
Preferred Qualifications
3+ years of full software development life cycle experience, including coding standards, code reviews, source control management, build processes, testing, and operations experience
Bachelor's degree in computer science or equivalent
Experience with Large Language Model Inference
Experience with GPU programming (TensorRT‑LLM)
Experience with Python, PyTorch, and C++ programming and performance optimization
Experience with Trainium and Inferentia Development
Amazon is an equal opportunity employer and does not discriminate on the basis of protected veteran status, disability, or other legally protected status. Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process, including support for the interview or onboarding process, please visit https://amazon.jobs/content/en/how-we-hire/accommodations for more information.
#J-18808-Ljbffr