Logo
Amazon

Software Development Engineer - Generative AI, AGIF | Inference Engine

Amazon, Boston, Massachusetts, us, 02298

Save Job

Software Development Engineer – Generative AI, AGIF | Inference Engine Join to apply for the

Software Development Engineer – Generative AI, AGIF | Inference Engine

role at

Amazon .

Responsibilities

Design, develop, test, and deploy high‑performance model inference capabilities, including multi‑modality, state‑of‑the‑art model architectures, latency, throughput, and cost optimizations.

Collaborate closely with engineers and scientists to influence overall strategy and define the team’s roadmap.

Drive system architecture, spearhead best practices, and mentor junior engineers.

Consult with scientists on emerging techniques, benchmark implementations, and implement production‑grade solutions that quickly reach deployment.

Collaborate with other science and engineering teams to deliver solutions and continually reduce operational load.

A Day in the Life • Consult with scientists to draw inspiration from emerging techniques and incorporate them into the roadmap. • Design and experiment with new algorithms from public and internal papers, benchmarking latency and accuracy. • Implement production‑grade solutions and advance them swiftly through deployment. • Collaborate with other teams to accomplish tasks and maintain operational excellence.

About the Team Our mission is to build best‑in‑class, fast, accurate, and cost‑efficient frontier model inference solutions and infrastructure that enable Amazon businesses to deliver maximum value to customers.

Basic Qualifications

3+ years of non‑internship professional software development experience.

Experience with software performance optimization.

Knowledge of Deep Learning and Transformer architectures.

Preferred Qualifications

3+ years of full software development life cycle experience, including coding standards, code reviews, source control management, build processes, testing, and operations.

Bachelor’s degree in computer science or equivalent.

Experience with Large Language Model inference.

Experience with GPU programming (TensorRT‑LLM).

Experience with Python, PyTorch, and C++ programming and performance optimization.

Experience with Trainium and Inferentia development.

Compensation The base pay for this position ranges from $129,300/year in our lowest geographic market up to $223,600/year in our highest geographic market. Pay is based on market location and may vary depending on job‑related knowledge, skills, and experience. Equity, sign‑on payments, and other components may be part of a total compensation package.

EEO Statement Amazon is an equal opportunity employer and does not discriminate on the basis of protected veteran status, disability, or other legally protected status. If you have a disability and need a workplace accommodation, please visit

https://amazon.jobs/content/en/how-we-hire/accommodations .

Company Information Amazon.com Services LLC Job ID: A3109601

#J-18808-Ljbffr