General Motors
Overview
Join to apply for the
Staff ML Engineer, Inference Platform
role at
General Motors . This position is eligible for relocation assistance. About the Team
The ML Inference Platform is part of the AI Compute Platforms organization within Infrastructure Platforms. Our team owns the cloud-agnostic, reliable, and cost-efficient platform that powers GM’s AI efforts. We support teams developing autonomous vehicles (L3/L4/L5) and other AI-driven products, enabling rapid innovation by optimizing for high-priority ML-centric use cases. Our platform supports serving state-of-the-art (SOTA) ML models for experimental and bulk inference, with a focus on performance, availability, concurrency, and scalability. We’re committed to maximizing GPU utilization across platforms (B200, H100, A100, and more) while maintaining reliability and cost efficiency. About the Role
We are seeking a Staff ML Infrastructure engineer to help build and scale robust Compute platforms for ML workflows. You’ll work closely with ML engineers and researchers to ensure efficient model serving and inference in production for workflows such as data mining, labeling, model distillation, simulations, and more. This is a high-impact opportunity to influence the future of AI infrastructure at GM. You will shape the architecture, roadmap, and user-experience of a robust ML inference service supporting real-time, batch, and experimental inference needs. The ideal candidate has experience designing distributed systems for ML, strong problem-solving skills, and a product mindset focused on platform usability and reliability. What you’ll be doing
Design and implement core platform backend software components. Collaborate with ML engineers and researchers to understand critical workflows, translate them into platform requirements, and deliver incremental value. Lead technical decision-making on model serving strategies, orchestration, caching, model versioning, and auto-scaling mechanisms. Drive monitoring, observability, and metrics to ensure reliability, performance, and resource optimization of inference services. Proactively research and integrate state-of-the-art model serving frameworks, hardware accelerators, and distributed computing techniques. Lead large-scale technical initiatives across GM’s ML ecosystem. Raise the engineering bar through technical leadership and best-practices. Contribute to open source projects; represent GM in relevant communities. Minimum Requirements
8+ years of industry experience, with focus on machine learning systems or high-performance backend services. Expertise in Go, Python, C++, or other relevant coding languages. Expertise in ML inference, model serving frameworks (e.g., Triton, RayServe, vLLM). Strong communication skills and a proven ability to drive cross-functional initiatives. Experience with cloud platforms such as GCP, Azure, or AWS. Ability to thrive in a dynamic, multi-tasking environment with evolving priorities. Preferred Qualifications
Hands-on experience building ML infrastructure platforms for model serving/inference. Experience designing interfaces, APIs, and clients for ML workflows. Experience with Ray framework and/or vLLM. Experience with distributed systems and large-scale data processing. Telemetry familiarity to inform product improvements. Familiarity with hardware acceleration (GPUs) and optimizations for inference workloads. Contributions to open-source ML serving frameworks. Compensation and Benefits
Compensation: The compensation information is a good faith estimate only. It is based on what a successful applicant might be paid in accordance with applicable state laws. The compensation may not be representative for positions outside of New York, Colorado, California, or Washington. Compensation: The expected base compensation for this role is: $195,000 - $298,000. Actual base compensation within the identified range will vary based on factors relevant to the position. Bonus Potential: An incentive pay program offers payouts based on company performance, job level, and individual performance. About GM
Our vision is a world with Zero Crashes, Zero Emissions and Zero Congestion. We embrace the responsibility to lead the change that will make our world better, safer and more equitable for all. Why Join Us
We believe in making a daily choice to drive meaningful change through our words, deeds, and culture. We want every employee to feel they belong to one General Motors team. Benefits Overview
From day one, we support your well-being at work and at home. Learn how GM supports a rewarding career by visiting Total Rewards resources. Non-Discrimination and Equal Employment Opportunities (U.S.)
General Motors is committed to a workplace free of unlawful discrimination and one that fosters inclusion and belonging. All employment decisions are made on a non-discriminatory basis without regard to sex, race, color, national origin, citizenship status, religion, age, disability, pregnancy or maternity status, sexual orientation, gender identity, veteran status, or any other protected status in accordance with federal, state and local laws. Applicants may be required to successfully complete role-related assessments and/or pre-employment screening where applicable. To learn more, visit How we Hire. Accommodations
General Motors offers accommodations to all job seekers. If you need a reasonable accommodation to assist with your job search or application, email us or call 800-865-7580. Please include the job title and requisition number.
#J-18808-Ljbffr
Join to apply for the
Staff ML Engineer, Inference Platform
role at
General Motors . This position is eligible for relocation assistance. About the Team
The ML Inference Platform is part of the AI Compute Platforms organization within Infrastructure Platforms. Our team owns the cloud-agnostic, reliable, and cost-efficient platform that powers GM’s AI efforts. We support teams developing autonomous vehicles (L3/L4/L5) and other AI-driven products, enabling rapid innovation by optimizing for high-priority ML-centric use cases. Our platform supports serving state-of-the-art (SOTA) ML models for experimental and bulk inference, with a focus on performance, availability, concurrency, and scalability. We’re committed to maximizing GPU utilization across platforms (B200, H100, A100, and more) while maintaining reliability and cost efficiency. About the Role
We are seeking a Staff ML Infrastructure engineer to help build and scale robust Compute platforms for ML workflows. You’ll work closely with ML engineers and researchers to ensure efficient model serving and inference in production for workflows such as data mining, labeling, model distillation, simulations, and more. This is a high-impact opportunity to influence the future of AI infrastructure at GM. You will shape the architecture, roadmap, and user-experience of a robust ML inference service supporting real-time, batch, and experimental inference needs. The ideal candidate has experience designing distributed systems for ML, strong problem-solving skills, and a product mindset focused on platform usability and reliability. What you’ll be doing
Design and implement core platform backend software components. Collaborate with ML engineers and researchers to understand critical workflows, translate them into platform requirements, and deliver incremental value. Lead technical decision-making on model serving strategies, orchestration, caching, model versioning, and auto-scaling mechanisms. Drive monitoring, observability, and metrics to ensure reliability, performance, and resource optimization of inference services. Proactively research and integrate state-of-the-art model serving frameworks, hardware accelerators, and distributed computing techniques. Lead large-scale technical initiatives across GM’s ML ecosystem. Raise the engineering bar through technical leadership and best-practices. Contribute to open source projects; represent GM in relevant communities. Minimum Requirements
8+ years of industry experience, with focus on machine learning systems or high-performance backend services. Expertise in Go, Python, C++, or other relevant coding languages. Expertise in ML inference, model serving frameworks (e.g., Triton, RayServe, vLLM). Strong communication skills and a proven ability to drive cross-functional initiatives. Experience with cloud platforms such as GCP, Azure, or AWS. Ability to thrive in a dynamic, multi-tasking environment with evolving priorities. Preferred Qualifications
Hands-on experience building ML infrastructure platforms for model serving/inference. Experience designing interfaces, APIs, and clients for ML workflows. Experience with Ray framework and/or vLLM. Experience with distributed systems and large-scale data processing. Telemetry familiarity to inform product improvements. Familiarity with hardware acceleration (GPUs) and optimizations for inference workloads. Contributions to open-source ML serving frameworks. Compensation and Benefits
Compensation: The compensation information is a good faith estimate only. It is based on what a successful applicant might be paid in accordance with applicable state laws. The compensation may not be representative for positions outside of New York, Colorado, California, or Washington. Compensation: The expected base compensation for this role is: $195,000 - $298,000. Actual base compensation within the identified range will vary based on factors relevant to the position. Bonus Potential: An incentive pay program offers payouts based on company performance, job level, and individual performance. About GM
Our vision is a world with Zero Crashes, Zero Emissions and Zero Congestion. We embrace the responsibility to lead the change that will make our world better, safer and more equitable for all. Why Join Us
We believe in making a daily choice to drive meaningful change through our words, deeds, and culture. We want every employee to feel they belong to one General Motors team. Benefits Overview
From day one, we support your well-being at work and at home. Learn how GM supports a rewarding career by visiting Total Rewards resources. Non-Discrimination and Equal Employment Opportunities (U.S.)
General Motors is committed to a workplace free of unlawful discrimination and one that fosters inclusion and belonging. All employment decisions are made on a non-discriminatory basis without regard to sex, race, color, national origin, citizenship status, religion, age, disability, pregnancy or maternity status, sexual orientation, gender identity, veteran status, or any other protected status in accordance with federal, state and local laws. Applicants may be required to successfully complete role-related assessments and/or pre-employment screening where applicable. To learn more, visit How we Hire. Accommodations
General Motors offers accommodations to all job seekers. If you need a reasonable accommodation to assist with your job search or application, email us or call 800-865-7580. Please include the job title and requisition number.
#J-18808-Ljbffr