Logo
Tesla Motors, Inc.

AI Engineer, ML Inference Optimization, Autonomy & Robotics

Tesla Motors, Inc., Palo Alto, California, United States, 94306

Save Job

What to Expect Tesla's AI team is pushing the frontier of real-world machine learning, building models that reason, predict, and act with human-level physical intelligence. We train and deploy large-scale ML systems powering products from Autopilot to Optimus.

As part of the Model Optimization group, you will work at the intersection of machine learning and systems, designing our most advanced models to run efficiently across Tesla's diverse compute stack, from data centers to edge AI accelerators. You will design the model architecture and engineer algorithmic optimizations that make large-scale model inference fast, reliable, and hardware-aware.

What You'll Do

Design, train, and deploy large neural networks that run efficiently on heterogeneous hardware (GPU, CPU, Tesla's in-house AI ASIC)

Develop and integrate quantization, sparsity, pruning, and distillation techniques to improve inference performance

Design inference algorithms that improve inference performance in terms of quantization and latency

Profile and improve latency, throughput, and memory efficiency for large ML models across edge and cloud environments

Collaborate with compiler and hardware engineers to co-design architectures for efficient real-time inference

Design and implement custom GPU kernels (CUDA / OpenCL) to accelerate model operations and post-processing pipelines

Conduct systematic benchmarking, scaling, and validation of inference performance across Tesla platforms

What You'll Bring

Proven experience in scaling and optimizing inference for large ML models, particularly transformers or similar architectures

Familiarity with quantization‑aware training, model compression, and distillation for edge and real‑time inference

Proficiency with Python and C++ (modern standards 14/17/20) and deep learning frameworks such as PyTorch, TensorFlow, or JAX

Strong understanding of computer systems and architecture, with experience deploying ML models on GPUs, TPUs, or NPUs

Hands‑on expertise with CUDA programming, low‑level performance profiling, and compiler‑level optimization (TensorRT, TVM, XLA)

Experience collaborating with compiler/hardware engineers to bridge model and system‑level optimization

Excellent problem‑solving skills and the ability to debug and tune high‑performance inference workloads

Compensation and Benefits Benefits Along with competitive pay, as a full‑time Tesla employee, you are eligible for the following benefits at day 1 of hire:

Aetna PPO and HSA plans > 2 medical plan options with $0 payroll deduction

Family‑building, fertility, adoption and surrogacy benefits

Dental (including orthodontic coverage) and vision plans, both have options with a $0 paycheck contribution

Company Paid (Health Savings Account) HSA Contribution when enrolled in the High Deductible Aetna medical plan with HSA

Healthcare and Dependent Care Flexible Spending Accounts (FSA)

401(k) with employer match, Employee Stock Purchase Plans, and other financial benefits

Company paid Basic Life, AD&D, short‑term and long‑term disability insurance

Employee Assistance Program

Sick and Vacation time (Flex time for salary positions), and Paid Holidays

Back‑up childcare and parenting support resources

Voluntary benefits to include: critical illness, hospital indemnity, accident insurance, theft & legal services, and pet insurance

Weight Loss and Tobacco Cessation Programs

Tesla Babies program

Commuter benefits

Employee discounts and perks program

Expected Compensation $124,000 - $420,000/annual salary + cash and stock awards + benefits Pay offered may vary depending on multiple individualized factors, including market location, job‑related knowledge, skills, and experience. The total compensation package for this position may also include other elements dependent on the position offered. Details of participation in these benefit plans will be provided if an employee receives an offer of employment.

#J-18808-Ljbffr