Luma AI
Product Engineer | Multimodal Interaction
Luma AI, Palo Alto, California, United States, 94306
Product Engineer | Multimodal Interaction
Join us to apply for the
Product Engineer | Multimodal Interaction
role at
Luma AI .
This role’s base pay range is $170,000 – $360,000 per year, depending on skills and experience.
The Opportunity Luma AI is defining the future of creative tools. We are moving beyond the prompt box to build intelligent interfaces where users collaborate with AI partners. We combine the research depth of a lab with the product obsession of a consumer app studio. You will work directly with world‑class researchers to productionize novel capabilities.
Where You Come In You will build the interface between human intention and machine intelligence. This role is about translating the capabilities of our multimodal models into magical, intuitive product experiences. You will solve the technical challenges of making complex, asynchronous agent actions feel responsive and alive.
What You Will Build
Visual Reasoning Systems: architect the backend systems that allow an agent to “see” a user’s canvas and make intelligent modifications.
Hybrid Workflows: build the bridges between synchronous user actions and asynchronous agent processing.
Research‑to‑Product Pipelines: partner with the research team to turn experimental model behaviors into stable, high‑fidelity product features.
The Profile We Are Looking For
Multimodal Experience: you have worked with systems involving video, images, or audio, and understand the unique challenges of media‑heavy applications.
Full‑Stack Fluency: while your focus is on the backend logic, you understand how API design impacts the frontend experience and latency.
Craft Obsession: you have a portfolio of high‑polish products and a passion for building tools that empower creators.
Seniority level Mid‑Senior level
Employment type Full‑time
Job function Engineering and Information Technology
Industries Software Development
Location Sunnyvale, CA
#J-18808-Ljbffr
Product Engineer | Multimodal Interaction
role at
Luma AI .
This role’s base pay range is $170,000 – $360,000 per year, depending on skills and experience.
The Opportunity Luma AI is defining the future of creative tools. We are moving beyond the prompt box to build intelligent interfaces where users collaborate with AI partners. We combine the research depth of a lab with the product obsession of a consumer app studio. You will work directly with world‑class researchers to productionize novel capabilities.
Where You Come In You will build the interface between human intention and machine intelligence. This role is about translating the capabilities of our multimodal models into magical, intuitive product experiences. You will solve the technical challenges of making complex, asynchronous agent actions feel responsive and alive.
What You Will Build
Visual Reasoning Systems: architect the backend systems that allow an agent to “see” a user’s canvas and make intelligent modifications.
Hybrid Workflows: build the bridges between synchronous user actions and asynchronous agent processing.
Research‑to‑Product Pipelines: partner with the research team to turn experimental model behaviors into stable, high‑fidelity product features.
The Profile We Are Looking For
Multimodal Experience: you have worked with systems involving video, images, or audio, and understand the unique challenges of media‑heavy applications.
Full‑Stack Fluency: while your focus is on the backend logic, you understand how API design impacts the frontend experience and latency.
Craft Obsession: you have a portfolio of high‑polish products and a passion for building tools that empower creators.
Seniority level Mid‑Senior level
Employment type Full‑time
Job function Engineering and Information Technology
Industries Software Development
Location Sunnyvale, CA
#J-18808-Ljbffr