Andromeda
Simulation and Test Engineer (Conversational AI) - US Based
Andromeda, San Francisco, California, United States, 94199
Simulation and Test Engineer (Conversational AI) – US Based
5 days ago Be among the first 25 applicants
This range is provided by Andromeda. Your actual pay will be based on your skills and experience — talk with your recruiter to learn more.
Base pay range A$150,000.00/yr - A$250,000.00/yr
About Us Andromeda Robotics is an ambitious social robotics company with offices in Melbourne and San Francisco, dedicated to creating robots that seamlessly and intelligently interact with the human world. Our first robot, Abi, is a testament to this vision—a custom‑built platform designed from the ground up to be a helpful aid and intuitive partner in aged care homes. We are a passionate, collaborative team of engineers who solve some of the most challenging problems in AI and robotics. To accelerate our development and ensure Abi's reliability, we are seeking a foundational member to build out our capabilities to train and test our robot in simulation.
Our Values
Deeply empathetic - Kindness and compassion are at the heart of everything we do.
Purposely playful - Play sharpens focus. It keeps us curious, fast and obsessed with the craft.
Relentlessly striving - With relentless ambition, an action bias and constant curiosity, we don't settle.
Strong when it counts - Tenacious under pressure, we expect problems and stay in motion to adapt and progress.
United in action - Different minds. Shared mission. No passengers.
The Role We are looking for a creative and driven Simulation and Test Engineer to build Andromeda's testing infrastructure for our conversational AI systems and embodied character behaviours. Your immediate focus will be creating robust test systems for Abi's voice-to-voice chatbot, social awareness perception, and gesture motor control. As this infrastructure matures, you'll extend it into simulation environments for generating synthetic training data for character animation and gesture models.
The Team You’ll work at the intersection of our character software, robotics, perception, conversational AI, controls, and audio engineering teams. We bring deep expertise from autonomous vehicles and robotics, including simulation backgrounds. You'll collaborate with product owners and technical specialists to define requirements, integrate systems, and ensure quality across our AI/ML stack.
Requirements Phase 1: Build The Test Foundation
Define and stand up synthetic test environments for our AI/ML conversational stack
Conversational AI testing: voice-to-voice chat quality, response appropriateness, tool calling accuracy
Memory system testing: context retention, recall accuracy, conversation coherence
Audio modelling and testing: multi‑speaker scenarios, room acoustics, voice activity detection
Perception system testing: social awareness (face detection, gaze tracking, person tracking)
Gesture appropriateness testing: Working with our Controls/ML team, create test infrastructure to validate that Abi's body gestures
CI/CD and automated regression testing for all AI/ML subsystems
Custodian of quality metrics: if they don't exist, work with stakeholders to elicit use cases, derive requirements, and establish measurable quality metrics
Requirements formalisation: you're skilled at gathering, documenting, and tracing requirements back to test cases
Phase 2: Scale To ML Training Infrastructure
Extend test environments into training data generation pipelines
Investigate and stand up simulation tools (e.g. Unity, Unreal Engine, Isaac Sim) to support our machine learning pipeline with synthetic data and validation infrastructure
Build infrastructure for fine‑tuning character animation models on simulated multi‑actor scenarios
Enable ML-generated gesture development to augment hand‑crafted animation workflows
Create virtual environments with diverse social interaction scenarios for training and evaluation
Success In This Role Looks Like Months 1-3, stabilise our conversational system with automated regression tests and measurable quality benchmarks. By month 6, deliver an integrated simulation environment enabling rapid testing and iteration across our AI/ML stack. You'll design tests that push our systems beyond their limits and find what's brittle. Through trade studies and make‑vs‑buy decisions, you'll establish the infrastructure, set up automatic regression tests, and trace test cases back to high‑level requirements. You'll be the final guardian, verifying that our AI and machine learning systems work as intended before integration with Abi's physical platform. Your work will directly impact the speed and quality of our development, ensuring that every software build is robust, reliable, and safe.
Key Responsibilities
Architect and Build: Design, develop, and maintain scalable test infrastructure for conversational AI, perception, and gesture control systems
Own Testing Pipeline: Develop a robust CI/CD pipeline for automated regression testing, enabling rapid iteration and guaranteeing quality before deployment
Develop Test Scenarios: Create diverse audio environments, multi‑actor social scenarios, and edge cases to rigorously test Abi's conversational and social capabilities
Model with Fidelity: Implement accurate models of Abi's hardware stack (cameras, microphone array, upper body motion) as needed for test and simulation scenarios
Enable Future ML Training: Design test infrastructure with an eye towards evolution into a simulation platform for generating synthetic training data for character animation and gesture models
Integrate and Collaborate: Work closely with the robotics, AI, and software teams to seamlessly integrate their stacks into the test infrastructure and define testing requirements
Analyse and Improve: Develop metrics, tools, and dashboards to analyse test data, identify bugs, track performance, and provide actionable feedback to the engineering teams
Ideally You Have
Bachelor or Masters in Computer Science, Robotics, Engineering, or a related field
5+ years of professional experience testing complex AI/ML systems (conversational AI, perception systems, or embodied AI)
Strong programming proficiency in Python (essential); C++ experience valuable
Hands‑on experience with LLM testing, voice AI systems, or chatbot evaluation frameworks
Understanding of audio processing, speech recognition, and/or computer vision fundamentals
Experience with testing frameworks and CI/CD tools (pytest, Jenkins, GitHub Actions, etc.)
Familiarity with ML evaluation metrics and experimental design
A proactive, first‑principles thinker who is excited by the prospect of owning a critical system at an early‑stage startup
Bonus Points
Experience with simulation platforms (e.g. Unity, Unreal Engine, NVIDIA Isaac Sim, Gazebo) and physics engines
Experience with character animation systems, motion capture data, or gesture generation
Knowledge of reinforcement learning, imitation learning, or synthetic data generation for training ML models
Experience with 3D modelling tools and game engine content creation
Understanding of ROS2 for robotics integration
Knowledge of sensor modelling techniques for cameras and audio
Experience building and managing large‑scale, cloud‑based simulation infrastructure
PhD in a relevant field
Benefits The expected base salary range for this role, when performed in our San Francisco office, is $150,000 - $250,000 USD, depending on skills and experience. The salary for this position may vary depending on factors such as job‑related knowledge, skills, and experience. The total compensation package may also include additional benefits or components based on the specific role. Details will be provided if an employment offer is made. If you're excited about this role but don't meet every requirement, that's okay—we encourage you to apply. At Andromeda Robotics, we celebrate diversity and are committed to creating an inclusive environment for all employees. Let's build the future together.
We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.
#J-18808-Ljbffr
This range is provided by Andromeda. Your actual pay will be based on your skills and experience — talk with your recruiter to learn more.
Base pay range A$150,000.00/yr - A$250,000.00/yr
About Us Andromeda Robotics is an ambitious social robotics company with offices in Melbourne and San Francisco, dedicated to creating robots that seamlessly and intelligently interact with the human world. Our first robot, Abi, is a testament to this vision—a custom‑built platform designed from the ground up to be a helpful aid and intuitive partner in aged care homes. We are a passionate, collaborative team of engineers who solve some of the most challenging problems in AI and robotics. To accelerate our development and ensure Abi's reliability, we are seeking a foundational member to build out our capabilities to train and test our robot in simulation.
Our Values
Deeply empathetic - Kindness and compassion are at the heart of everything we do.
Purposely playful - Play sharpens focus. It keeps us curious, fast and obsessed with the craft.
Relentlessly striving - With relentless ambition, an action bias and constant curiosity, we don't settle.
Strong when it counts - Tenacious under pressure, we expect problems and stay in motion to adapt and progress.
United in action - Different minds. Shared mission. No passengers.
The Role We are looking for a creative and driven Simulation and Test Engineer to build Andromeda's testing infrastructure for our conversational AI systems and embodied character behaviours. Your immediate focus will be creating robust test systems for Abi's voice-to-voice chatbot, social awareness perception, and gesture motor control. As this infrastructure matures, you'll extend it into simulation environments for generating synthetic training data for character animation and gesture models.
The Team You’ll work at the intersection of our character software, robotics, perception, conversational AI, controls, and audio engineering teams. We bring deep expertise from autonomous vehicles and robotics, including simulation backgrounds. You'll collaborate with product owners and technical specialists to define requirements, integrate systems, and ensure quality across our AI/ML stack.
Requirements Phase 1: Build The Test Foundation
Define and stand up synthetic test environments for our AI/ML conversational stack
Conversational AI testing: voice-to-voice chat quality, response appropriateness, tool calling accuracy
Memory system testing: context retention, recall accuracy, conversation coherence
Audio modelling and testing: multi‑speaker scenarios, room acoustics, voice activity detection
Perception system testing: social awareness (face detection, gaze tracking, person tracking)
Gesture appropriateness testing: Working with our Controls/ML team, create test infrastructure to validate that Abi's body gestures
CI/CD and automated regression testing for all AI/ML subsystems
Custodian of quality metrics: if they don't exist, work with stakeholders to elicit use cases, derive requirements, and establish measurable quality metrics
Requirements formalisation: you're skilled at gathering, documenting, and tracing requirements back to test cases
Phase 2: Scale To ML Training Infrastructure
Extend test environments into training data generation pipelines
Investigate and stand up simulation tools (e.g. Unity, Unreal Engine, Isaac Sim) to support our machine learning pipeline with synthetic data and validation infrastructure
Build infrastructure for fine‑tuning character animation models on simulated multi‑actor scenarios
Enable ML-generated gesture development to augment hand‑crafted animation workflows
Create virtual environments with diverse social interaction scenarios for training and evaluation
Success In This Role Looks Like Months 1-3, stabilise our conversational system with automated regression tests and measurable quality benchmarks. By month 6, deliver an integrated simulation environment enabling rapid testing and iteration across our AI/ML stack. You'll design tests that push our systems beyond their limits and find what's brittle. Through trade studies and make‑vs‑buy decisions, you'll establish the infrastructure, set up automatic regression tests, and trace test cases back to high‑level requirements. You'll be the final guardian, verifying that our AI and machine learning systems work as intended before integration with Abi's physical platform. Your work will directly impact the speed and quality of our development, ensuring that every software build is robust, reliable, and safe.
Key Responsibilities
Architect and Build: Design, develop, and maintain scalable test infrastructure for conversational AI, perception, and gesture control systems
Own Testing Pipeline: Develop a robust CI/CD pipeline for automated regression testing, enabling rapid iteration and guaranteeing quality before deployment
Develop Test Scenarios: Create diverse audio environments, multi‑actor social scenarios, and edge cases to rigorously test Abi's conversational and social capabilities
Model with Fidelity: Implement accurate models of Abi's hardware stack (cameras, microphone array, upper body motion) as needed for test and simulation scenarios
Enable Future ML Training: Design test infrastructure with an eye towards evolution into a simulation platform for generating synthetic training data for character animation and gesture models
Integrate and Collaborate: Work closely with the robotics, AI, and software teams to seamlessly integrate their stacks into the test infrastructure and define testing requirements
Analyse and Improve: Develop metrics, tools, and dashboards to analyse test data, identify bugs, track performance, and provide actionable feedback to the engineering teams
Ideally You Have
Bachelor or Masters in Computer Science, Robotics, Engineering, or a related field
5+ years of professional experience testing complex AI/ML systems (conversational AI, perception systems, or embodied AI)
Strong programming proficiency in Python (essential); C++ experience valuable
Hands‑on experience with LLM testing, voice AI systems, or chatbot evaluation frameworks
Understanding of audio processing, speech recognition, and/or computer vision fundamentals
Experience with testing frameworks and CI/CD tools (pytest, Jenkins, GitHub Actions, etc.)
Familiarity with ML evaluation metrics and experimental design
A proactive, first‑principles thinker who is excited by the prospect of owning a critical system at an early‑stage startup
Bonus Points
Experience with simulation platforms (e.g. Unity, Unreal Engine, NVIDIA Isaac Sim, Gazebo) and physics engines
Experience with character animation systems, motion capture data, or gesture generation
Knowledge of reinforcement learning, imitation learning, or synthetic data generation for training ML models
Experience with 3D modelling tools and game engine content creation
Understanding of ROS2 for robotics integration
Knowledge of sensor modelling techniques for cameras and audio
Experience building and managing large‑scale, cloud‑based simulation infrastructure
PhD in a relevant field
Benefits The expected base salary range for this role, when performed in our San Francisco office, is $150,000 - $250,000 USD, depending on skills and experience. The salary for this position may vary depending on factors such as job‑related knowledge, skills, and experience. The total compensation package may also include additional benefits or components based on the specific role. Details will be provided if an employment offer is made. If you're excited about this role but don't meet every requirement, that's okay—we encourage you to apply. At Andromeda Robotics, we celebrate diversity and are committed to creating an inclusive environment for all employees. Let's build the future together.
We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.
#J-18808-Ljbffr