Let’s get started
By clicking ‘Next’, I agree to the Terms of Service
and Privacy Policy, and consent to receive emails from Rise
Jobs / Job page
Simulation and Test Engineer (Conversational AI) - US Based image - Rise Careers
Job details

Simulation and Test Engineer (Conversational AI) - US Based

About Us

Andromeda Robotics is an ambitious social robotics company with offices in Melbourne and San Francisco, dedicated to creating robots that seamlessly and intelligently interact with the human world. Our first robot, Abi, is a testament to this vision—a custom-built platform designed from the ground up to be a helpful aid and intuitive partner in aged care homes. We are a passionate, collaborative team of engineers who solve some of the most challenging problems in AI and robotics. To accelerate our development and ensure Abi's reliability, we are seeking a foundational member to build out our capabilities to train and test our robot in simulation. 

Our Values

Deeply empathetic - Kindness and compassion are at the heart of everything we do.

Purposely playful - Play sharpens focus. It keeps us curious, fast and obsessed with the craft.

Relentlessly striving - With relentless ambition, an action bias and constant curiosity, we don’t settle.

Strong when it counts - Tenacious under pressure, we expect problems and stay in motion to adapt and progress.

United in action - Different minds. Shared mission. No passengers.

The Role

We are looking for a creative and driven Simulation and Test Engineer to build Andromeda's testing infrastructure for our conversational AI systems and embodied character behaviours. Your immediate focus will be creating robust test systems for Abi's voice-to-voice chatbot, social awareness perception, and gesture motor control. As this infrastructure matures, you'll extend it into simulation environments for generating synthetic training data for character animation and gesture models.

The Team

You'll work at the intersection of our character software, robotics, perception, conversational AI, controls, and audio engineering teams. We bring deep expertise from autonomous vehicles and robotics, including simulation backgrounds. You'll collaborate with product owners and technical specialists to define requirements, integrate systems, and ensure quality across our AI/ML stack.

Phase 1: Build The Test Foundation

  • Define and stand up synthetic test environments for our AI/ML conversational stack
  • Conversational AI testing: voice-to-voice chat quality, response appropriateness, tool calling accuracy
  • Memory system testing: context retention, recall accuracy, conversation coherence
  • Audio modelling and testing: multi-speaker scenarios, room acoustics, voice activity detection
  • Perception system testing: social awareness (face detection, gaze tracking, person tracking)
  • Gesture appropriateness testing: Working with our Controls/ML team, create test infrastructure to validate that Abi's body gestures
  • CI/CD and automated regression testing for all AI/ML subsystems
  • Custodian of quality metrics: if they don't exist, work with stakeholders to elicit use cases, derive requirements, and establish measurable quality metrics
  • Requirements formalisation: you're skilled at gathering, documenting, and tracing requirements back to test cases

Phase 2: Scale To ML Training Infrastructure

Our approach to gesture generation requires high-fidelity synthetic interaction data at scale. You'll investigate and build the infrastructure to generate this data, working closely with our character software team to define requirements and validate approaches.

  • Extend test environments into training data generation pipelines
  • Investigate and stand up simulation tools (e.g. Unity, Unreal Engine, Isaac Sim) to support our machine learning pipeline with synthetic data and validation infrastructure
  • Build infrastructure for fine-tuning character animation models on simulated multi-actor scenarios
  • Enable ML-generated gesture development to augment hand-crafted animation workflows
  • Create virtual environments with diverse social interaction scenarios for training and evaluation

Success In This Role Looks Like

Months 1-3, stabilise our conversational system with automated regression tests and measurable quality benchmarks. By month 6, deliver an integrated simulation environment enabling rapid testing and iteration across our AI/ML stack.

You'll design tests that push our systems beyond their limits and find what's brittle. Through trade studies and make-vs-buy decisions, you'll establish the infrastructure, set up automatic regression tests, and trace test cases back to high-level requirements. You'll be the final guardian, verifying that our AI and machine learning systems work as intended before integration with Abi's physical platform. Your work will directly impact the speed and quality of our development, ensuring that every software build is robust, reliable, and safe.

Key Responsibilities

  • Architect and Build: Design, develop, and maintain scalable test infrastructure for conversational AI, perception, and gesture control systems
  • Own Testing Pipeline: Develop a robust CI/CD pipeline for automated regression testing, enabling rapid iteration and guaranteeing quality before deployment
  • Develop Test Scenarios: Create diverse audio environments, multi-actor social scenarios, and edge cases to rigorously test Abi's conversational and social capabilities
  • Model with Fidelity: Implement accurate models of Abi's hardware stack (cameras, microphone array, upper body motion) as needed for test and simulation scenarios
  • Enable Future ML Training: Design test infrastructure with an eye towards evolution into a simulation platform for generating synthetic training data for character animation and gesture models
  • Integrate and Collaborate: Work closely with the robotics, AI, and software teams to seamlessly integrate their stacks into the test infrastructure and define testing requirements
  • Analyse and Improve: Develop metrics, tools, and dashboards to analyse test data, identify bugs, track performance, and provide actionable feedback to the engineering teams

Ideally You Have

  • Bachelor or Masters in Computer Science, Robotics, Engineering, or a related field
  • 5+ years of professional experience testing complex AI/ML systems (conversational AI, perception systems, or embodied AI)
  • Strong programming proficiency in Python (essential); C++ experience valuable
  • Hands-on experience with LLM testing, voice AI systems, or chatbot evaluation frameworks
  • Understanding of audio processing, speech recognition, and/or computer vision fundamentals
  • Experience with testing frameworks and CI/CD tools (pytest, Jenkins, GitHub Actions, etc.)
  • Familiarity with ML evaluation metrics and experimental design
  • A proactive, first-principles thinker who is excited by the prospect of owning a critical system at an early-stage startup

Bonus Points

  • Experience with simulation platforms (e.g. Unity, Unreal Engine, NVIDIA Isaac Sim, Gazebo) and physics engines
  • Experience with character animation systems, motion capture data, or gesture generation
  • Knowledge of reinforcement learning, imitation learning, or synthetic data generation for training ML models
  • Experience with 3D modelling tools and game engine content creation
  • Understanding of ROS2 for robotics integration
  • Knowledge of sensor modelling techniques for cameras and audio
  • Experience building and managing large-scale, cloud-based simulation infrastructure
  • PhD in a relevant field

The expected base salary range for this role, when performed in our San Francisco office, is $150,000 - $250,000 USD, depending on skills and experience. The salary for this position may vary depending on factors such as job-related knowledge, skills, and experience. The total compensation package may also include additional benefits or components based on the specific role. Details will be provided if an employment offer is made.

If you’re excited about this role but don’t meet every requirement, that’s okay—we encourage you to apply. At Andromeda Robotics, we celebrate diversity and are committed to creating an inclusive environment for all employees. Let’s build the future together.

We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.

Average salary estimate

$200000 / YEARLY (est.)
min
max
$150000K
$250000K

If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.

Similar Jobs
Photo of the Rise User
Posted 3 hours ago

Lead Java-focused quality assurance for AI data training projects, providing code reviews, technical evaluations, and contributor support on a remote US-only contract.

jsrglobal Hybrid USA - NC - Durham - Hamlin Rd
Posted 23 hours ago

KBI Biopharma is hiring a hands-on Data Reviewer / Associate Scientist to perform GMP analytical data review and ensure accuracy of documentation for biopharmaceutical projects.

Photo of the Rise User
Posted 2 hours ago

Technical QA Lead specializing in HTML/CSS to review front-end training data, guide contributors, and maintain quality standards for remote AI data projects.

Photo of the Rise User
Posted 3 hours ago

Evolv Technology is hiring a Quality Analytics Intern to analyze quality data, build automated dashboards, and support reliability improvements for fielded product platforms.

Photo of the Rise User
Posted 7 hours ago

Iru is hiring a Quality Engineer to own functional testing for a development team and grow Playwright-based automation for its AI-driven security platform.

MATCH
Calculating your matching score...
FUNDING
SENIORITY LEVEL REQUIREMENT
TEAM SIZE
No info
HQ LOCATION
No info
EMPLOYMENT TYPE
Full-time, hybrid
DATE POSTED
November 28, 2025
Risa star 🔮 Hi, I'm Risa! Your AI
Career Copilot
Want to see a list of jobs tailored to
you, just ask me below!