At Commure, our mission is to simplify healthcare. We have bold ambitions to reimagine the healthcare experience, setting a new standard for how care is delivered and experienced across the industry. Our growing suite of AI solutions spans ambient AI clinical documentation, provider copilots, autonomous coding, revenue cycle management and more — all designed for providers & administrators to focus on what matters most: providing care.
Healthcare is a $4.5 trillion industry with more than $500 billion spent annually on administrative costs, and Commure is at the heart of transforming it. We power over 500,000 clinicians across hundreds of care sites nationwide – more than $10 billion flows through our systems and we support over 100 million patient interactions. With new product launches on the horizon, expansion into additional care segments, and a bold vision to tackle healthcare's most pressing challenges, our ambition is to move from upstart innovator to the industry standard over the next few years.
Commure was recently named to Fortune’s Future 50 list for 2025 and is backed by world-class investors including General Catalyst, Sequoia, Y Combinator, Lux, Human Capital, 8VC, Greenoaks Capital, Elad Gil, and more. Commure has achieved over 300% year-over-year growth for the past two years and this is only the beginning. Healthcare's moment for AI-powered transformation is here, and we're building the technology to power it. Come join us in shaping the future of healthcare.
We’re building the next generation of data infrastructure to support administrative healthcare services (RCM, EHR, ambient scribing, and more). Our systems process large-scale patient-related data and require secure, performant, and compliant handling of sensitive healthcare information. We’re looking for a Data Engineer who can design, optimize, and extend our data pipelines and platforms, enabling analytics, observability, and secure development workflows.
Data Pipeline Development
Design, implement, and optimize ETL/ELT pipelines for large-scale PostgreSQL datasets (11TB+ production, 5TB staging)
Build scalable ingestion workflows into ClickHouse Cloud using Iceberg tables on AWS S3 and AWS Glue
Develop processes for anonymizing and preparing healthcare data in staging environments to support development and research without exposing PHI
Implement robust validation and reconciliation checks to ensure data quality and HIPAA-compliant handling
Data Modeling & Warehousing
Develop and maintain schemas to support both OLTP (PostgreSQL) and OLAP (ClickHouse/Iceberg) workloads
Optimize query performance for analytics while minimizing load on production databases
Extend our data warehouse to enable ad-hoc analysis, BI tool integrations, and healthcare-specific reporting use cases
Data Observability & Performance
Build tools and dashboards to monitor schema changes, query performance, and pipeline health across PostgreSQL, ClickHouse, and Glue/S3
Implement alerting, logging, and performance tuning strategies for production and staging environments
Collaborate with engineers and analysts to proactively identify bottlenecks and scalability improvements
Data Integrations & APIs
Integrate structured healthcare data flows between EHR systems, RCM platforms, and internal services
Build APIs or connectors to surface analytical and operational data securely to downstream consumers
Ensure interoperability across GCP (SQL) and AWS (ClickHouse, Glue, S3) platforms
Infrastructure & Governance
Manage and evolve our hybrid-cloud data infrastructure (GCP SQL + AWS Glue/ClickHouse)
Enforce access management, encryption, and anonymization controls aligned with HIPAA and healthcare compliance standards
Partner with security and compliance teams to implement best practices in sensitive data handling
Collaboration & Knowledge Sharing
Work closely with analysts, scribe technology developers, and product engineers to capture data requirements
Document schemas, pipelines, and workflows to ensure maintainability and cross-team understanding
Mentor team members and advocate for data engineering best practices across the company
What You Have
Bachelor’s degree in Computer Science, Engineering, Information Systems, or equivalent experience
Proven experience in data engineering at scale (multi-TB datasets, OLTP + OLAP systems)
Strong SQL expertise with PostgreSQL and experience tuning queries for high-volume transactional databases
Hands-on experience with Python, Java, and SQL for data processing and pipeline orchestration
Familiarity with ClickHouse or other analytical databases, and data lake formats (Iceberg, Parquet, ORC)
Experience with AWS Glue (ETL, catalog) and S3-based data lakes
Understanding of cloud-native services in both Google Cloud (Cloud SQL) and AWS
Knowledge of data anonymization and governance techniques for sensitive healthcare data (HIPAA familiarity a plus)
Experience with monitoring/observability tools for data infrastructure (e.g., Grafana, dbt metrics, or custom solutions)
Strong problem-solving and debugging skills; ability to balance technical rigor with business needs
Effective communicator and collaborator across engineering, analytics, and product teams
Commure + Athelas is committed to creating and fostering a diverse team. We are open to all backgrounds and levels of experience, and believe that great people can always find a place. We are committed to providing reasonable accommodations to all applicants throughout the application process.
Please be aware that all official communication from us will come exclusively from email addresses ending in @getathelas.com, @commure.com or @augmedix.com. Any emails from other domains are not affiliated with our organization.
Employees will act in accordance with the organization’s information security policies, to include but not limited to protecting assets from unauthorized access, disclosure, modification, destruction or interference nor execute particular security processes or activities. Employees will report to the information security office any confirmed or potential events or other risks to the organization. Employees will be required to attest to these requirements upon hire and on an annual basis.
If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.
Work on Yahoo’s Consumer Data team building and maintaining GCP-based data pipelines and GA4 integrations to enable analytics and product insights across Yahoo properties.
Lead the design and scaling of a modern, cloud-based data platform as a Data Engineering Manager for a US-based company operating remotely.
AssistRx is hiring a remote Business Intelligence Developer to build and optimize Tableau and ThoughtSpot analytics, create dbt data models on Snowflake, and support enterprise reporting needs.
Senior Data Engineer to lead end-to-end design and delivery of Snowflake/Airflow data products, drive data quality and observability, and introduce AI-driven automation across the platform.
A remote Data Migration Associate is needed to execute and validate platform-to-platform data transfers, support migration planning, and maintain accurate documentation to ensure data integrity.
Kobie is hiring a Senior Database Developer to build and tune scalable SQL-based data solutions for high-volume loyalty platforms in a flexible remote environment.
Lead enterprise data strategy and architecture for a fast-evolving fintech/retail finance platform, designing cloud-native data platforms, governance, and reporting to enable analytics-driven decision making.
Experienced data engineer needed to architect, build, and optimize Azure-based data pipelines and warehouse solutions for LEARFIELD's data strategy team.
Sandpiper Productions is hiring a part-time remote Data/AI Engineer to develop Python-based data pipelines, enhance web projects, and implement AI/LLM automations while working overlapping hours with EST.
KCI Technologies seeks a Data Support Specialist to manage, clean, validate, and present geospatial and tabular data for engineering projects while supporting field verification and client-facing workflows.
A cloud-focused Data Engineer (or Senior Data Engineer) role to build and optimize scalable GCP data pipelines and infrastructure that power AI/ML and LLM workflows for a fast-moving, collaborative team.
Lead and scale a modern data engineering team for a US-based organization, building robust data infrastructure, governance, and analytics pipelines in a hands-on management role.
Pfizer is hiring a Database Analyst (Associate) to develop and maintain clinical data capture solutions supporting clinical trials, data review, and regulatory submissions within the CDIS organization.
Commure is the first operating system designed for healthcare. By unifying disparate datasets, surfacing meaningful insights, accelerating performance through a suite of intuitive applications, and enabling seamless innovation across the industry,...
12 jobs