Healthcare is in crisis and the people behind the results deserve better. With data exploding across wearables, lab tests, and patient–doctor interactions, we’re entering an era where data is abundant.
Junction is building the infrastructure layer for diagnostic healthcare, making patient data accessible, actionable, and automated across labs and devices. Our mission is simple but ambitious: use health data to unlock unprecedented insight into human health and disease.
If you're passionate about how technology can supercharge healthcare, you’ll fit right in.
Backed by Creandum, Point Nine, 20VC, YC, and leading angels, we’re working to solve one of the biggest challenges of our time: making healthcare personalized, proactive, and affordable. We’re already connecting millions and scaling fast.
Short on time?
Who you are: A data engineer with solid software engineering fundamentals who can build, own, and scale reliable data pipelines and warehouse infrastructure.
Ownership: You’ll shape our data foundation from ingestion through transformation — and make it analytics-ready at scale.
Salary: $160K - $200k+ equity
Time zone: Preferably NYC; EST required.
Why we need you
Junction powers modern diagnostics at scale and as we grow, our platform is becoming increasingly data-intensive. The way we move, structure, and surface data directly affects our ability to support customers, deliver real-time insights, and unlock the next generation of diagnostics products.
We’re hiring our first Data Engineer to take ownership of that foundation.
Build and run pipelines that turn raw, messy healthcare data into clean, trusted, usable information
Power customer products, internal analytics, and the AI models behind our next wave of diagnostics
Design how data flows through an entire diagnostics ecosystem — not just maintain ETLs
Build scalable, cloud-native pipelines on GCP and eliminate bottlenecks as we scale
Hunt down edge cases, build guardrails for quality, and ship systems other engineers rely on daily
If you love untangling complexity and building data systems that truly make an impact, you’ll fit right in — and the systems you build will unlock new products and accelerate everything we ship.
What you’ll be doing day to day
Designing and operating ingestion, transformation, and replication pipelines on GCP
Managing orchestration and streamlining ELT/ETL workflows (e.g., Temporal)
Creating clean, scalable, analytics-ready schemas in BigQuery
Implementing monitoring, alerting, testing, and observability across data flows
Integrating data from APIs, operational databases, and unstructured sources
Collaborating with product, engineering, analytics, and compliance on secure, high-quality data delivery
Requirements
Solid engineering fundamentals and experience building pipelines from scratch
Python and SQL fluency; comfortable across relational + NoSQL systems
Experience with orchestrators like Temporal, Airflow, or Dagster
Hands-on with BigQuery, BigTable, and core GCP data tooling
Ability to turn messy, ambiguous data problems into clear, scalable solutions
Startup or small-team experience; comfortable moving fast with ownership
Communication skills, attention to detail, and a bias toward clarity and reliability
You don’t need to tick every box to fit in here. If the problems we’re solving genuinely interest you and you know you can contribute, we’d love to talk.
Nice to have
Experience with HIPAA/PHI or regulated healthcare data
Background with time-series data or event-driven architectures
Familiarity with dbt or similar transformation frameworks
Experience with healthcare, diagnostics, or ML/AI workloads
How you'll be compensated
Salary: $160K - $200k + early stage options
Your salary is dependant on your location and experience level, generated by our salary calculator. Read more in our handbook here.
Generous early stage options (extended exercise post 2 years employment) - you will receive 3 offers based on how much equity you'd like
Regular in person offsites, last were in Morocco and Tenerife
Bi-weekly team happy hours & events remotely
Monthly learning budget of $300 for personal development/productivity
Flexible, remote-first working - including $1K for home office equipment
25 days off a year + national holidays
Healthcare cover depending on location
Oh and before we forget:
Backend Stack: Python (FastAPI), Go, PostgreSQL, Google Cloud Platform (Cloud Run, GKE, Cloud BigTable, etc), Temporal Cloud
Frontend Stack: TypeScript, Next.js
API docs are here: https://docs.junction.com/
Company handbook is here with engineering values + principles
Important details before applying:
We only hire folks physically based in GMT and EST timezones - more information here.
We do not sponsor visas right now given our stage
If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.
Senior Associate Data Engineer to develop and maintain ETL/Databricks pipelines and integrations that power Rx Savings Solutions' analytics and operational reporting.
UC Irvine Libraries is hiring a Digital Initiatives Librarian to manage and grow digital collections and infrastructure while supporting digital scholarship, rights review, and campus outreach.
Lead enterprise risk analytics strategy and architecture at Lilly to turn cross-functional risk data into executive insights that drive strategic decisions and measurable business outcomes.
Promise is hiring a Data Engineer to build and scale reliable data pipelines, warehouses, and data products that enable data-driven decisions across product, operations, and finance.
Lead the Commercial Real Estate BI function to design scalable analytics, reporting, and visualizations that drive strategic portfolio and executive decisions at Flagstar Bank.
Artsy is hiring an experienced data leader to define and operationalize a company-wide data strategy, translating analytics into product and revenue growth for its marketplace.
Senior Data Engineer needed to design and maintain complex ETL and backend systems (Python, NodeJS, Java) for Elsevier's clinical AI and healthcare data products.
Amigo is hiring a Senior Software Engineer (Data) to design and operate Databricks streaming and batch pipelines that enable analytics, research, and continuous improvement of clinical AI agents.
Join Flyway Health as a founding Data/AI engineer to build and scale AI-driven data pipelines and agent systems that power enterprise life-sciences insights.
F5 is hiring a seasoned Director of Enterprise Data Management to build enterprise-scale data governance, security, and AI oversight that enables responsible, high-impact analytics.
Exegy is hiring a Technical Business Analyst in New York to support market data modeling, normalization, and feed integration for its market data engineering team.
Senior Data Engineer needed to lead real-time data ingestion, transformation, and visualization across cloud and streaming platforms for Highmark/enGen (U.S. citizens only).
NVIDIA seeks a Principal Data Platform Architect to lead architecture and delivery of large-scale observability and data platforms for AI and HPC clusters.