Overview
Skills
Job Details
Core Responsibilities
Design, own, and maintain data pipelines that aggregate lab throughput, logistics, and customer behavior
Build and scale backend infrastructure to support real-time delivery coordination across a global network of 10+ labs
Maintain, optimize, and evolve the data warehouse architecture using tools like Snowflake and DBT
Develop robust models and analytical tools to monitor performance, ensure quality, and identify bottlenecks
Convert raw, complex datasets into actionable insights that drive product and operational decision-making
Work cross-functionally with engineers, scientists, and business stakeholders to ensure data systems reflect real-world processes
Lead root-cause investigations when anomalies appear in data, identifying upstream issues and system dynamics
Required Qualifications
5 10 years of experience in software or data engineering
At least 2 years working at a Seed to Series B startup with fewer than 100 employees
BS, MS, or PhD in Computer Science (or related field) from a top 40 U.S. university
Deep expertise in SQL for querying, modeling, and performance tuning
Proficiency in Python and its core data libraries (e.g., Pandas, NumPy)
Hands-on experience with ETL/ELT tools such as DBT, data warehouses like Snowflake, and backend frameworks like Flask or Django
Comfortable using Git and working in Linux-based environments
Strong sense of ownership and the ability to thrive in high-intensity environments (60+ hours/week expected)
Preferred Qualifications
Experience in biotech, genomics, or lab automation environments
Built data platforms supporting logistics, sequencing pipelines, or operational analytics
Prior experience as a founder or early engineering hire at a data-driven startup
Familiarity with modern BI tools such as Sigma
Track record of personal or side projects with tangible real-world impact