Position :: Data Integration Engineer
Location :: 100% Remote
Duration :: 6+ Months
Interview :: Phone and Video
Job Description:
Start Date: ASAP post Public Trust granted 3-4 weeks
Data Integration Engineer
Note: Looking for 2 to 4 years of experience … Healthcare is preferred, not mandatory.
Role Overview: seeking a Data Integration Engineer to support healthcare data integration efforts within an Azurebased data platform. This role is handson and deliveryfocused, with a strong emphasis on Pythondriven data pipelines, Azure Synapse, and healthcare interoperability (FHIR/HL7). A core expectation of this role is the ability to design, test, and validate data pipelines in environments where upstream specifications may be incomplete or inconsistent. The ideal candidate brings strong technical judgment, repeatable testing strategies, and the ability to raise data quality standards across the team.
Key Responsibilities
- Design, build, and maintain data pipelines in Azure Synapse using Python.
- Implement and operate a medallion data architecture (Bronze, Silver, Gold layers).
- Ingest, transform, and publish data in CSV, Parquet, and XML formats.
- Perform complex data mapping and transformation across healthcare data sources.
- Work directly with HL7 and FHIR healthcare data standards.
- Define and execute data pipeline testing strategies, including:
- Validation of transformations and mappings
- Data completeness, accuracy, and consistency checks
- Repeatable, teamadoptable testing approaches
- Operate effectively in situations where test cases or specs are not clearly provided, helping establish defensible validation criteria.
- Serve as a technical lead, setting patterns and best practices the broader team can follow.
- Collaborate closely with engineering, QA, and stakeholders to improve data quality and delivery outcomes.
Minimum Requirements:
This is a Health IT opportunity and previous experience working in Health IT and with healthcare data and data standards is preferred.
- 2–5 years of experience in data integration or data engineering roles.
- Hands-on experience with HL7, FHIR, X12, or similar healthcare data formats.
- Proficiency with Git for version control and collaborative development.
- Experience using Terraform to deploy or manage cloud infrastructure.
- General knowledge of cloud environments (Azure, AWS, or Google Cloud Platform).
- Working knowledge of Azure Synapse or similar cloud data platforms.
- Experience working with Parquet file formats in data engineering workflows.
- Strong SQL and/or Python skills for data manipulation and validation.