Overview
Remote
Depends on Experience
Accepts corp to corp applications
Contract - Independent
Contract - W2
Contract - 12 Month(s)
No Travel Required
Unable to Provide Sponsorship
Skills
Data
Hadoop
Python
Job Details
One of our HealthCare clients in Texas is urgently looking for a Data Engineer.
Job Summary
We are seeking an experienced Data Engineer with strong expertise in Hadoop ecosystems and Python to design, build, and maintain scalable data solutions for a healthcare client. The ideal candidate will work with large, complex healthcare datasets to enable analytics, reporting, and data-driven decision-making while ensuring data quality, security, and compliance.
Key Responsibilities
- Design, develop, and maintain scalable data pipelines using Hadoop and related big data technologies
- Process and transform large volumes of structured and unstructured healthcare data
- Develop and optimize ETL/ELT workflows using Python
- Work with Hadoop components such as HDFS, Hive, Spark, and YARN
- Collaborate with data scientists, analysts, and business stakeholders to support analytics and reporting needs
- Ensure data quality, integrity, and reliability across data platforms
- Implement performance tuning and optimization for big data jobs
- Adhere to healthcare data standards and regulatory requirements (e.g., HIPAA)
- Support data ingestion from multiple sources such as databases, APIs, and flat files
- Troubleshoot and resolve data pipeline and production issues
Required Qualifications
- Bachelor’s degree in Computer Science, Engineering, or a related field
- 7+ years of experience as a Data Engineer or similar role
- Strong hands-on experience with Hadoop ecosystem (HDFS, Hive, Spark)
- Proficiency in Python for data processing and automation
- Experience with SQL and data warehousing concepts
- Understanding of data modeling and big data architecture
Preferred Qualifications
- Experience working with healthcare data (claims, EHR/EMR, clinical, or payer data)
- Knowledge of Spark (PySpark) for distributed data processing
- Experience with workflow orchestration tools (e.g., Airflow, Oozie)
- Exposure to cloud platforms (AWS, Azure, or Google Cloud Platform)
- Understanding of healthcare standards such as HL7, FHIR
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.