Lead ETL Developer - PySpark/Python

Overview

On Site
Depends on Experience
Accepts corp to corp applications
Contract - W2
Contract - Independent
Contract - 12 Month(s)

Skills

ETL Developer
Python/Pyspark

Job Details

Job Summary:

We are seeking a highly skilled and experienced Senior ETL Developer with strong expertise in PySpark, Python, along with a solid background in ETL (Extract, Transform, Load) processes across relational databases and cloud platforms. The ideal candidate will have a minimum of years of hands-on

experience developing scalable data pipelines and ETL workflows using modern tools and frameworks.

Responsibilities:

  • Design, develop, and maintain ETL pipelines using PySpark and Python for large-scale data processing
  • Implement validation logic as part of ETL workflows.
  • Work with relational databases (.g., Oracle PostgreSQL, SQL Server) to extract, transform, and load data efficiently,
  • Develop and optimize data pipelines on cloud platforms (AWS) using Spark-based frameworks (Data Frames)
  • Collaborate with data architects, analysts, and other stakeholders to understand requirements and deliver reliable data solutions
  • Ensure data integrity, performance, and quality in all ETL processes.
  • Participate in code reviews, testing, and deployment of data solutions,
  • Work on-site in a U,S, office location four days per week, with one day optional for remote work,

Required Qualifications:

  • 5+ years of experience in ETL development with a strong foundation in data integration, transformation, and loading,
  • Proven experience writing efficient code in PySpark and Python, especially for large-scale data sets,
  • Hands-on experience with creating ETL workflows.
  • Strong experience working with relational databases and writing complex SQL queries
  • Solid understanding of data modeling, performance tuning, and best practices in ETL design.
  • Experience with cloud-based data processing using Spark on AWS, Azure, or Google Cloud Platform.
  • Strong communication and collaboration skills.
  • Must be authorized to work in the U.S. and willing to work from the office 4 days/week.

Preferred Qualifications:

  • Experience with data orchestration tools like Apache Airflow or similar.
  • Familiarity with data governance and security best practices.
  • Experience in Agile/Scrum environments.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.