Data Engineer Databricks | PySpark | Scala

Overview

On Site
Depends on Experience
Contract - Independent
Contract - W2
Contract - 12 Month(s)
Able to Provide Sponsorship

Skills

PySpark
Databricks
Data Engineering
Extract
Transform
Load
Scala

Job Details

Job Description Summary:

We are seeking an experienced Data Engineer with 9+ years of experience in enterprise-scale data engineering environments. The ideal candidate should have deep expertise in Databricks, PySpark, and Scala, with a strong background in building and optimizing ETL pipelines and managing Delta Lake architectures. Strong SQL skills and experience with data modeling and Spark performance tuning are essential. Onsite, face-to-face interview in Dallas, TX is mandatory.

Roles and Responsibilities:

  • Design and develop scalable ETL pipelines using Databricks, PySpark, and Scala
  • Manage and optimize Delta Lake architectures for batch and streaming data processing
  • Fine-tune and troubleshoot high-volume Apache Spark jobs in a distributed environment
  • Build and maintain scalable data models to support reporting and analytics use cases
  • Develop complex SQL queries and implement data quality and validation checks
  • Collaborate with architects, analysts, and DevOps teams to deliver reliable data solutions
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.