Overview
On Site
Contract - W2
Contract - 6
Skills
Java
Scala
pyspark
Snowflake
APACHE SPARK
Job Details
Data Engineer Snowflake & Apache Spark
Location : Plano, TX (5 Days Onsite)
Contract-6 Months+
Location : Plano, TX (5 Days Onsite)
Contract-6 Months+
Required Qualifications:
Certifications:
SnowPro Core / Advanced Certification (e.g., SnowPro Advanced: Architect, Data Engineer, etc.)
Databricks Certified Associate Developer for Apache Spark (latest version preferred)
3+ years of experience working with Snowflake, including schema design, query optimization, and Snowpipe/Streams/Tasks.
2+ years of hands-on development with Apache Spark (PySpark, Scala, or Java) in Databricks or open-source environments.
Strong understanding of distributed computing, data lakes, and modern data architectures.
Certification is Mandatory.
Certifications:
SnowPro Core / Advanced Certification (e.g., SnowPro Advanced: Architect, Data Engineer, etc.)
Databricks Certified Associate Developer for Apache Spark (latest version preferred)
3+ years of experience working with Snowflake, including schema design, query optimization, and Snowpipe/Streams/Tasks.
2+ years of hands-on development with Apache Spark (PySpark, Scala, or Java) in Databricks or open-source environments.
Strong understanding of distributed computing, data lakes, and modern data architectures.
Certification is Mandatory.
Technical Skills:
Proficient in SQL, Spark (RDD/DataFrame APIs), and Python or Scala
Experience with cloud platforms (AWS, Azure, or Google Cloud Platform), especially integrating Snowflake and Databricks
Familiarity with data modeling, data quality, and orchestration tools (e.g., Airflow, Prefect)
Knowledge of CI/CD pipelines and version control (e.g., Git, GitHub Actions)
Proficient in SQL, Spark (RDD/DataFrame APIs), and Python or Scala
Experience with cloud platforms (AWS, Azure, or Google Cloud Platform), especially integrating Snowflake and Databricks
Familiarity with data modeling, data quality, and orchestration tools (e.g., Airflow, Prefect)
Knowledge of CI/CD pipelines and version control (e.g., Git, GitHub Actions)
Preferred Qualifications:
Experience with Delta Lake, MLflow, and Data Governance frameworks
Familiarity with real-time data streaming (Kafka, Spark Structured Streaming)
Strong communication and documentation skills
Experience working in Agile/Scrum teams
Experience with Delta Lake, MLflow, and Data Governance frameworks
Familiarity with real-time data streaming (Kafka, Spark Structured Streaming)
Strong communication and documentation skills
Experience working in Agile/Scrum teams
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.