Data Engineer (Snowflake & Apache Spark) | Plano, TX | Long term contract

  • Plano, TX
  • Posted 2 days ago | Updated 2 days ago

Overview

On Site
$50 - $57
Contract - W2
Contract - 18 Month(s)

Skills

Snowflake
Apache Spark
Databricks
SnowPro Core

Job Details

Job Title: Data Engineer Snowflake & Apache Spark (Certified)

Location: Plano TX

Job type: Contract

Job Summary:

We are seeking a skilled and certified Data Engineer with proven hands-on experience in Snowflake and Apache Spark. The ideal candidate holds a SnowPro Core or Advanced Certification and a Databricks Certified Associate Developer for Apache Spark, demonstrating their expertise in modern data platforms and distributed data processing. This role will involve building scalable data pipelines, optimizing data storage, and enabling real-time and batch data processing.

Key Responsibilities:

  • Design, develop, and maintain robust, scalable ETL/ELT pipelines using Apache Spark and Snowflake.
  • Leverage Databricks for data processing, transformation, and analytics in distributed environments.
  • Develop efficient SQL and Spark applications to process and analyze large volumes of data.
  • Implement and maintain data warehousing solutions using Snowflake with best practices for performance, cost, and security.
  • Collaborate with data scientists, analysts, and business stakeholders to meet data needs.
  • Ensure data quality and integrity through unit testing, data validation, and monitoring.
  • Optimize and troubleshoot Spark jobs, SQL queries, and Snowflake data workflows.
  • Integrate with various data sources (cloud storage, APIs, RDBMS) and tools (Airflow, DBT, etc.).
  • Apply data governance and compliance policies in data pipeline design and execution.

Required Qualifications:

  • Certifications:
    • SnowPro Core / Advanced Certification (e.g., SnowPro Advanced: Architect, Data Engineer, etc.)
    • Databricks Certified Associate Developer for Apache Spark (latest version preferred)
  • Experience:
    • 3+ years of experience working with Snowflake, including schema design, query optimization, and Snowpipe/Streams/Tasks.
    • 2+ years of hands-on development with Apache Spark (PySpark, Scala, or Java) in Databricks or open-source environments.
    • Strong understanding of distributed computing, data lakes, and modern data architectures.
  • Technical Skills:
    • Proficient in SQL, Spark (RDD/DataFrame APIs), and Python or Scala
    • Experience with cloud platforms (AWS, Azure, or Google Cloud Platform), especially integrating Snowflake and Databricks
    • Familiarity with data modeling, data quality, and orchestration tools (e.g., Airflow, Prefect)
    • Knowledge of CI/CD pipelines and version control (e.g., Git, GitHub Actions)

Preferred Qualifications:

  • Experience with Delta Lake, MLflow, and Data Governance frameworks
  • Familiarity with real-time data streaming (Kafka, Spark Structured Streaming)
  • Strong communication and documentation skills
  • Experience working in Agile/Scrum teams

Applicant Consent:
By submitting your application, you agree to ApTask's () and , and provide your consent to receive SMS and voice call communications regarding employment opportunities that match your resume and qualifications. You understand that your personal information will be used solely for recruitment purposes and that you can withdraw your consent at any time by contacting us at or . Message frequency may vary. Msg & data rates may apply.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.