Snowflake Data Engineer(Fulltime) for Dallas TX (Day 1 Onsite)

Overview

On Site
Full Time

Skills

snowflake
snowpro

Job Details

Fulltime job Opportunity

Tittle : Snowflake Data Engineer

Location: Dallas TX (Day 1 Onsite)

Key Responsibilities

  • Design and implement Snowflake schemas (star, snowflake, data vault) optimized with micro-partitioning, clustering keys, materialized views, and search optimization services.
  • Build real-time and batch ingestion pipelines into Snowflake using Snowpipe, Kafka Connect, Fivetran, Matillion, Informatica, or dbt.
  • Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture).
  • Use Zero-Copy Cloning for environment management, testing, and sandboxing.
  • Apply Time Travel and Fail-safe features for data recovery and auditing.
  • Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake.
  • Design integrations with cloud storage (S3, Azure ADLS, GCS) for staging and external tables.
  • Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace.
  • Enable semi-structured data handling (JSON, Avro, Parquet, ORC, XML) using VARIANT columns and lateral flattening.
  • Integrate Snowflake with BI tools (Power BI, Tableau) via live connections and semantic layers.
  • Implement RBAC (Role-Based Access Control), Row Access Policies, and Dynamic Data Masking for data security.
  • Optimize compute usage with multi-cluster warehouses, resource monitors, and query performance tuning.
  • Manage cost optimization strategies (warehouse auto-suspend, query profiling, storage/compute separation).
  • Integrate with data catalog & governance platforms (Collibra, Alation, Informatica CDGC) using Snowflake metadata and APIs.
  • Work with domain teams to deliver data products leveraging Snowflake's data mesh-friendly features.
  • Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers.
  • Support CI/CD automation for Snowflake code deployment using GitHub Actions, Azure DevOps, or dbt Cloud.

Qualifications

  • Education: Bachelor's or Master's in Computer Science, Data Engineering, or related field.
  • Experience:
    • 10+ years of data engineering experience, with 5+ years in Snowflake Data Cloud.
    • Expertise in SQL optimization and Snowflake performance tuning.
    • Hands-on with Snowpipe, Streams & Tasks, Snowpark, Zero-Copy Cloning, and Secure Data Sharing.
    • Proficiency in Python, Scala, or Java for Snowpark development.
    • Experience integrating with cloud platforms like AWS.
    • Exposure to ETL/ELT tools (Informatica, Matillion, Fivetran).
    • Familiarity with CI/CD, Git, DevOps practices for data operations.
  • Preferred Certifications:
    • SnowPro Core Certification

Key Skills

  • Snowflake-native feature design and implementation (Snowpark, Streams, Time Travel, Secure Data Sharing)
  • Data ingestion (Snowpipe, CDC, Kafka, Fivetran)
  • Semi-structured data handling (VARIANT, JSON, Avro, Parquet)
  • Advanced SQL and performance tuning
  • Data governance (RBAC, masking, lineage, catalogs)
  • Cloud data platform integrations (AWS S3, Azure ADLS, Google Cloud Platform GCS)
  • BI and analytics tool integration
  • Cost optimization and warehouse orchestration

Thanks &Regards

Rahul Sharma | Team Lead
Amaze Systems Inc

E: |

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About Amaze Systems Inc