Role: Snowflake Data Engineer
Location: Hybrid (Dallas TX)
Key Responsibilities
· Design and implement Snowflake schemas (star, snowflake, data vault) optimized with micro-partitioning, clustering keys, materialized views, and search optimization services.
· Build real-time and batch ingestion pipelines into Snowflake using Snowpipe, Kafka Connect, Fivetran, Matillion, Informatica, or dbt.
· Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture).
· Use Zero-Copy Cloning for environment management, testing, and sandboxing.
· Apply Time Travel and Fail-safe features for data recovery and auditing.
· Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake.
· Design integrations with cloud storage (S3, Azure ADLS, GCS) for staging and external tables.
· Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace.
· Enable semi-structured data handling (JSON, Avro, Parquet, ORC, XML) using VARIANT columns and lateral flattening.
· Integrate Snowflake with BI tools (Power BI, Tableau) via live connections and semantic layers.
· Implement RBAC (Role-Based Access Control), Row Access Policies, and Dynamic Data Masking for data security.
· Optimize compute usage with multi-cluster warehouses, resource monitors, and query performance tuning.
· Manage cost optimization strategies (warehouse auto-suspend, query profiling, storage/compute separation).
· Integrate with data catalog & governance platforms (Collibra, Alation, Informatica CDGC) using Snowflake metadata and APIs.
· Work with domain teams to deliver data products leveraging Snowflake’s data mesh-friendly features.
· Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers.
· Support CI/CD automation for Snowflake code deployment using GitHub Actions, Azure DevOps, or dbt Cloud.
Qualifications
Education: Bachelor’s or Master’s in Computer Science, Data Engineering, or related field.
Experience:
· 10+ years of data engineering experience, with 5+ years in Snowflake Data Cloud.
· Expertise in SQL optimization and Snowflake performance tuning.
· Hands-on with Snowpipe, Streams & Tasks, Snowpark, Zero-Copy Cloning, and Secure Data Sharing.
· Proficiency in Python, Scala, or Java for Snowpark development.
· Experience integrating with cloud platforms like AWS.
· Exposure to ETL/ELT tools (Informatica, Matillion, Fivetran).
· Familiarity with CI/CD, Git, DevOps practices for data operations.
· Preferred Certifications:
· SnowPro Core Certification
Key Skills
· Snowflake-native feature design and implementation (Snowpark, Streams, Time Travel, Secure Data Sharing)
· Data ingestion (Snowpipe, CDC, Kafka, Fivetran)
· Semi-structured data handling (VARIANT, JSON, Avro, Parquet)
· Advanced SQL and performance tuning
· Data governance (RBAC, masking, lineage, catalogs)
· Cloud data platform integrations (AWS S3, Azure ADLS, Google Cloud Platform GCS)
· BI and analytics tool integration
· Cost optimization and warehouse orchestration