We are seeking a Lead Snowflake Data Engineer to design, own, and deliver end-to-end data engineering solutions in modern cloud environments. This role focuses on building scalable, high-performance data pipelines using Snowflake and Cortex AI, with full lifecycle ownership from ingestion and transformation to modeling, optimization, and consumption.
Key Responsibilities
Lead the design and development of end-to-end ELT pipelines using Snowflake
Architect scalable data models optimized for performance, cost, and analytics consumption
Build and maintain backend data services using Python and PySpark
Leverage Snowflake Cortex AI to enable advanced analytics and intelligent data products
Drive performance tuning across pipelines, including query optimization, clustering, and warehouse scaling
Enforce best practices in data governance, security, and compliance
Collaborate across business, analytics, and engineering teams to deliver high-quality solutions
Provide technical leadership and mentorship to engineering teams
Communicate architecture decisions and trade-offs effectively in client-facing environments
Required Qualifications & Technical Expertise
10+ years of experience, or equivalent ownership of production-grade data platforms
Deep expertise in:
Snowflake (data modeling, performance tuning, optimization)
Python and PySpark
Advanced SQL
Proven ability to design and deliver end-to-end data pipelines (ingestion transformation modeling consumption) in cloud environments (AWS preferred)
Required: Ownership of at least one production-grade Snowflake pipeline end-to-end
Strong foundation in modern data warehousing:
Dimensional modeling (star/snowflake schemas)
ELT/ETL design patterns
Data marts and optimization strategies
Experience with distributed data processing and large-scale datasets
Hands on experience with Snowflake Cortex AI integration
Working knowledge of React.js or similar frameworks
Strong understanding of data governance, security, and compliance
Ability to:
Clearly explain and defend architectural decisions
Design systems that perform reliably at scale
Balance performance, cost, and maintainability
Technical Depth (Must Be Demonstrated)
Candidates should be able to clearly explain and apply the following in real-world scenarios:
Snowflake Performance & Scaling
Warehouse scaling modes (auto-scale, multi-cluster) and when to use them
Clustering keys and performance trade-offs
Cost vs performance optimization strategies
Snowflake Storage & Optimization
Micro-partitioning and its impact on pruning and query performance
Practical optimization techniques for large datasets
End-to-End Pipeline Design
Designing a complete ELT pipeline using Snowflake
Deciding where transformations should occur (Snowflake vs external processing)
Ensuring scalability, maintainability, and performance across the pipeline Engagement.