Overview
Skills
Job Details
Lead Databricks Engineer Role Summary
Architecture & Engineering Leadership:
Leads the end-to-end architecture, design, and implementation of big data pipelines on Databricks Lakehouse Platform, integrating with Azure, AWS, or Google Cloud Platform.Data Pipeline & ETL Expertise:
Develops scalable, high-performance ETL/ELT pipelines using PySpark, Delta Lake, and SQL, enabling real-time and batch processing of large datasets.Team & Project Oversight:
Mentors junior engineers, drives Agile delivery, performs code reviews, and ensures best practices in data engineering and CI/CD (using Git, Azure DevOps, or Jenkins).Cross-Platform Integration:
Integrates Databricks with external systems such as Kafka, Snowflake, Power BI, Synapse, and cloud storage (S3, ADLS), ensuring data quality, security, and governance.Performance, Security & Governance:
Optimizes jobs and clusters for cost and performance. Implements Unity Catalog, role-based access controls, and compliance policies for secure enterprise data usage.
Must-Have Skills:
Databricks (PySpark, Delta Lake, MLflow, Unity Catalog)
Cloud (Azure Databricks, AWS/Google Cloud Platform optional)
Python, SQL, Spark
CI/CD, Git, DevOps, Terraform (optional)
Data Governance, Performance Tuning