Overview
Skills
Job Details
ob Title: Data Engineer (Databricks + Azure)
Client: One of our Consulting Clients (Global Analytics & Digital Transformation Firm)
Location: Columbus, OH (Remote/Hybrid)
Duration: Full-Time
About the Role
We are seeking a highly skilled Data Engineer with deep expertise in Databricks and Azure Cloud to join a decision analytics and data engineering team within one of our global consulting clients. The role involves building, optimizing, and maintaining large-scale data pipelines that fuel enterprise analytics, reporting, and AI-driven insights-primarily supporting clients in the insurance and financial services domains.
Key Responsibilities
Data Pipeline Development & Optimization
- Design, build, and enhance ETL/ELT data pipelines using Azure Data Factory, Databricks (PySpark, SQL, Python), and related services.
- Develop and manage Delta Live Tables, Autoloader, and Unity Catalog within the Databricks ecosystem for structured, incremental data processing.
- Implement data ingestion, transformation, and validation frameworks that ensure high performance, scalability, and reliability.
- Monitor data pipelines, troubleshoot issues, and ensure optimal system performance and SLA adherence.
Data Modeling & Architecture
- Collaborate with business analysts and reporting teams to define logical and physical data models supporting analytical and operational needs.
- Implement data warehousing and lakehouse solutions using Azure Data Lake and Delta Lake.
- Optimize data structures for query performance, cost efficiency, and reusability.
Data Quality, Governance & Automation
- Design and implement robust data quality checks and validation mechanisms to maintain integrity across sources and transformations.
- Automate repetitive processes using scripts, parameterized pipelines, and reusable frameworks.
- Conduct periodic audits and compliance checks aligned with governance policies.
- Contribute to metadata management, documentation, and lineage tracking.
Required Skills & Experience
- 7 12 years of experience in Data Engineering with proven expertise in Databricks and Azure Cloud ecosystems.
- Strong hands-on experience in PySpark, Python, and SQL for data transformation, validation, and performance tuning.
- Solid understanding of Delta Lake architecture, ETL/ELT frameworks, and data warehousing principles.
- Proficiency with Azure services including Data Factory (ADF), Data Lake (ADLS), and Databricks Notebooks.
- Experience with Delta Live Tables, Unity Catalog, and Autoloader for batch and streaming data processing.
- Strong background in data modeling, performance optimization, and automation scripting.
- Familiarity with Agile methodologies and DevOps-based deployment practices (Git, CI/CD preferred).
- Strong analytical, communication, and problem-solving skills to collaborate effectively across diverse teams.
- Preferred: Exposure to insurance, healthcare, or financial services data ecosystems.
Nice to Have
- Experience in data migration projects (on-prem to cloud or multi-cloud).
- Familiarity with Delta Sharing, Databricks SQL Warehouses, or MLflow for advanced use cases.
- Experience with data cataloging, lineage, or quality frameworks such as Purview, Collibra, or Great Expectations.
- Exposure to BI/reporting tools like Power BI or Tableau for end-to-end integration understanding.