Databricks Engineer - Commercial Insurance - Fulltime

Overview

Remote
On Site
Full Time

Skills

Azure
Insurance
databricks
Data lineage

Job Details

ob Title: Data Engineer (Databricks + Azure)

Client: One of our Consulting Clients (Global Analytics & Digital Transformation Firm)

Location: Columbus, OH (Remote/Hybrid)

Duration: Full-Time

About the Role

We are seeking a highly skilled Data Engineer with deep expertise in Databricks and Azure Cloud to join a decision analytics and data engineering team within one of our global consulting clients. The role involves building, optimizing, and maintaining large-scale data pipelines that fuel enterprise analytics, reporting, and AI-driven insights-primarily supporting clients in the insurance and financial services domains.

Key Responsibilities

Data Pipeline Development & Optimization

  • Design, build, and enhance ETL/ELT data pipelines using Azure Data Factory, Databricks (PySpark, SQL, Python), and related services.
  • Develop and manage Delta Live Tables, Autoloader, and Unity Catalog within the Databricks ecosystem for structured, incremental data processing.
  • Implement data ingestion, transformation, and validation frameworks that ensure high performance, scalability, and reliability.
  • Monitor data pipelines, troubleshoot issues, and ensure optimal system performance and SLA adherence.

Data Modeling & Architecture

  • Collaborate with business analysts and reporting teams to define logical and physical data models supporting analytical and operational needs.
  • Implement data warehousing and lakehouse solutions using Azure Data Lake and Delta Lake.
  • Optimize data structures for query performance, cost efficiency, and reusability.

Data Quality, Governance & Automation

  • Design and implement robust data quality checks and validation mechanisms to maintain integrity across sources and transformations.
  • Automate repetitive processes using scripts, parameterized pipelines, and reusable frameworks.
  • Conduct periodic audits and compliance checks aligned with governance policies.
  • Contribute to metadata management, documentation, and lineage tracking.

Required Skills & Experience

  • 7 12 years of experience in Data Engineering with proven expertise in Databricks and Azure Cloud ecosystems.
  • Strong hands-on experience in PySpark, Python, and SQL for data transformation, validation, and performance tuning.
  • Solid understanding of Delta Lake architecture, ETL/ELT frameworks, and data warehousing principles.
  • Proficiency with Azure services including Data Factory (ADF), Data Lake (ADLS), and Databricks Notebooks.
  • Experience with Delta Live Tables, Unity Catalog, and Autoloader for batch and streaming data processing.
  • Strong background in data modeling, performance optimization, and automation scripting.
  • Familiarity with Agile methodologies and DevOps-based deployment practices (Git, CI/CD preferred).
  • Strong analytical, communication, and problem-solving skills to collaborate effectively across diverse teams.
  • Preferred: Exposure to insurance, healthcare, or financial services data ecosystems.

Nice to Have

  • Experience in data migration projects (on-prem to cloud or multi-cloud).
  • Familiarity with Delta Sharing, Databricks SQL Warehouses, or MLflow for advanced use cases.
  • Experience with data cataloging, lineage, or quality frameworks such as Purview, Collibra, or Great Expectations.
  • Exposure to BI/reporting tools like Power BI or Tableau for end-to-end integration understanding.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About Acunor Infotech