Databricks Architect

Hybrid in Los Angeles, CA, US • Posted 30+ days ago • Updated 3 days ago
Contract W2
Contract Independent
No Travel Required
Hybrid
$60 - $80/hr
Company Branding Image
Fitment

Dice Job Match Score™

🤯 Applying directly to the forehead...

Job Details

Skills

  • Databricks Architect
  • Databricks
  • Databricks Lakehouse
  • Data Architect
  • Data Engineering Architect
  • Data Platform Architect
  • Lead Data Engineer
  • Principal Data Engineer
  • AWS
  • AWS Data Engineer
  • PySpark
  • Spark
  • Spark SQL
  • Delta Lake
  • Medallion Architecture
  • Bronze Silver Gold Architecture
  • Unity Catalog
  • Databricks Workflows
  • Databricks Jobs
  • Delta Live Pipelines
  • Structured Streaming
  • ETL
  • ELT
  • Data Pipelines
  • Big Data
  • Distributed Data Systems
  • S3
  • AWS Glue
  • Glue Catalog
  • IAM
  • Lambda
  • Kinesis
  • API Gateway
  • Data Lake
  • Lakehouse Architecture
  • Data Governance
  • Data Optimization
  • ZORDER
  • Data Partitioning
  • CI/CD
  • Git
  • Terraform
  • DevOps
  • MLflow
  • MLOps

Summary

Job Title: Databricks Architect

Location : Los Angeles CA  (Hybrid)

Hire type : FTE / CTH

Role Overview

The Databricks Architect is responsible for designing, implementing, and optimizing scalable data engineering and analytics solutions on the Databricks Lakehouse Platform on AWS. This role requires deep expertise in distributed data processing, Delta Lake–based architectures, and modern data engineering best practices. The architect will partner with crossfunctional teams to define data strategies, ensure platform reliability, and enable advanced analytics, ML, and BI workloads across the enterprise.


Must Demonstrate (Critical Architectural Capabilities)

  • Designing Databricks-based Lakehouse architectures on AWS
  • Clear separation of compute layer vs. serving layer
  • Low-latency API/data delivery strategy (cannot rely solely on Spark)
  • Caching strategies for performance acceleration and cost efficiency
  • Data partitioning and optimization strategy, including file-size tuning
  • Ability to handle multi-terabyte structured time-series datasets
  • Skill in distilling architectural significance from complex requirements
  • Strong curiosity and requirementprobing mindset
  • Playercoach leadership style (hands-on engineering + design guidance)

Key Responsibilities

Architecture & Design

  • Architect end-to-end Databricks Lakehouse solutions on AWS for ingestion, processing, storage, and consumption.
  • Define and implement Delta Lake patterns including Medallion Architecture (Bronze/Silver/Gold).
  • Lead design of scalable data pipelines using PySpark, Spark SQL, Workflows, and Delta Live Pipelines.
  • Architect solutions for structured, semi-structured, and time-series workloads.
  • Ensure architectures support low-latency delivery, servinglayer separation, and high performance.

Engineering & Implementation

  • Build robust ETL/ELT pipelines using Databricks Notebooks, Jobs, and Workflows.
  • Implement streaming and incremental data processing as needed using Structured Streaming.
  • Optimize Spark jobs with partitioning, caching, ZORDER, file compaction, and shuffle reduction.
  • Implement CI/CD automation using Databricks Repos, GitLab/GitHub, and Infrastructure-as-Code.

AWS Cloud & Platform Expertise

  • Architect Databricks solutions using AWS-native services including:
    • S3 (data storage)
    • Glue Catalog (metadata governance)
    • IAM (identity & access control)
    • Lambda / API Gateway (low-latency serving mechanisms)
    • Kinesis (streaming ingestion)
  • Ensure security, governance, and compliance via Unity Catalog, RBAC, and encryption standards.
  • Monitor workloads and optimize cluster sizing, autoscaling, and cost controls.

Collaboration & Leadership

  • Partner with data engineers, ML engineers, BI teams, and business stakeholders.
  • Serve as a Databricks SME, defining best practices, standards, and architectural patterns.
  • Conduct architectural reviews and guide teams on solution choices.
  • Lead PoCs, evaluate new Databricks features, and drive platform adoption across teams.

Quality, Governance & Observability

  • Define standards for data quality, lineage, observability, and operational governance.
  • Implement automated testing frameworks for pipelines and notebooks.
  • Establish monitoring dashboards, performance baselines, and reliability KPIs.

Required Skills & Experience

Technical Skills

  • 7+ years in data engineering or data architecture.
  • 3+ years hands-on with Databricks.
  • Strong expertise in Spark, PySpark, SQL, and distributed data systems.
  • Deep understanding of Delta Lake features (ACID, OPTIMIZE, ZORDER, Auto Loader).
  • Experience with workflows/orchestration and Databricks REST APIs.
  • Hands-on expertise with AWS, specifically:
    • S3
    • Glue / Glue Catalog
    • IAM
    • Lambda
    • Kinesis
  • Familiarity with CI/CD, Git, DevOps, and IaC (Terraform preferred).

Soft Skills

  • Strong analytical and problemsolving abilities.
  • Excellent communication and stakeholder management.
  • Ability to lead design discussions and guide engineering teams.
  • Strong documentation and architectural blueprinting abilities.

Preferred Qualifications

  • Databricks Certifications such as:
    • Databricks Certified Data Engineer Professional
    • Databricks Certified Machine Learning Professional
    • Databricks Lakehouse Fundamentals
  • Experience with MLflow, Feature Store, or MLOps pipelines.
  • Experience in regulated industries (BFSI, healthcare, etc.).

 

 

 

Thanks & Stay Safe!

 

-

Saurabh Dhiman

Ph: EXT 114 

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 91170785
  • Position Id: DD008
  • Posted 30+ days ago

Company Info

About Digital Dhara LLC

Wouldn’t it be amazing to have access to a global network of the world’s smartest people—people with specialized IT and engineering skills? Partnering with us instantly gives you the ability to bring experts or multi-disciplined teams into your business. 

You can rely on us to manage entire services on your behalf or deliver your products and services end-to-end at our global Delivery Centers and Centers of Excellence.

We believe that having the speed, agility, expertise and scale to outpace technology actually comes down to one very simple solution – the right people, working together, in the right ways. Let us help you connect smarter.

About_Company_OneAbout_Company_Two
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

Los Angeles, California

3d ago

Easy Apply

Full-time

130,000 - 150,000

Phoenix, Arizona

7d ago

Easy Apply

Contract, Third Party

Depends on Experience

Search all similar jobs