AI/ML Lead with Databricks

Hybrid in Los Angeles, CA, US • Posted 3 hours ago • Updated 3 hours ago
Contract Corp To Corp
Contract Independent
Contract W2
Hybrid
$70 - $80/hr
Fitment

Dice Job Match Score™

🛠️ Calibrating flux capacitors...

Job Details

Skills

  • AI/ML
  • Databricks
  • AWS
  • Delta Lake
  • S3
  • Unity Catalog
  • Python
  • PySpark
  • MLflow
  • MLOps
  • SQL

Summary

Job Title: AI/ML Lead with Databricks

Location: Los Angeles CA (Hybrid)

Duration: 6+ Months

Job Description

Role Overview

We are seeking an experienced AI/ML Architect with deep hands-on expertise in Databricks on AWS to lead the design and implementation of scalable, highperformance data and machine learning platforms. The ideal candidate combines architectural thinking with strong engineering execution, demonstrating the ability to build modern lakehouse systems, optimize largescale pipelines, and drive analytical and ML capabilities across the organization.

This role requires working with large, multi-terabyte datasets, advanced analytics, and endtoend ML lifecycle management using Databricks, Python, PySpark, and AWS-native services.

Must Demonstrate (Critical Competencies)

  • Designing Databricksbased lakehouse architectures on AWS (Delta Lake + S3 + Unity Catalog).
  • Clear separation of compute vs. serving layers in distributed architectures.
  • Low-latency API strategy where Spark is insufficient (e.g., leveraging optimized services or caching).
  • Caching strategies to accelerate reads and reduce compute cost.
  • Data partitioning, file size tuning, and optimization strategies for large-scale pipelines.
  • Experience handling multi-terabyte structured timeseries workloads.
  • Ability to distill architectural significance from ambiguous business requirements.
  • Strong curiosity, questioning, and requirementprobing mindset.
  • Playercoach approach: hands-on technical depth + ability to guide design.

Key Responsibilities

AI/ML & Advanced Analytics

  • Develop, train, and optimize ML models using Python, PySpark, MLflow, and Databricks Machine Learning.
  • Conduct exploratory data analysis (EDA) to identify patterns, trends, and insights in large datasets.
  • Deploy ML models into production using MLflow, Databricks Workflows, or other MLOps pipelines.
  • Build analytics solutions such as forecasting, anomaly detection, segmentation, or recommendation systems.
  • Design ML architectures aligned with Databricks Lakehouse on AWS.
  • Data Engineering & Lakehouse Architecture
  • Architect and build scalable ETL/ELT pipelines using PySpark, SQL, and Databricks Workflows.
  • Implement Delta Lake best practices, including OPTIMIZE, ZORDER, partitioning, and schema evolution.
  • Design lakehouse layers (Bronze/Silver/Gold) with strong separation of compute and serving layers.
  • Optimize cluster performance and jobs using Spark tuning, caching, and shuffle minimization.
  • Work with multi-terabyte, time-series, highvelocity data in a distributed environment.
  • Ensure robust data availability for downstream ML and analytics workloads.

AWS Cloud Integration

  • Architect end-to-end data and ML solutions using AWS services, including:
  • S3 for storage
  • IAM for identity & access
  • Glue Catalog for metadata management
  • Networking for secure, highthroughput data movement
  • Integrate Databricks with AWS-native compute, API layers, and low-latency endpoints.
  • Business Collaboration & Leadership
  • Translate business problems into scalable analytical or ML architectures.
  • Communicate complex statistical and architectural concepts to nontechnical stakeholders.
  • Collaborate with product, engineering, and business leaders to drive data-informed initiatives.
  • Provide design leadership while remaining hands-on in execution.

Skills & Qualifications

Required

  • Bachelor s or Master s in Computer Science, Data Science, Engineering, Statistics, or related field.
  • 10+ years of experience in data engineering, ML engineering, or AI/ML architecture roles.
  • Deep expertise in Databricks on AWS, including:
  • PySpark / Spark SQL
  • Databricks Notebooks
  • Delta Lake
  • Unity Catalog
  • MLflow
  • Databricks Jobs & Workflows
  • Strong programming ability in Python (pandas, numpy, scikit-learn).
  • Demonstrated experience with large-scale, multi-terabyte data processing.
  • Strong understanding of ML algorithms, distributed systems, and data optimization.

Preferred

  • Experience with MLOps and production deployment pipelines.
  • Strong grasp of AWS-native data and compute services.
  • Understanding of CI/CD using GitHub Actions, GitLab CI, or similar.
  • Familiarity with deep learning frameworks (TensorFlow, PyTorch).

Key Competencies

  • Strong analytical and problem-solving skills.
  • Ability to work in fast-paced, highly collaborative environments.
  • Excellent communication and presentation abilities.
  • Self-driven with exceptional attention to architectural detail.

With Regards,

Dinesh B

CS Solutions, Inc

7525 Mitchell Road, Suite 106, Eden Prairie, MN 55344

Email ID -

Direct Number

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: CSSOL
  • Position Id: 8903022
  • Posted 3 hours ago
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

Los Angeles, California

Today

Easy Apply

Contract

Depends on Experience

Hybrid in Los Angeles, California

Today

Easy Apply

Full-time

$140,000 - $160,000

Los Angeles, California

6d ago

Easy Apply

Third Party

Los Angeles, California

Today

Easy Apply

Third Party, Contract

$50 - $60

Search all similar jobs