Databricks Data Engineer :: San Fran, CA(Hybrid, Need only locals with DL Copy)

Overview

On Site
Full Time
Accepts corp to corp applications
Contract - W2
Contract - Independent
Contract - longterm

Skills

Analytics
Data Processing
Extract
Transform
Load
ELT
Collaboration
Management
Data Quality
Terraform
Command-line Interface
Regulatory Compliance
Privacy
Computer Science
Data Engineering
Apache Spark
Python
Scala
SQL
Data Lake
Cloud Computing
Microsoft Azure
Amazon Web Services
Google Cloud
Google Cloud Platform
Amazon S3
Continuous Integration
Continuous Delivery
Version Control
Git
Orchestration
Database
Workflow
Data Warehouse
Performance Tuning
Big Data
Data Governance
Unity
Business Intelligence
Microsoft Power BI
Tableau
Databricks
Machine Learning (ML)
LinkedIn
Oracle UCM

Job Details

Title :- Databricks Data Engineer

Location :- San Fran, CA(Hybrid, Need only locals with DL Copy)

Type :- Contract role

Job Summary:-
We are seeking a skilled Data Engineer with hands-on Databricks experience to design, build, and optimize large-scale data pipelines and analytics solutions. You will work with cross-functional teams to enable scalable data processing using the Databricks Lakehouse Platform on Azure.

Key Responsibilities:

  • Design and implement ETL/ELT pipelines using Databricks, Delta Lake, and Apache Spark
  • Collaborate with data scientists, analysts, and stakeholders to deliver clean, reliable, and well-modeled data
  • Build and manage data workflows with Databricks Jobs, Notebooks, and Workflows
  • Optimize Spark jobs for performance, reliability, and cost-efficiency
  • Maintain and monitor data pipelines, ensuring availability and data quality
  • Implement CI/CD practices for Databricks notebooks and infrastructure-as-code (e.g., Terraform, Databricks CLI)
  • Document data pipelines, datasets, and operational processes
  • Ensure compliance with data governance, privacy, and security policies


Qualifications:

  • Bachelor's or Master's in Computer Science, Data Engineering, or a related field
  • 5+ years of experience in data engineering or a similar role
  • Strong hands-on experience with Databricks and Apache Spark (Python, Scala, or SQL)
  • Proficiency with Delta Lake, Unity Catalog, and data lake architectures
  • Experience with cloud platforms (Azure, AWS, or Google Cloud Platform), especially data services (e.g., S3, ADLS, BigQuery)
  • Familiarity with CI/CD pipelines, version control (Git), and job orchestration tools (Airflow, DB Workflows)
  • Strong understanding of data warehousing concepts, performance tuning, and big data processing


Preferred Skills:

  • Experience with MLflow, Feature Store, or other machine learning tools in Databricks
  • Knowledge of data governance tools like Unity Catalog or Purview
  • Experience integrating BI tools (Power BI, Tableau) with Databricks
  • Databricks certification(s) (Data Engineer Associate/Professional, Machine Learning, etc.)

Thanks & Regards

Akshit Sisonia - Sr. Technical Recruiter

Email -

LinkedIn -

STELLENT IT A Nationally Recognized Minority Certified Enterprise

"Happiness can be found, even in the darkest of times, if one only remembers to turn on the light."
- JK Rowling

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.