Overview
Skills
Job Details
Location: CA
Job Summary:We are seeking a skilled and proactive Azure Databricks Developer with strong proficiency in Python to join our data engineering team. The ideal candidate will be responsible for designing, developing, and optimizing large-scale data pipelines and analytics solutions on the Azure Databricks platform.
Key Responsibilities:-
Design, develop, and maintain data pipelines and ETL/ELT processes using Azure Databricks and Python.
-
Implement data ingestion from various sources (structured, semi-structured, and unstructured).
-
Write clean, scalable, and optimized Spark code using PySpark within Databricks.
-
Collaborate with data architects, data analysts, and business stakeholders to understand data requirements.
-
Ensure data quality, integrity, and compliance with governance and security standards.
-
Use Azure services like Data Lake, Data Factory, Synapse, Blob Storage, and Key Vault in daily workflows.
-
Monitor, troubleshoot, and improve the performance of existing pipelines.
-
Participate in code reviews, testing, and deployment using CI/CD tools and practices.
-
Bachelor's or Master's degree in Computer Science, Engineering, or related field.
-
8+ years of hands-on experience with Azure Databricks and Python.
-
Strong knowledge of Apache Spark, PySpark, and distributed data processing.
-
Experience working with Azure Data Services (ADF, ADLS, Synapse, etc.).
-
Proficient in developing production-level ETL workflows.
-
Good understanding of data modeling, SQL, and data warehouse concepts.
-
Experience with version control (Git) and CI/CD pipelines.
-
Familiarity with Agile/Scrum methodologies.
-
Azure certification (e.g., DP-203: Data Engineering on Microsoft Azure).
-
Experience with Delta Lake, MLflow, or Databricks SQL.
-
Exposure to DevOps practices and monitoring tools like Datadog or Azure Monitor.
-
Experience working in cloud-native data lakehouse architectures.