Overview
Remote
Depends on Experience
Contract - W2
Contract - Independent
Contract - 1 Year(s)
Skills
Azure Data Factory
Azure Databricks
Scala
ADF
Delta table
Job Details
Job Title: Azure Data Engineer - Databricks with Scala
- Job Location: India : Remote Role
- Experience Level: Mid to Senior
- Must-Have Skills:
- Azure Databricks with Scala (MUST HAVE)
- Experience with Azure Data Factory (ADF) dataflow and ADF pipelines (MUST HAVE)
- Practical experience in handling error or invalid formats in incoming data.
- Experience working with Delta table formats (Must)
Job Description:
We are looking for a skilled Data Engineer with expertise in Databricks and Scala to join our team. The ideal candidate will have hands-on experience in building robust data pipelines, working with Azure Data Factory, and managing data in Delta table formats. This role offers an exciting opportunity to work on cutting-edge data projects and contribute to the success of our analytics initiatives.
Responsibilities:
- Design, develop, and maintain data pipelines using Databricks with Scala, ensuring scalability, reliability, and efficiency.
- Implement and optimize Azure Data Factory (ADF) dataflows and pipelines to facilitate seamless data movement and transformation.
- Handle error or invalid formats in incoming data streams, implementing robust error handling mechanisms to ensure data integrity.
- Work with Delta table formats to manage and process large volumes of structured and unstructured data efficiently.
- Collaborate with cross-functional teams to understand data requirements and implement solutions that meet business needs.
- Conduct performance tuning and optimization of data pipelines to improve throughput and reduce latency.
- Document technical designs, specifications, and procedures to ensure knowledge sharing and maintain system integrity.
- Stay updated with emerging technologies and industry trends related to data engineering and analytics.
Requirements:
- Bachelor's degree in Computer Science, Engineering, or a related field.
- 5+ years of experience in data engineering with a focus on Databricks and Scala.
- Strong proficiency in Apache Spark and Scala programming language.
- Hands-on experience with Azure Data Factory (ADF) dataflow and pipeline development.
- Experience working with Delta Lake and managing Delta table formats.
- Solid understanding of data processing concepts, data modeling, and ETL best practices.
- Ability to troubleshoot and debug complex data pipeline issues effectively.
- Excellent communication and collaboration skills, with the ability to work independently and in a team environment
- Nice to Have:
- Experience with other cloud platforms such as AWS or Google Cloud Platform.
- Knowledge of Python programming language.
- Familiarity with containerization technologies such as Docker and Kubernetes.
- Certification in Azure Data Engineering or related field.