Azure DataBricks with Pyspark

Overview

Remote
$60 - $70
Accepts corp to corp applications
Contract - 12 month(s)
No Travel Required

Skills

ORACLE
bi
oic
paas

Job Details

Title: Azure DataBricks with Pyspark and Python
Location: Seattle, WA (Remote)
Duration: 12+ Months

Job Description:

  • Design, build, and deploy data extraction, transformation, and loading processes and pipelines from various sources including databases, APIs, and data files.
  • Develop and support data pipelines within a Cloud Data Platform, such as Databricks
  • Build data models that reflect domain expertise, meet current business needs, and will remain flexible as strategy evolves
  • Monitor and optimize Databricks cluster performance, ensuring cost-effective scaling and resource utilization
  • Demonstrates ability to communicate technical concepts to non-technical audiences both in written and verbal form
  • Demonstrates strong understanding with coding and programming concepts to build data pipelines (e.g. data transformation, data quality, data integration, etc.)
  • Demonstrates strong understanding of database storage concepts (data lake, relational databases, NoSQL, Graph, data warehousing)
  • Implement and maintain Delta Lake for optimized data storage, ensuring data reliability, performance, and versioning
  • Automate CI/CD pipelines for data workflows using Azure DevOps
  • Collaborate with cross-functional teams to support data governance using Databricks Unity Catalog
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About Purview Infotech