Data Architect with Microservices

Overview

Remote
Depends on Experience
Accepts corp to corp applications
Contract - W2
Contract - Independent

Skills

GCP

Job Details

Job Title: Data Architect with Microservices
Location: Remote
Duration: Long Term Contract
 
Responsibilities:
 · Architect, design and develop core data platform components with a microservices architecture, abstracting platform, and infrastructure intricacies.
· Create and maintain essential data platform SDKs and libraries, adhering to industry best practices.
· Design and develop connector frameworks and modern connectors to source data from disparate application both on-prem and cloud.
· Design and optimize data storage, processing, and querying performance for large-scale datasets using industry best practices.
· Design and develop data quality frameworks and processes to ensure the accuracy and reliability of data.
· Collaborate with data scientists, analysts, and cross functional teams to design data models, database schemas and data storage solutions.
· Design and develop advanced analytics and machine learning capabilities on the data platform.
· Design and develop observability and data governance frameworks and practices.
· Stay up to date with the latest data engineering trends, technologies, and best practices.
· Drive the deployment and release cycles, ensuring a robust and scalable platform.
 
 
Requirements:
 · 10+ (for senior) 15+ (for principal) of proven experience in modern cloud data engineering, data architectures, data warehousing, and software engineering.
· Expertise in architecting, designing, and building end to end data platforms in the Google Cloud Platform environment using BigQuery and other services while adhering to best practices guidelines such as open standards, cost, performance, time to market and minimize vendor lock
· Solid experience building data platforms in Google Cloud Platform environment.
· Solid experience designing and developing modular, distributed data platform components with a microservices architecture. Strong experience with Docker, Kubernetes, APIs is needed.
· Proficiency in data engineering tools and technologies - SQL, Python, Spark, DBT, Airflow, Kafka.
· Solid experience implementing data lineage, data quality and data observability for big data workflows.
· Strong experience with modern data modeling, data architecture, and data governance principles.
· Excellent experience with DataOps principles and test automation.
· Excellent experience with observability tools - Grafana and Datadog
· Applicants must be authorized to work for any employer in the US without visa requirements. This position is not eligible for an employment visa sponsorship.
 
 
Nice to have:
· Experience with Data Mesh architecture.
· Experience building Semantic layers for data platforms.
· Experience building scalable IoT architectures
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About TechMiners LLC