Data Infrastructure Engineer (NEED & LOCAL)

Overview

Remote
Depends on Experience
Contract - Independent
Contract - W2
Contract - 12 Month(s)

Skills

Amazon DynamoDB
Amazon Web Services
Apache Kafka
Apache Spark
Business Intelligence
Business Operations
Cloud Computing
Dashboard
Data Analysis
Data Governance
Data Integrity
Data Masking
Databricks
Extract
Transform
Load
Finance
Golang
Kubernetes
Legal
Management
Orchestration
PostgreSQL
Python
Real-time
Recruiting
Regulatory Compliance
SQL
Snow Flake Schema
Startups
Streaming
Terraform
TypeScript
Warehouse
Writing

Job Details

Job Title:- Data Infrastructure Engineer

Location:- NYC (Remote)

Duration:- 3-month C2H

Job Description

Responsibilities:

  • Architect, build, and maintain modern and robust real-time and batch data analytics pipelines.
  • Develop and maintain declarative data models and transformations.
  • Implement data ingestion integrations for streaming and traditional sources such as Postgres, Kafka, and DynamoDB.
  • Deploy and configure BI tooling for data analysis.
  • Work closely with product, finance, legal, and compliance teams to build dashboards and reports to support business operations, regulatory obligations, and customer needs.
  • Establish, communicate, and enforce data governance policies.
  • Document and share best practices with regards to schema management, data integrity, availability, and security.
  • Protect and limit access to sensitive data by implementing a secure permissioning model and establishing data masking and tokenization processes.
  • Identify and communicate data platform needs, including additional tooling and staffing.
  • Work with cross-functional teams to define requirements, plan projects, and execute on the plan.

Qualifications

  • 5+ years of professional engineering and data analytics experience, startup experience a plus.
  • Strong proficiency and comfort using SQL and Python to perform complex data analysis.
  • Recent experience building automation tooling and pipelines using a general purpose language such as Python, Golang, and/or Typescript.
  • Experience with modern data pipeline and warehouse technologies (e.g. Snowflake, Databricks, Apache Spark, AWS Glue)
  • Strong proficient writing declarative data models and transformations using modern technologies (e.g. dbt)
  • Experience building and maintaining cloud-based data lakes.
  • Prior experience with integrating real-time data streaming technologies (e.g. Kafka, Spark)
  • Prior experience with configuring and maintaining modern data orchestration platforms (e.g. Airflow)
  • Comfort with infrastructure-as-code tooling (e.g. Terraform) and container orchestration platforms (Kubernetes)
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.