Data Engineering Lead - Fulltime

New York, NY, US • Posted 3 hours ago • Updated 3 hours ago
Full Time
On-site
Depends on Experience
Fitment

Dice Job Match Score™

👾 Reticulating splines...

Job Details

Skills

  • Data Engineer Lead
  • AWS Redshift
  • Spark
  • Python
  • Collibra
  • Snowflake/Databricks
  • Tableau.
  • Data bricks
  • MCP tools
  • LangChain
  • LlamaIndex
  • ETL/ELT

Summary

Job Title: Data Engineering Lead

Skill: AWS redshift

Minimum Experience: 10 - 15 Years

Location: New York, NY

Job Description

Must Have Technical/Functional Skills

AWS Data Engineering Services (EMR/Glue, Redshift, Aurora, S3, Lambda), Spark, Python, Collibra, Snowflake/Databricks, Tableau.

Roles & Responsibilities

  • Ingest and model data from APIs, files/SFTP, and relational sources; implement layered architectures (raw/clean/serving) using PySpark/SQL and dbt, Python.
  • Design and operate pipelines with Prefect (or Airflow), including scheduling, retries, parameterization, SLAs, and well documented runbooks.
  • Build on cloud data platforms, leveraging S3/ADLSS for storage and a Spark platform (e.g., Databricks or equivalent) for compute; manage jobs, secrets, and access.
  • Publish governed data services and manage their lifecycle with Azure API Management (APIM) authentication/authorization, policies, versioning, quotas, and monitoring.
  • Enforce data quality and governance through data contracts, validations/tests, lineage, observability, and proactive alerting.
  • Optimize performance and cost via partitioning, clustering, query tuning, job sizing, and workload management.
  • Uphold security and compliance (e.g., PII handling, encryption, masking) in line with firm standards.
  • Collaborate with stakeholders (analytics, AI engineering, and business teams) to translate requirements into reliable, production ready datasets.
  • Enable AI/LLM use cases by packaging datasets and metadata for downstream consumption, integrating via Model Context Protocol (MCP) where appropriate.
  • Continuously improve platform reliability and developer productivity by automating routine tasks, reducing technical debt, and maintaining clear documentation.
  • 4 15 years of professional data engineering experience.
  • Strong Python, SQL, and Spark (PySpark) skills, and/or Kafka. Snowflake (Snowpipe, Tasks, Streams) as a complementary warehouse.
  • Databricks (Delta formats, workflows, cataloging) or equivalent Spark platforms.
  • Minimum 1 yr of experience in Data bricks (Hands-on).
  • Integrating datasets into MCP tools/providers for LLM/agent applications; familiarity with frameworks such as LangChain or LlamaIndex.
  • Hands-on experience building ETL/ELT with Prefect (or Airflow), dbt, Spark, and/or Kafka.
  • Experience onboarding datasets to cloud data platforms (storage, compute, security, governance).
  • Familiarity with Azure/AWS/Google Cloud Platform data services (e.g., S3/ADLSS; Redshift/BigQuery; Glue/ADF).
  • Git-based workflows CI/CD and containerization with Docker (Kubernetes a plus).
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 10120177
  • Position Id: 8916719
  • Posted 3 hours ago
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

New York, New York

Yesterday

Easy Apply

Full-time

Depends on Experience

New York, New York

Today

Full-time

USD 140,000.00 - 170,000.00 per year

Hybrid in New York, New York

Today

Easy Apply

Full-time

$80,000 - $100,000

New York, New York

Today

Full-time

USD 171,000.00 - 260,000.00 per year

Search all similar jobs