Enterprise Data Lake (EDL) Technical Lead

Hybrid in Dallas, TX, US • Posted 9 hours ago • Updated 9 hours ago
Contract Corp To Corp
Contract W2
Contract Independent
75% Travel Required
Able to Sponsor
Hybrid
Depends on Experience
Fitment

Dice Job Match Score™

🧠 Analyzing your skills...

Job Details

Skills

  • EDL
  • Kafka

Summary

Enterprise Data Lake Technical Lead 
Any Visa works 
Relocation works 
C2C works 
Hybrid 3 days onsite in Dallas Texas 
Direct Client Role 
12+ Months Contract 

The Enterprise Data Lake (EDL) Technical Lead is responsible for owning the design, implementation, and engineering leadership of the enterprise data lake platform. This role ensures the platform is scalable, reliable, secure, and consumable across analytics, reporting, and operational workloads. 

The Tech Lead drives technical decisions and collaborates closely with Architecture, Product Management, Data Engineering, Security, and downstream consumers to deliver robust data platform capabilities.



Enterprise Data Lake Technical Lead 

Primary Skills :

Apache Kafka
Databricks platform
DataHub or Collibra/Alation/Apache Atlas
RDBMS and Modern Data lake
Kubernetes - Functional knowledge

Secondary : 
Cloud Platform (Azure and/ or AWS)
C# and or Go or java
Terraform
ELT kind of services

Position Overview 

The Enterprise Data Lake (EDL) Technical Leadis responsible for owning the design, implementation, and engineering leadership of the enterprise data lake platform. This role ensures the platform is scalable, reliable, secure, and consumable across analytics, reporting, and operational workloads. 

The Tech Lead drives technical decisions and collaborates closely with Architecture, Product Management, Data Engineering, Security, and downstream consumers to deliver robust data platform capabilities. 

 

Key Responsibilities 

Technical Leadership 

  • Own the design and implementation of enterprise-scale data lake solutions on Azure/AWS Cloud 

  • Define technical standards and best practices for data platform components 

  • Drive technical decision-making across data ingestion, storage, processing, and governance layers 

  • Provide technical mentorship and guidance to data engineering and platform teams 

  • Collaborate with Architecture team on overall architectural alignment and strategy 

  • Partner with Product Management, Data Engineering, Security, SRE, and downstream consumers 

Data Governance & Catalog Management (Critical) 

  • Own the design and implementation of enterprise data catalog and metadata management layer 

  • Build and operate data catalog using DataHub or equivalent metadata platform 

  • Design and implement automated metadata ingestion, lineage tracking, and data discovery capabilities 

  • Design and implement data governance policies, data quality rules, and compliance frameworks 

  • Enable self-service data discovery and access management for downstream consumers 

  • Establish data ownership, stewardship models, and metadata standards across the enterprise 

  • Build APIs and services for metadata-driven data operations and workflows 

Platform Engineering & Development 

  • Design and implement real-time and batch data ingestion pipelines using Apache Kafka 

  • Optimize data processing workflows on Databricks platform (Delta Lake, Spark optimization, Unity Catalog) 

  • Design integration patterns between RDBMS sources and data lake (CDC, batch ETL, replication) 

  • Optimize database performance, query tuning, and indexing strategies across relational and distributed systems 

  • Implement infrastructure as code using Terraform for automated provisioning and management 

  • Design and deploy containerized data services on Kubernetes clusters 

  • Develop data platform services and tooling using C# and Go 

  • Build database migration tools, data validation frameworks, and connectivity services 

Operations & Reliability 

  • Ensure platform scalability, reliability, and security across all data lake components 

  • Implement monitoring, logging, and observability solutions for data infrastructure 

  • Optimize Linux-based systems for data processing workloads 

  • Establish CI/CD pipelines for data platform deployments 

  • Ensure platform security, compliance, and data governance standards (GDPR, SOC2, etc.) 

  • Drive cost optimization and performance tuning initiatives 

  • Implement database backup/recovery strategies and disaster recovery planning 

 

Required Qualifications 

Technical Expertise 

Data Streaming & Processing 

  • 5+ years with Apache Kafka (streaming architecture, Kafka Connect, Schema Registry, stream processing) 

  • 3+ years with Databricks (Delta Lake, Apache Spark optimization, Unity Catalog, cluster management) 

  • Deep understanding of data lake architecture patterns (Bronze/Silver/Gold, medallion architecture) 

Data Governance & Cataloging 

  • 3+ years with DataHub or similar metadata management platforms (Alation, Collibra, Apache Atlas) 

  • Deep experience building and operating enterprise data catalog systems 

  • Expertise in automated metadata extraction, lineage tracking, and impact analysis 

  • Experience with data quality frameworks and metadata-driven data operations 

  • Knowledge of data governance policies, data classification, and compliance automation 

  • Understanding of data discovery, access control, and self-service analytics enablement 

Relational Database Systems 

  • 5+ years with enterprise RDBMS platforms including: 

  • SQL Server (T-SQL, SSIS, SSRS, replication, Always On Availability Groups) 

  • PostgreSQL (advanced query optimization, partitioning, extensions, streaming replication) 

  • MySQL (replication, clustering, performance tuning) 

  • Strong SQL skills (complex queries, stored procedures, window functions, query optimization) 

  • Database design principles (normalization, indexing strategies, schema design, partitioning) 

  • Change Data Capture (CDC) patterns and implementation (Debezium, Azure Data Factory, AWS DMS, custom solutions) 

  • Database migration experience (schema migration, data migration, zero-downtime migrations) 

Cloud & Infrastructure 

  • 5+ years with Azure or AWS Cloud including: 

  • Azure: Data Lake Storage (Gen2), Event Hubs, AKS, Azure SQL, Key Vault, Azure AD, Monitor 

  • AWS: S3, MSK/Kinesis, EKS, RDS/Aurora, Secrets Manager, IAM, CloudWatch 

  • Cloud-native data services, networking, security, and IAM 

  • 3+ years with Kubernetes (deployment strategies, scaling, monitoring, service mesh, Helm) 

  • 3+ years with Terraform (modules, state management, multi-environment deployments, multi-cloud) 

Programming & Development 

  • Strong proficiency in any programming languages like C#, Go 

  • Expert-level SQL across multiple database platforms 

  • Experience with Python for data engineering tasks (preferred) 

  • Familiarity with Shell scripting (Bash) 

Leadership & Experience 

  • 7+ years in data engineering, platform engineering, or database engineering roles 

  • 3+ years in technical leadership capacity (Tech Lead, Principal Engineer) 

  • Proven track record of delivering large-scale data infrastructure projects 

  • Experience leading teams of 5-10+ engineers 

  • Strong architectural design and system thinking capabilities 

  • Experience migrating legacy RDBMS workloads to modern data lake architectures 

  • Demonstrated ability to balance technical excellence with business needs 

 

Preferred Qualifications 

Additional Technologies 

  • Experience with cloud ETL services (Azure Data Factory, AWS Glue, Azure Stream Analytics, AWS Lambda) 

  • Knowledge of managed database features (Azure SQL elastic pools/hyperscale, AWS RDS/Aurora serverless) 

  • Knowledge of additional streaming technologies (Apache Flink) 

  • Experience with database sharding and horizontal partitioning strategies at scale 

  • Familiarity with NoSQL databases (Cosmos DB, DynamoDB, MongoDB, Cassandra, Redis) 

  • Experience with Apache Iceberg 

  • Knowledge of data observability tools 

    Education: Any Bachelors or Similar

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 10428868
  • Position Id: 8949491
  • Posted 9 hours ago
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

Hybrid in Dallas, Texas

Yesterday

Easy Apply

Contract, Third Party

Depends on Experience

Dallas, Texas

Yesterday

Easy Apply

Contract

Depends on Experience

Irving, Texas

Today

Easy Apply

Contract

$50 - $55

Irving, Texas

Today

Easy Apply

Contract

$53.56 - $60.35

Search all similar jobs