LEAD Data Engineer / Lead Data Platform Engineer

  • Dallas, TX
  • Posted 14 hours ago | Updated 14 hours ago

Overview

Remote
Hybrid
Depends on Experience
Contract - W2

Skills

development/support
administration
Python
Databricks
Apache Spark
Delta Lake
AWS CloudOps
Cloud Security

Job Details

Lead Data Platform Engineer
Location: Miramar, Dallas or Remote


Sr. Data Platform Engineer who thrives in a hybrid role 60% administration and 40% development/support to help us scale our data and DataOps infrastructure. You ll work with cutting-edge technologies like Databricks, Apache Spark, Delta Lake, and AWS CloudOps, Cloud Security, while supporting mission-critical data pipelines and integrations. If you re a hands-on engineer with strong Python skills, deep AWS experience, and a knack for solving complex data challenges, we want to hear from you.

Key Responsibilities Design, develop, and maintain scalable ETL pipelines and integration frameworks. Administer and optimize Databricks and Apache Spark environments for data engineering workloads. Build and manage data workflows using AWS services such as Lambda, Glue, Redshift, SageMaker, and S3. Support and troubleshoot DataOps pipelines, ensuring reliability and performance across environments. Automate platform operations using Python, PySpark, and infrastructure-as-code tools. Collaborate with cross-functional teams to support data ingestion, transformation, and deployment. Provide technical leadership and mentorship to junior developers and third-party teams. Create and maintain technical documentation and training materials. Troubleshoot recurring issues and implement long-term resolutions.

Minimum Qualifications Bachelor s or Master s degree in Computer Science or related field. 5+ years of experience in data engineering or platform administration. 3+ years of experience in integration framework development with a strong emphasis on Databricks, AWS, and ETL.

Required Technical Skills Strong programming skills in Python and PySpark. Expertise in Databricks, Apache Spark, and Delta Lake. Proficiency in AWS CloudOps, Cloud Security, including configuration, deployment, and mon

Meta driven, real-time streaming pipelines development, Distributed data processing, Data Orchestration, Unstructured/Semi/Structured data processing, Flexible data modeling and Semantic Engineering, Knowledge graphs, Data-as-a-service, Data Observability, Quality frameworks, Data Syndication, Data Fabric development, Data Market place development, Cognitive search engine development, Master data management, Data Governance, Data Migration; Experience in specific tech stacks include: Kafka, Databricks, DeltaLake, Python, Pandas, Spark, pySpark, AirFlow, Neo4j, GraphDB, MongoDB, PostgreSQL, OWL, Python functions, New Relic, Grafana, OpenLineage, Apache atlas, Databricks Unity catalog, DLT, Great Expectations, Databricks Delta-sharing

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About Akshaya Inc