Data Engineer - Onsite- Dallas TX or Charlotte, NC. Must have strong experience in PySpark, Spark, Python and Hadoop.

  • Dallas, TX
  • Posted 2 days ago | Updated 2 days ago

Overview

On Site
Depends on Experience
Accepts corp to corp applications
Contract - W2
Contract - Independent

Skills

Apache Hadoop
Apache Spark
Application Development
Collaboration
Conflict Resolution
Data Analysis
Data Processing
Data Quality
Decision-making
FOCUS
Geospatial Analysis
Knowledge Transfer
Management
Mentorship
Problem Solving
PySpark
Python
Storage
Testing
User Experience
Workflow

Job Details

Job Title: Data Engineer with PySpark, Spark, Python, Hadoop experience Experience: 12 years
Job Summary: With a focus on Apache Hadoop, Python, and Spark, the candidate will develop and implement data-driven strategies. This hybrid role offers the opportunity to work with cutting-edge technologies in a dynamic environment, contributing to impactful projects that enhance the company's geospatial capabilities.
Required Skills: Apache Hadoop, Hadoop, Python, Spark Pyspark, PySpark
Responsibilities: - Develop and implement geospatial data solutions using Apache Hadoop and Spark to drive business insights. - Collaborate with cross-functional teams to integrate geospatial data into existing systems and workflows. - Analyze complex datasets using Python and PySpark to identify trends and patterns that inform decision-making. - Design and optimize data pipelines for efficient processing and storage of geospatial information. - Provide technical expertise in the deployment and maintenance of geospatial data platforms. - Ensure data quality and integrity through rigorous testing and validation processes. - Lead the development of innovative geospatial applications that enhance user experience and functionality. - Oversee the integration of geospatial data with other data sources to create comprehensive datasets. - Mentor junior engineers in best practices for geospatial data analysis and application development. - Stay updated with the latest advancements in geospatial technologies and incorporate them into projects. - Collaborate with stakeholders to understand business needs and translate them into technical requirements. - Document processes and methodologies to ensure knowledge transfer and continuity. - Contribute to the company's strategic goals by delivering high-quality geospatial solutions that impact society positively.
Qualifications: - Possess a strong background in Apache Hadoop and Spark for geospatial data processing. - Demonstrate proficiency in Python and PySpark for data analysis and application development. - Have experience in designing and optimizing data pipelines for large-scale data environments. - Show expertise in integrating geospatial data with other data sources for comprehensive analysis. - Exhibit strong problem-solving skills and the ability to work collaboratively in a hybrid work model.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About Keylent