Data Engineer_ W2

Overview

On Site
Up to $75
Contract - W2
Contract - 12 Month(s)

Skills

Amazon Redshift
Amazon S3
Amazon Web Services
Analytics
Apache Spark
Big Data
Business Operations
Cloud Computing
Collaboration
Communication
Conflict Resolution
Continuous Delivery
Continuous Improvement
Continuous Integration
Data Engineering
Data Processing
Data Quality
Data Security
Data Warehouse
Database
Distributed Computing
ELT
Electronic Health Record (EHR)
Extract
Transform
Load
Git
NoSQL
Performance Tuning
Problem Solving
PySpark
Python
Regulatory Compliance
Reporting
Unstructured Data
Version Control

Job Details

Job Title: Data Engineer Location: Wilmington, DE
Job Summary:

We are seeking a highly skilled Data Engineer with strong experience in PySpark/Spark, AWS cloud services, and Python. The ideal candidate will be responsible for designing, building, and optimizing scalable data pipelines and data processing solutions to support analytics, reporting, and business operations.

Key Responsibilities:

Design, develop, and maintain scalable ETL/ELT data pipelines using PySpark/Spark.

Build and optimize large-scale data processing systems on AWS (S3, EMR, Glue, Lambda, Redshift, etc.).

Write efficient, reusable, and robust code in Python for data transformation and processing.

Collaborate with data architects, analysts, and business stakeholders to understand requirements and implement data solutions.

Work on data ingestion from various structured/unstructured data sources.

Ensure data quality, integrity, and reliability through validation, monitoring, and performance tuning.

Implement best practices for data security, governance, and compliance.

Troubleshoot data pipeline issues and provide timely resolutions.

Participate in code reviews and contribute to continuous improvement of data engineering processes.

Required Skills:

Strong hands-on experience with PySpark/Spark for big data processing.

Proficient in AWS services such as S3, Glue, Lambda, EMR, Redshift, IAM, CloudWatch, etc.

Advanced programming skills in Python.

Experience working with relational and NoSQL databases.

Strong understanding of ETL/ELT frameworks, data warehousing concepts, and distributed computing.

Ability to handle large datasets and optimize performance.

Familiarity with CI/CD pipelines and version control (Git).

Excellent problem-solving and communication skills.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About Learn Beyond Consulting LLC