Overview
Skills
Job Details
Job Title: Data Engineer Location: Wilmington, DE
Job Summary:
We are seeking a highly skilled Data Engineer with strong experience in PySpark/Spark, AWS cloud services, and Python. The ideal candidate will be responsible for designing, building, and optimizing scalable data pipelines and data processing solutions to support analytics, reporting, and business operations.
Key Responsibilities:
Design, develop, and maintain scalable ETL/ELT data pipelines using PySpark/Spark.
Build and optimize large-scale data processing systems on AWS (S3, EMR, Glue, Lambda, Redshift, etc.).
Write efficient, reusable, and robust code in Python for data transformation and processing.
Collaborate with data architects, analysts, and business stakeholders to understand requirements and implement data solutions.
Work on data ingestion from various structured/unstructured data sources.
Ensure data quality, integrity, and reliability through validation, monitoring, and performance tuning.
Implement best practices for data security, governance, and compliance.
Troubleshoot data pipeline issues and provide timely resolutions.
Participate in code reviews and contribute to continuous improvement of data engineering processes.
Required Skills:
Strong hands-on experience with PySpark/Spark for big data processing.
Proficient in AWS services such as S3, Glue, Lambda, EMR, Redshift, IAM, CloudWatch, etc.
Advanced programming skills in Python.
Experience working with relational and NoSQL databases.
Strong understanding of ETL/ELT frameworks, data warehousing concepts, and distributed computing.
Ability to handle large datasets and optimize performance.
Familiarity with CI/CD pipelines and version control (Git).
Excellent problem-solving and communication skills.