Overview
Hybrid1 2 / Days per Week
Depends on Experience
Full Time
Skills
SAP ECC
S/4HANA
SQL
Python
and PySpark
ETL/ELT
Job Details
Data Engineer
Location Newark, NJ / Edison, NJ (Hybrid 1 2 / Days per Week)
Responsibilities:
- Extract and transform data from SAP systems into cloud-based storage solutions.
- Manage and structure data within SQL-based Data Lakes for efficient querying and access.
- Design and maintain ETL/ELT pipelines to ingest data into Amazon Redshift for analytics and reporting.
- Develop and schedule AWS Glue jobs to automate data processing workflows using PySpark or Spark.
- Optimize Redshift performance through efficient schema design, indexing, and data modeling.
- Ensure data quality, consistency, and integrity across the entire data pipeline.
- Collaborate with cross-functional teams to gather data requirements and support business intelligence needs.
- Monitor, troubleshoot, and document data pipelines, ensuring reliability and scalability.
Required:
- Minimum of 8 years of hands-on experience in data engineering or related roles.
- Experience extracting and integrating data from SAP systems (e.g., SAP ECC, S/4HANA).
- Proficiency in SQL, Python, and PySpark for data processing and transformation.
- Hands-on experience with AWS services: Glue (ETL jobs), Redshift, S3, and CloudWatch.
- Strong understanding of data lake architecture and file formats like Parquet, ORC, JSON.
- Knowledge of data warehousing, ETL/ELT pipelines, and performance optimization in Redshift.
- Excellent problem-solving, communication, and collaboration skills.
Preferred Qualifications:
- Experience of working in the Utility Industry
- Experience integrating SAP IS-U (Industry Solution for Utilities) data.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.