Overview
Hybrid1 to 2 days a week onsite
$100,000 - $120,000
Full Time
25% Travel
Skills
AWS
Medallion
Data Engineering
Job Details
Job Title: Sr. AWS Data Engineer
Location: Newark, NJ
Work Mode: Hybrid 1 to 2 days a week onsite
Duration: Fulltime with ComTec
Eligibility Requirements: Candidates must have legal authorization to work in the United States without the need for current or future employer sponsorship.
We are seeking a skilled AWS Data Engineer who has experience working with Python, SQL, Glue, RedShift, Lambda, Airflow, Medallion Architecture, and Step Functions.
Responsibilities:
- Design, build, and optimize ETLs using Python, SQL, Lambda, Glue, RedShift, Airflow, Step Functions, and other AWS services.
- Develop and maintain ETL workflows using Python, SQL, and AWS-native tools.
- Create SQL queries to segment, manipulate, and formatdata.
- Provide Run/DevOps support for data services, ensuring high availability and performance.
- Collaborate with data scientists, business analysts, SAP functional SMEs and other stakeholders to resolve data-related issues and improve system reliability.
- Implement batch job scheduling and manage data dependencies using tools like Airflow or Step Functions.
- Maintain documentation for data flows, incident resolution, and operational procedures.
- Participate in on-call rotations and incident response for production systems.
- Build automations to ingest, transfer, move, upload, and manipulatedata.
- Build or maintain data ingestion pipelines that move data from source systems into Snowflake.
- Create and manage data models to ensure data integrity and facilitate efficient data analysis.
- Implement and maintain data security and compliance measures, including access controls, encryption, and data masking.
- Ensure data quality, accuracy, and consistency through data validation, cleansing, and monitoring.
Requirements:
- Bachelor s degree in computer science, Engineering, or related field.
- 3 5 years of experience in data engineering or production support roles.
- Strong proficiency in AWS services: S3, Glue, Lambda, Redshift, DMS, CloudWatch.
- Good understanding of Medallion architecture
- Solid programming skills in Python and SQL.
- Experience with DevOps tools: GitLab, Jenkins, Bitbucket, Maven.
- Familiarity with data processing frameworks: Spark, Hive, Kafka.
- Must have experience with Airflow and Step Functions.
- Advanced SQL query development proficiency
- Understanding of data modelling principles and techniques.
- Knowledge of data security best practices and compliance requirements.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.