Overview
Skills
Job Details
Job Role: ETL developer with AWS Cloud
Location: Remote
Experience: 12+ Years
ETL Tools:
Talend
Database:
Snowflake, Oracle, Amazon RDS (Aurora, Postgres), DB2, SQL server and Casandra,
Big Data and Amazon Services:
Apache Sqoop, AWS S3, Hue, AWS CLI, Amazon EMR, Amazon MSK, Amazon Sagemaker, Apache Spark
Data Modeling Tools:
ArchiMate (not mandated- secondary/preferred), Erwin, Oracle Data Modeler (secondary/preferred)
Scheduling Tools: Autosys, SFTP, AirFlow (preferred. This should not be an issue, any resource can learn how to use it)
Key Responsibilities:
Designing, building, and automating ETL processes using AWS services like Apache Sqoop, AWS S3, Hue, AWS CLI, Amazon EMR, Amazon MSK, Amazon Sagemaker Apache Spark.
Developing and maintaining data pipelines to move and transform data from diverse sources into data warehouses or data lakes.
Ensuring data quality and integrity through validation, cleansing, and monitoring ETL processes.
Optimizing ETL workflows for performance, scalability, and cost efficiency within the AWS environment.
Troubleshooting and resolving issues related to data processing and ETL workflows.
Implementing and maintaining security measures and compliance standards for data pipelines and infrastructure.
Documenting ETL processes, data mappings, and system architecture.
Implementing security measures such as IAM roles and access controls.
Diagnosing and resolving issues related to AWS services, infrastructure, and applications.
Proficiency in Big data tool and AWS services: Including Apache Sqoop, AWS S3, Hue, AWS CLI, Amazon EMR, Amazon MSK, Amazon Sagemaker, Apache Spark relevant to data storage and processing.
Strong SQL skills:
For querying databases and manipulating data during the transformation process.