Overview
On Site
Depends on Experience
Accepts corp to corp applications
Contract - W2
Skills
RDBMS
PostgreSQL
SQL
JSON
DataStage
AWS Glue
PySpark
Python
Job Details
Key Responsibilities:
- Design, develop, and maintain scalable data pipelines and ETL processes using AWS Glue, DataStage, PySpark, and Python.
- Work with RDBMS (e.g., PostgreSQL, SQL Server) to extract, transform, and load data.
- Handle diverse data formats (e.g., JSON, Parquet) to support analytics and data science initiatives.
- Leverage AWS services such as S3, Glue Catalog, and related ecosystem tools to build robust, high-performance data platforms.
- Optimize and monitor workflows for performance, cost-efficiency, and reliability.
- Collaborate with data analysts, architects, and business stakeholders to understand requirements and deliver solutions.
- Ensure data quality, security, and compliance with company and industry standards.
Required Skills and Qualifications:
- Proven experience with RDBMS (PostgreSQL, SQL Server) and complex SQL queries.
- Strong knowledge of data formats (JSON, Parquet) and their efficient usage in big data pipelines.
- Hands-on expertise in AWS ecosystem for data engineering (S3, Glue, DataStage).
- Proficiency in PySpark and Python for data processing and transformation.
- Experience building and maintaining robust data platforms for analytics and BI.
- Familiarity with data pipeline orchestration, performance tuning, and cost optimization.
- Strong problem-solving skills and the ability to work collaboratively in a fast-paced environment.
Thanks
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.