Database Developer / ETL with Python

Overview

On Site
Depends on Experience
Full Time

Skills

SQL
Python
PL/SQL
Pandas
Transact-SQL
Stored Procedures
Jenkins
SQL. ETL
Database Systems
Data Modeling
CI/CD

Job Details

Key Responsibilities:

  • Design, develop, and optimize T-SQL and PL/SQL queries, stored procedures, functions, and triggers for efficient, high-performance data processing across relational databases.

  • Perform data extraction, transformation, and loading (ETL/ELT) using a combination of tools and languages such as Informatica, Python (pandas, polars, pyarrow, SQLAlchemy), and SQL-based scripts.

  • Build and automate end-to-end ETL pipelines to ingest, cleanse, transform, and load data from heterogeneous sources including flat files, APIs, RDBMS, and cloud storage.

  • Architect and execute data migration strategies, including on-prem to cloud migrations leveraging services like Snowflake, Azure SQL, Google Cloud Platform BigQuery, or AWS RDS.

  • Optimize and maintain large-scale database systems such as Teradata, Oracle, PostgreSQL, and SQL Server.

  • Utilize Python scripting for custom transformations, data quality checks, batch automation, and integration with third-party APIs or services.

  • Contribute to and maintain overall data architecture and modeling standards, ensuring scalability, integrity, and performance tuning across data layers.

  • Build and maintain CI/CD pipelines using tools like Jenkins, Git, GitLab CI/CD, and Bitbucket Pipelines for automated testing and deployment of database objects and ETL workflows.

  • Collaborate with stakeholders including data architects, analysts, DevOps, and QA teams to ensure delivery of reliable and accurate data solutions.


Preferred Qualifications:

  • Proven experience with Teradata, Oracle, and advanced PL/SQL programming.

  • Deep expertise in Python for ETL development, with strong command of pandas, polars, numpy, pyarrow, and SQLAlchemy.

  • Hands-on experience with Informatica PowerCenter/Cloud or similar ETL tools.

  • Working knowledge of Snowflake, including tasks like staging, transformation, and performance tuning (nice to have).

  • Experience managing version-controlled ETL projects using Git, and deploying changes using CI/CD pipelines (Jenkins, GitHub Actions, or GitLab).

  • Familiarity with cloud data platforms (Azure, Google Cloud Platform, AWS) and working with tools like Airflow, DBT, or Dataform is a plus.

  • Strong understanding of data governance, data quality, and metadata management best practices.

  • Ability to handle multiple concurrent priorities, deliver independently, and estimate effort reliably in Agile/Scrum environments.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About HStechnologies LLC