Data Engineer with Python

Overview

On Site
$Negotiable
Accepts corp to corp applications
Contract - W2
Contract - long term

Skills

HTML
WEB SERVICES
sql server reporting services
relational databases
data warehousing
Oracle Applications
APACHE SPARK
Big Data
IBM DB2
GOVERNANCE
Microsoft SQL Server
Airflow
Data Ingestion
BigQuery
Data Pipelines
PySpark
DataBricks
COMPUTER PROGRAMMING
OOP Languages
SQL databases
Application Programming Interfaces (APIs)
Python (Programming Language)
Extract Transform Load (ETL)
Self Motivation
Front End Software Development
JavaScript (Programming Language)
SQL Server Integration Services
SQL Stored Procedures
Curiosity
Directed Acyclic Graph (Directed Graphs)
Oracle SQL Developer
Data Systems

Job Details

Job Title: Google Cloud Platform Data Engineer with Python
Location: New York, NY (Hybrid 3 days/week onsite; local candidates only; diversity candidates strongly encouraged)

---

### About the Role:

We are seeking a talented Data Engineer with Python expertise to join our NYC-based team. The ideal candidate will be an individual contributor with a strong background in programming, relational databases, and ETL pipeline development. You will work on building robust data solutions, integrating disparate systems, and ensuring high-quality data for analytics and reporting. Local candidates able to commute to downtown Manhattan are required.

---

### Key Responsibilities:

- Develop, enhance, and maintain ETL pipelines in Python from scratch.
- Analyze and develop complex SQL queries and procedures across multiple RDBMS (SQL Server, DB2, Oracle).
- Design and implement data integrations via APIs and web services.
- Model data: create schemas, define relationships, and design efficient table structures.
- Work with big data and cloud data warehouses (BigQuery, Databricks Delta Lakehouse).
- Implement data ingestion, cleansing, governance, and reporting processes.
- Develop and schedule Airflow DAGs for workflow orchestration.
- Collaborate cross-functionally with business and technical teams.
- Independently understand and enhance existing codebases.

---

### Required Skills and Experience:

- Local to NYC/tri-state area; able to work in-person 3 days/week.
- Strong programming skills in Python; knowledge of OOP principles.
- SQL expertise: can create, optimize, and review queries and stored procedures in SQL Server, Oracle PL/SQL, and DB2.
- Experience with ETL/data pipeline development in large-scale environments.
- Hands-on with data warehouse technologies (BigQuery, Databricks Delta Lakehouse, etc.).
- Familiar with Airflow for pipeline orchestration.
- Experience with RDBMS (MS SQL, Oracle), including SSIS and SSRS.
- Knowledge of Apptio is a plus.
- Exposure to PySpark and Spark is a plus.
- Basic front-end skills: JavaScript, HTML (for data or tool integrations).

---

### Nice to Have:

- Eagerness to learn new technologies and tools.
- Experience with additional cloud platforms or big data technologies.

---

### Diversity & Inclusion:

We strongly encourage applications from women, minorities, and other underrepresented groups in tech.

---

To Apply:
Please submit your resume and a brief cover letter outlining your relevant experience and confirming your ability to commute to downtown Manhattan three days per week.

---

Let me know if you'd like a more concise version or a tailored email template for outreach!
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.