Data Engineer

  • Chicago, IL
  • Posted 3 days ago | Updated 16 hours ago

Overview

On Site
Hybrid
Accepts corp to corp applications
Contract - W2
Contract - Independent
Contract - 6+ month(s)

Skills

Python
GCP
CI/CD
GKE
SQL
Kubernetes
pyspark
Kafka
Spark SQL
Airflow
API
Rest
webhooks
PostgreSQL
YAML
Grafana

Job Details

Job Summary (List Format):

- 100% remote position
- Experience in healthcare or logistics domains is highly desired
- Design, build, and maintain data pipelines using Python and PySpark
- Develop and manage workflows with Airflow or similar orchestration tools
- Work extensively with Google Cloud Platform services, including BigQuery, GCS, Pub/Sub, Cloud Run, Functions, and Cloud SQL
- Implement real-time data ingestion using Kafka, webhooks, and file-based methods
- Integrate APIs via REST/webhooks
- Deploy and manage Kubernetes environments, preferably GKE
- Write and optimize queries in BigQuery SQL and PostgreSQL
- Design YAML/config-driven data pipelines
- Perform schema transformation, hashing, and data quality framework tasks
- Contribute to CI/CD pipelines, observability, and develop lightweight dashboards using Grafana, Streamlit, or Flask UI
- Bonus experience in PostgreSQL, CI/CD, monitoring, dashboarding, or lightweight UI development
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About DCode Talent