Overview
Remote
On Site
Full Time
Skills
Video
Extract
Transform
Load
Microsoft Windows
Value Engineering
Analytical Skill
Data Processing
Storage
Amazon S3
GCS
Orchestration
Management
Testing
Data Warehouse
Snow Flake Schema
Migration
ELT
Real-time
Apache Kafka
Streaming
Use Cases
Soft Skills
Debugging
Python
SQL
Apache Spark
Apache Flink
Amazon Web Services
Electronic Health Record (EHR)
Cloud Computing
Google Cloud
Google Cloud Platform
Workflow
Git
Continuous Integration
Continuous Delivery
Java
Apache Hive
Oracle
Communication
Job Details
Position: Data Engineer
Location: CA - 100% Remote
Interview Mode - Video
Duration: 12+ months
Core Technical Skills
1. Python (Required)
4. Cloud Platforms (Google Cloud Platform or AWS)
8. Java (Nice to Have)
Platform & Architecture
9. Data Warehouse Experience
Soft Skills & Remote Readiness
12. Independent Work
Pre-Qualification Checklist
Skill / Tool
Required
Confirmed (Yes/No)
Notes
Python
Yes
SQL
Yes
Spark / Flink / Beam
Yes
Google Cloud Platform or AWS
Yes
BigQuery / EMR
Yes
Dataproc
Yes
Cloud Run
Yes
Airflow / Prefect / Dagster
Yes
Google Cloud Workflows
Yes
Git / CI/CD
Yes
dbt / Dataform
Nice to Have
Java
Nice to Have
Hive
Nice to Have
Oracle
Nice to Have
Communication & Autonomy
Yes
#TB_EN
Location: CA - 100% Remote
Interview Mode - Video
Duration: 12+ months
Core Technical Skills
1. Python (Required)
- Years of experience:
- Describe a production-grade data pipeline you've developed using Python.
- Comfort level with complex queries (joins, windows, CTEs):
- Share an example of a query you've written for analytical purposes.
- Tools used (select all that apply): Spark, Flink, Beam
- Which one(s) have you used in production, and for what purpose?
- Describe a project where one of these tools was key for performance, scale, or streaming.
4. Cloud Platforms (Google Cloud Platform or AWS)
- Platforms used:
- Experience with:
- AWS EMR
- Google Cloud Platform BigQuery
- Google Dataproc
- Google Cloud Run
- Cloud-native storage (e.g., S3, GCS)
- Relevant cloud APIs
- Tools used: (Airflow, Prefect, Dagster, Google Cloud Workflows, etc.)
- How have you managed retries, alerts, dependencies, or branching logic?
- Experience level:
- Approach to managing models, testing, and deployment environments
- Git strategy familiarity:
- Tools used for automating pipeline deployment or testing
8. Java (Nice to Have)
- Years of experience:
- What types of data-related or backend projects have you used Java for?
Platform & Architecture
9. Data Warehouse Experience
- Have you worked with:
- BigQuery
- Snowflake
- Dremio
- Hive (Nice to Have)
- Oracle (Nice to Have)
- Describe a migration or integration project involving one of these systems.
- What's your approach to designing scalable, testable ELT workflows?
- Any experience optimizing for performance or cost?
- Experience with Kafka, Spark Streaming, Flink, or Beam:
- Use case and your role in implementation
Soft Skills & Remote Readiness
12. Independent Work
- This role is remote and requires autonomy. Share an example of successfully delivering a solution independently.
- Describe a time you had to debug a complex data issue. What tools or methods did you use?
Pre-Qualification Checklist
Skill / Tool
Required
Confirmed (Yes/No)
Notes
Python
Yes
SQL
Yes
Spark / Flink / Beam
Yes
Google Cloud Platform or AWS
Yes
BigQuery / EMR
Yes
Dataproc
Yes
Cloud Run
Yes
Airflow / Prefect / Dagster
Yes
Google Cloud Workflows
Yes
Git / CI/CD
Yes
dbt / Dataform
Nice to Have
Java
Nice to Have
Hive
Nice to Have
Oracle
Nice to Have
Communication & Autonomy
Yes
#TB_EN
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.