Overview
On Site
Depends on Experience
Contract - W2
Skills
GCP
Python
Data Flow
DataStream
CDC (Change Data Capture)
Cloud Functions
Cloud Run
Pub Sub
BigQuery
Cloud Storage
GitHub
Terraform
SAP SLT
Job Details
Role: Data Engineer
Location Houston, TX(Hybrid)
Job description:
- Develop, construct, test and maintain data acquisition pipelines for large volumes of structed and unstructured data. This includes batch and real-time processing (in Google Cloud)
- Build large and complex datasets based on business requirements
- Construct big data pipeline architecture
- Identify opportunities for data acquisition via working with stakeholders and business clients
- Translate business needs to technical requirements
- Leverage a variety of tools in the Google Cloud Ecosystem such as Python, Data Flow, DataStream, CDC (Change Data Capture), Cloud Functions, Cloud Run, Pub Sub, BigQuery, Cloud Storage to integrate systems and data pipelines
- Use logs & alerts to effectively monitor pipelines
- Use SAP SLT to replicate SAP tables to Google Cloud using SLT
- Develop JSON messaging structure for integrating with various applications
- Leverage DevOps and CI/CD practices (GitHub, Terraform) to ensure the reliability and scalability of data pipelines
- Partition/Cluster and retrieve content in Big Query and use IAM roles & Policy Tags to secure the data
- Use roles to secure access to datasets, authorized views to share data between projects
- Design and build an ingestion pipeline using Rest API
- Recommends ways to improve data quality, reliability, and efficiency
Experience:
- Requires
- 4+ years of experience in a data engineering role Google Cloud Platform Preferred
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.