Here is OUR CLIENT REQUIREMENT which can be filled immediately.
Position: Data Engineer (Python, PySpark, Scala, SQL, Google Cloud Platform)
Location: NJ/ NC
Duration: Full-Time
Years of Exp: 10+ Years
Role Summary:
Looking for a Data Engineer with expertise in Python, PySpark, Scala, SQL, and Google Cloud Platform (Google Cloud Platform) to design and maintain scalable data pipelines and workflows.
Key Responsibilities:
Develop and maintain ETL processes using PySpark on Google Cloud Platform.
Optimize and troubleshoot data processing jobs for performance and reliability.
Implement data transformations and create pipelines to support analytics.
Collaborate with data scientists and analysts to deliver business insights.
Monitor and maintain cloud infrastructure related to data processing on Google Cloud Platform.
Document technical solutions and provide support for data-related issues.
Required Skills:
Hands-on experience with Google Cloud Platform services (BigQuery, Dataproc, Dataflow, Cloud Storage, Pub/Sub).
Proficiency in PySpark, Scala, and SQL for large-scale data processing.
Experience with workflow orchestration tools like Apache Airflow or Cloud Composer.
Familiarity with Databricks and CI/CD practices.
Strong problem-solving and analytical skills.
Process Skills:
Strong communication and collaboration in Agile teams.
Ability to lead technical discussions and mentor junior developers.