Technical Skills:
· PySpark – 3 to 5 years
· Google Cloud Platform Dataflow – 3 to 5 years
· Big Data Engineering – 3 to 5 years
Role Description:
As a Google Cloud Platform Data Pipeline Developer & BigQuery Data Engineer, you will be responsible for designing, developing, and maintaining data pipelines on Google Cloud Platform. Your role will involve working with BigQuery to analyze and optimize large datasets, as well as implementing ETL processes to ensure data quality and integrity. Proficiency in foundational Google Cloud Platform services, BigQuery, data engineering principles, SQL, and PySpark will be essential for success in this position. Additionally, experience with ETL tools such as Qlik Data Integration or FiveTran would be a valuable asset in this role.
In this role, you will have the opportunity to work on complex data projects that require a deep understanding of cloud-based data processing and analysis. Your expertise in Google Cloud Platform and BigQuery will be crucial in developing efficient data pipelines that meet the organization''s data processing needs. Strong SQL skills will be necessary for querying and manipulating data within BigQuery, while experience with PySpark will enable you to perform advanced data transformations and analytics. Overall, your ability to leverage these tools and technologies effectively will be key to driving data-driven decision-making within the organization.
As a Google Cloud Platform Data Pipeline Developer & BigQuery Data Engineer, you will play a critical role in shaping the organization''s data infrastructure and analytics capabilities. Your experience in data engineering and cloud-based data processing will be instrumental in driving the success of data projects and initiatives. By utilizing your skills in Google Cloud Platform, BigQuery, SQL, and PySpark, you will be able to design and implement scalable data solutions that support the organization''s data-driven goals. Additionally, your familiarity with ETL tools will enable you to streamline data integration processes and enhance overall data quality.
Responsibilities:
· Develop and maintain Google Cloud Platform data pipelines
· Design and optimize BigQuery data structures
· Utilize SQL and PySpark for data processing
· Collaborate with team members on data engineering projects
· Implement ETL processes using tools like Google Cloud Platform Dataflow, Qlik Data Integration, or FiveTran