Job Title - Google Cloud Platform Engineer
Location - Hartford, CT
Duration - 3 Months followed by C2H
Position Description:
Design and implement robust data pipelines using Google Cloud Platform (Google Cloud Platform) services such as BigQuery, Cloud Storage, and Pub/Sub.
Develop and manage workflows using Cloud Composer (Apache Airflow) for efficient scheduling and orchestration.
Write clean, efficient, scalable code in Python, leveraging advanced programming techniques.
Craft complex SQL queries in BigQuery, including window functions, CTEs, and performance tuning strategies.
Build and maintain real-time data processing systems using Apache Kafka.
Model and manage NoSQL databases, particularly MongoDB, with a focus on scalable schema design.
Utilize Shell scripting and perform Linux system administration tasks to support data infrastructure.
Conduct data profiling and implement validation techniques to ensure data quality and integrity.
Develop and maintain API integration scripts for seamless service automation and data exchange.
Troubleshoot and resolve data-related issues with strong analytical and problem-solving skills.
Create and maintain data flow diagrams to clearly communicate architecture and pipeline logic to stakeholders.
Required Skills:
3+ years of proven hands-on experience with Google Cloud Platform data services.
Experience in Agile methodology with preferred experience in SAFe methodology
Experience working in onshore/offshore support model collaborating work with offshore teams.
Strong understanding of distributed systems and data engineering best practices.
Excellent communication and documentation skills.
Ability to work independently and collaboratively in a fast-paced environment.
Bonus Skills:
Google Cloud Platform Google Cloud Professional Data Engineer Certification
Experience working with GitHub, RTC, automation tools
Experience AI programming, IVR Technologies like AYAYA, CISCO, Chatbot etc