Google Cloud Data Architect- Highly Preferred OCP exp


OMG Technologies
Dice Job Match Score™
🔗 Matching skills to job...
Job Details
Skills
- Apache Hadoop
- Access Control
- Advanced Analytics
- Analytics
- Apache Avro
- Apache Beam
- Apache Hive
- Apache Parquet
- Apache Pig
- Apache Spark
- Apache Sqoop
- Cloud Storage
- Clustering
- Communication
- Continuous Delivery
- Continuous Integration
- Build Automation
- Business Intelligence
- Change Data Capture
- Cloud Computing
- Data Flow
- Data Governance
- Data Lake
- Distributed File System
- Extract
- Transform
- Load
- Data Processing
- Data Quality
- Data Warehouse
- Decision-making
- Cost Control
- Dashboard
- GCS
- Generative Artificial Intelligence (AI)
- Git
- Good Clinical Practice
- Google Cloud
- Meta-data Management
- Natural Language
- OCP
- Google Cloud Platform
- Identity Management
- Layout
- Management
- Migration
- Privacy
- Data Architecture
- Data Compression
- Data Engineering
- SQL
- Scalability
- Semantics
- DevOps
- ELT
- FOCUS
- PySpark
- Python
- Query Optimization
- Reporting
- Version Control
- Solution Architecture
- Sprint
- Streaming
- Testing
- Trend Analysis
- Use Cases
- Workflow
Summary
Role : Google Cloud Data Architect IAM Data Modernization
Location : Dallas, TX / Charlotte, NC (Hybrid 4 days office)
Highly Preferred OCP exp
Project/Program
Identity & Access Management (IAM) Data Modernization migration of an onpremises SQL data warehouse to a targetstate Data Lake on Google Cloud (Google Cloud Platform), enabling metrics & reporting, advanced analytics, and GenAI use cases (natural language querying, accelerated summarization, crossdomain trend analysis) leveraging PySparkbased processing, cloudnative DevOps CI/CD pipelines, and containerized deployments on OpenShift (OCP) to deliver scalable, secure, and highperformance data solutions.
About Program/Project
The IAM Data Modernization project involves migrating an on-premises SQL data warehouse to a target state Data Lake in Google Cloud Platform cloud environment. Key highlights include:
- Integration Scope: 30+ source system data ingestions and multiple downstream integrations
- Capabilities: Metrics, reporting, and Gen AI use cases with natural language querying, advanced pattern/trend analysis, faster summarizations, and cross-domain metric monitoring
- Benefits:
- Scalability and access to advanced cloud functionality
- Highly available and performant semantic layer with historical data support
- Unified data strategy for executive reporting, analytics, and Gen AI across cyber domains
This modernization establishes a single source of truth for enterprise-wide data-driven decision-making.
Required Skills
DevOps / CICD
- Experience implementing CI/CD pipelines for data and analytics workloads
- Familiarity with Gitbased source control, build automation, and deployment strategies
Containers & Platform
- Experience with OpenShift Container Platform (OCP) for deploying data workloads and services
- Understanding of containerized architecture, scaling, and environment management
- Proven ability to build CI/CD pipelines for data and infrastructure workloads
- Experience managing secrets securely using Google Cloud Platform Secret Manager
- Ownership of observability, SLOs, dashboards, alerts, and runbooks
- Proficiency in logging, monitoring, and alerting for data pipelines and platform reliability
Big Data & Processing
- Handson experience with PySpark for ETL/ELT, data transformation, and performance optimization
- Solid understanding of distributed data processing concepts
Data & Cloud Architecture
- Strong experience designing data platforms on Google Cloud Platform (Google Cloud Platform)
- Experience with Data Lakes, data warehousing, and largescale migration programs
Data Lake Architecture & Storage
- Proven experience designing and implementing data lake architectures (e.g., Bronze/Silver/Gold or layered models).
- Strong knowledge of Cloud Storage (GCS) design, including bucket layout, naming conventions, lifecycle policies, and access controls
- Experience with Hadoop/HDFSarchitecture, distributed file systems, and data locality principles
- Hands-on experience with columnar data formats (Parquet, Avro, ORC) and compression techniques
- Expertise in partitioning strategies, backfills, and large-scale data organization
- Ability to design data models optimized for analytics and BI consumption
Data Ingestion & Orchestration
- Experience building batch and streaming ingestion pipelinesusing Google Cloud Platform-native services
- Knowledge of Pub/Sub-based streaming architectures, event schema design, and versioning
- Strong understanding of incremental ingestion and CDC patterns, including idempotency and deduplication
- Hands-on experience with workflow orchestrationtools (Cloud Composer / Airflow)
- Ability to design robust error handling, replay, and backfill mechanisms
Data Processing & Transformation
- Experience developing scalable batch and streaming pipelinesusing Dataflow (Apache Beam) and/or Spark (Dataproc)
- Strong proficiency in BigQuery SQL, including query optimization, partitioning, clustering, and cost control.
- Hands-on experience with Hadoop MapReduceand ecosystem tools (Hive, Pig, Sqoop)
- Advanced Python programming skillsfor data engineering, including testing and maintainable code design
- Experience managing schema evolutionwhile minimizing downstream impact
Analytics & Data Serving
- Expertise in BigQuery performance optimizationand data serving patterns
- Experience building semantic layers and governed metricsfor consistent analytics
- Familiarity with BI integration, access controls, and dashboard standards
- Understanding of data exposure patterns via views, APIs, or curated datasets
Data Governance, Quality & Metadata
- Experience implementing data catalogs, metadata management, and ownership models
- Understanding of data lineagefor auditability and troubleshooting
- Strong focus on data quality frameworks, including validation, freshness checks, and alerting
- Experience defining and enforcing data contracts, schemas, and SLAs
Good to have
Security, Privacy & Compliance
- Hands-on experience implementing fine-grained access controlsfor BigQuery and GCS
- Experience with Sprint planning and helping team technically.
- Strong stakeholder communication and solutionarchitecture skills
Qualifications
- Experience: [10 14]+ years in DevOps and Data Architecture, 5+ years designing on Pyspark/Google Cloud Platform/OCP at scale; prior onprem cloud migration a must.
- Education: Bachelor s/Master s in Computer Science, Information Systems, or equivalent experience.
- Certifications: Google Cloud Professional Cloud Architect/DevOps/OCP (required or within 3 months). Plus: Professional Data Engineer, Security Engineer.
Other Job Details:
- Job Type: C2C or W2
- Pay Rate: $60-65 hr on C2C / $55/hr on W2
- Duration: 12 months (high possibility of extension)
- Location: Dallas, TX / Charlotte, NC (Hybrid 4 days in office)
- Docs Required: ID proof will be required
Please review the job description and let me know if it aligns with your experience. Looking forward to your response.
- Dice Id: 91093052
- Position Id: 8964735
- Posted 12 hours ago
Company Info
About OMG Technologies
From our humble beginnings as a small IT service provider in 2006, OMG Technologies has grown to become a reliable provider of management and IT consulting services, while still maintaining agility of a small company. With multiple clients in various industries, OMG Technologies continues to grow at a good, steady and consistent pace based on our client focused approach.
Whatever it Takes to deliver value is our CORE approach
Similar Jobs
It looks like there aren't any Similar Jobs for this job yet.
Search all similar jobs