Role : Google Cloud Data Architect IAM Data Modernization
Location : Dallas, TX Iselin NJ and Charlotte NC/Onsite
Need Passport Number
Project/Program
Identity & Access Management (IAM) Data Modernization migration of an onpremises SQL data warehouse to a targetstate Data Lake on Google Cloud (Google Cloud Platform), enabling metrics & reporting, advanced analytics, and GenAI use cases (natural language querying, accelerated summarization, crossdomain trend analysis).
About Program/Project
The IAM Data Modernization project involves migrating an on-premises SQL data warehouse to a target state Data Lake in Google Cloud Platform cloud environment. Key highlights include:
- Integration Scope: 30+ source system data ingestions and multiple downstream integrations
- Capabilities: Metrics, reporting, and Gen AI use cases with natural language querying, advanced pattern/trend analysis, faster summarizations, and cross-domain metric monitoring
- Benefits:
- Scalability and access to advanced cloud functionality
- Highly available and performant semantic layer with historical data support
- Unified data strategy for executive reporting, analytics, and Gen AI across cyber domains
This modernization establishes a single source of truth for enterprise-wide data-driven decision-making.
Required Skills
Data Lake Architecture & Storage
- Proven experience designing and implementing data lake architectures (e.g., Bronze/Silver/Gold or layered models).
- Strong knowledge of Cloud Storage (GCS) design, including bucket layout, naming conventions, lifecycle policies, and access controls
Experience with Hadoop/HDFS architecture, distributed file systems, and data locality principles
- Hands-on experience with columnar data formats (Parquet, Avro, ORC) and compression techniques
- Expertise in partitioning strategies, backfills, and large-scale data organization
- Ability to design data models optimized for analytics and BI consumption
Qualifications
- Experience: [10 14]+ years in data engineering/architecture, 5+ years designing on Google Cloud Platform at scale; prior onprem cloud migration a must.
- Education: Bachelor s/Master s in Computer Science, Information Systems, or equivalent experience.
- Certifications: Google Cloud Professional Cloud Architect (required or within 3 months). Plus: Professional Data Engineer, Security Engineer.
Data Ingestion & Orchestration
Experience building batch and streaming ingestion pipelines using Google Cloud Platform-native services
Knowledge of Pub/Sub-based streaming architectures, event schema design, and versioning
Strong understanding of incremental ingestion and CDC patterns, including idempotency and deduplication
Hands-on experience with workflow orchestration tools (Cloud Composer / Airflow)
Ability to design robust error handling, replay, and backfill mechanisms
Data Processing & Transformation
Experience developing scalable batch and streaming pipelines using Dataflow (Apache Beam) and/or Spark (Dataproc)
Strong proficiency in BigQuery SQL, including query optimization, partitioning, clustering, and cost control.
Hands-on experience with Hadoop MapReduce and ecosystem tools (Hive, Pig, Sqoop)
Advanced Python programming skills for data engineering, including testing and maintainable code design
Experience managing schema evolution while minimizing downstream impact