Lead solution architecture and technical governance for a large-scale historical data migration (3B+ records) from DB2 to Salesforce using a PostgreSQL staging/ETL layer. Own end-to-end migration strategy, scalability/performance design, data quality and reconciliation, and phased (throttled) capability activation with incremental delta loads to enable reliable, auditable Production cutover.
Qualifications
● 20+ years in data architecture/engineering, leading large-scale migrations (hundreds of millions to billions of records) from legacy/mainframe sources (e.g., DB2) to cloud/SaaS targets.
● Strong SQL + PostgreSQL (staging/ETL persistence): schema design, partitioning, bulk loading, tuning, and operational rigor (HA/DR/backup, auditability, reconciliation).
● Delta/incremental load design (CDC/watermarking, micro-batching), with idempotent processing and replay/backfill strategies.
● Salesforce data loading at scale: object modeling, external IDs, load sequencing, Bulk API patterns, and governor/locking constraints.
● Data quality, de-duplication, and survivorship frameworks for customer/transaction domains; strong validation and reconciliation patterns.
● Cloud experience on AWS supporting large-scale data platforms/migration factories (e.g., S3, RDS/Aurora PostgreSQL, Glue, EMR/Spark, Lambda/Step Functions, IAM/KMS, CloudWatch).