Role: Data Architect
Location: Plano, TX (Onsite from Day 1)
Job Type: Full Time
Role:
We are seeking a hands-on Data Architect to design and evolve an AWS based data platform spanning streaming ingestion (Kafka), API/enterprise integration (MuleSoft), containerized data services (EKS), data lake on S3, interactive query with Athena, and analytics/reporting on Snowflake and Tableau.
Skills:
Data Architecture, Kafka connect, Snowflake, Big Data
The opportunity:
Architecture & Design:
Own the end to end data architecture across ingestion, storage, processing, serving, and visualization layers.
Define canonical data models and domain data contracts; lead conceptual/logical/physical data modeling and schema design for batch and streaming use cases.
Establish reference architectures and patterns for event driven and API led data integration (Kafka, MuleSoft).
Design secure, multi account AWS topologies (VPC, IAM, KMS) for data workloads; enforce governance, lineage, and cataloging.
Platform Enablement (New Platform Build out):
Lead the blueprint and incremental rollout of a new AWS data platform, including landing raw curated zones on S3, Athena for ad hoc/interactive SQL, and Snowflake for governed analytics and reporting.
Define platform SLAs/SLOs, cost guardrails, and chargeback/showback models; optimize storage/compute footprints.
Partner with DevOps to run containerized data services on EKS (e.g., stream processors, microservices, connectors) and automate with CI/CD.
Required:
- 8+ years in data architecture/engineering with 3+ years architecting on AWS.
- Proven design of S3 based data lakes with robust partitioning, lifecycle policies, and metadata/catalog strategy.
- Hands on experience with Kafka (topic design, schema evolution, consumer groups, throughput/latency tuning).
- Practical MuleSoft integration design (API led connectivity, RAML/OAS, policies, governance).
- Production experience with Amazon EKS for data/streaming microservices and connectors.
- Strong SQL and performance tuning with Athena; expertise selecting file formats/partitioning for cost/perf.
- Data warehousing on Snowflake (ELT, clustering, resource monitors, security) and delivering analytics via Tableau.
- Mastery of data modeling (3NF, dimensional/star, data vault), data contracts, and event modeling.
- Solid foundations in security, IAM/KMS, networking for data platforms, and cost management.