Onsite role from day one to Dallas, TX. NO C2C allowed.
Role Overview
We are seeking a highly skilled AWS Infrastructure Engineer to design, build, and manage scalable, secure, and highly available AWS infrastructure. This role will be responsible for creating AWS infrastructure from the ground up, defining policies, roles and standards, and deploying and managing resources using Infrastructure as Code (Terraform). The engineer will also support data platforms, including S3, EMR , Kafka and other AWS services , ensuring performance, reliability, versioning , scalability and cost efficiency.
Key Responsibilities
AWS Infrastructure & Architecture
- Design and implement AWS cloud infrastructure following best practices for security, scalability, and availability
- Architect solutions across multiple availability zones for high availability and fault tolerance
- Define and enforce AWS resource naming standards and tagging strategies
- Implement IAM policies, bucket policies, and security controls aligned with organizational governance
- Manage core AWS services including VPC, EC2, S3, IAM, CloudWatch, and EMR
Amazon S3 & Data Storage
- Design and manage Amazon S3 buckets, including:
- Bucket policies and access controls
- Encryption, versioning, and logging
- Define and implement S3 lifecycle management policies for cost optimization and data retention
- Establish data partitioning and versioning strategies for large-scale datasets
Infrastructure as Code & Automation
- Develop, deploy, and maintain AWS infrastructure using Terraform
- Create reusable Terraform modules and enforce IaC best practices
- Manage infrastructure deployments across multiple environments (dev, test, prod)
- Integrate Terraform deployments with CI/CD pipelines
- Perform infrastructure upgrades and changes with minimal downtime
Data & EMR Platform
- Design, deploy, and manage AWS EMR clusters
- Cluster sizing, node roles (master/core/task), and configurations
- Auto-scaling and performance tuning
- Manage cluster lifecycle (provisioning, scaling, patching, termination)
- Optimize Spark, Hive, and Hadoop workloads for performance and cost
- Integrate and manage Kafka for streaming data pipelines
- Define Kafka partitioning and scaling strategies
Monitoring, Operations & Documentation
- Implement monitoring, logging, and alerting using Amazon CloudWatch and related tools
- Troubleshoot infrastructure, EMR, and data platform issues
- Document architecture, standards, and operational procedures in Confluence
- Collaborate with data engineers, security teams, and application teams
Required Skills & Experience
- 10+ years of experience in AWS infrastructure and cloud engineering
- Strong hands-on experience with Terraform and Infrastructure as Code
- Deep knowledge of AWS services: S3, EC2, VPC, IAM, EMR, CloudWatch
- Experience designing highly available, multi-AZ architectures
- Solid understanding of security best practices and policy management
- Hands-on experience with Big Data platforms (EMR, Spark, Hive, Hadoop)
- Experience with Kafka, including partitioning and scaling strategies
- Strong scripting skills (Python, Bash)
- Experience working in Linux-based environments
Preferred Qualifications
- AWS Certified Solutions Architect (Associate or Professional)
- AWS Certified Data Analytics Specialty
Familiarity with CI/CD tools (GitHub Actions, Jenkins, GitLab CI)
- Experience in large-scale or regulated enterprise environments
What Success Looks Like
- AWS infrastructure is fully codified, repeatable, and secure
- EMR and Kafka platforms scale reliably with predictable performance
- Clear standards, policies, and documentation are adopted across teams
- Infrastructure costs are optimized without sacrificing reliability