Sr. DataOps Engineer

Overview

On Site
Depends on Experience
Accepts corp to corp applications
Contract - Independent
Contract - W2
Contract - 12 Month(s)

Skills

DataOps
CloudOps
Automation
AWS
Databricks
Python
Communication
Computer Networking
Bash
Caching
Cloud Computing
Cloud Security
Cloud Storage
Amazon SageMaker
Amazon Web Services
Artificial Intelligence
Backup Administration
Access Control
Agile
Dashboard
Data Governance
Data Modeling
Continuous Integration
Continuous Integration and Development
Disaster Recovery
Cost Management
Data Warehouse
FOCUS
Failover
GitHub
GitLab
Database
DevOps
DevSecOps
Encryption
Extract
Transform
Load
Good Clinical Practice
Amazon EC2
Amazon RDS
Amazon S3
Regulatory Compliance
Orchestration
Lifecycle Management
Prompt Engineering
Provisioning
Remote Desktop Services
High Availability
Incident Management
Storage
System On A Chip
Network
Optimization
Real-time
SaaS
Scripting
Scripting Language
Kubernetes
LangChain
Machine Learning (ML)
Management
Microsoft Azure
Collaboration
Continuous Delivery
EBS
Virtual Private Cloud
Google Cloud Platform
HIPAA
IaaS
Terraform
Testing
Unity
Workflow

Job Details

Position: Sr. DataOps Engineer

Location: Onsite (Hybrid, Duluth, GA)

Experience: 8 years in CloudOps or DataOps

Skill Matrix

Skills

Years of Exp

Year lastly Used

Rating Out of 10

DataOps / CloudOps

Automation in AWS

Databricks

AWS

Python

Job Description:

We are seeking an experienced Senior DataOps Engineer to join our team. This candidate will have a strong background in DevOps, DataOps, or Cloud Engineering practices, with extensive experience in automating the CICD pipelines and modern data stack technologies.

Key Responsibilities:

  • Develop and maintain robust, scalable data pipelines and infrastructure automation workflows using GitHub, AWS, and Databricks.
  • Implement and manage CI/CD pipelines using GitHub Actions and GitLab CI/CD for automated infrastructure deployment, testing, and validation.
  • Deploy and manage Databricks LLM Runtime or custom Hugging Face models within Databricks notebooks and model serving endpoints.
  • Manage and optimize Cloud Infrastructure costs, usage, and performance through tagging policies, right-sizing EC2 instances, storage tiering strategies, and auto-scaling.
  • Set up infrastructure observability and performance dashboards using AWS CloudWatch for real-time insights into cloud resources and data pipelines.
  • Develop and manage Terraform or CloudFormation modules to automate infrastructure provisioning across AWS accounts and environments.
  • Implement and enforce cloud security policies, IAM roles, encryption mechanisms (KMS), and compliance configurations.
  • Administer Databricks Workspaces, clusters, access controls, and integrations with Cloud Storage and identity providers.
  • Enforce DevSecOps practices for infrastructure-as-code, ensuring all changes are peer-reviewed, tested, and compliant with internal security policies.
  • Coordinate cloud software releases, patching schedules, and vulnerability remediations using Systems Manager Patch Manage.
  • Automate AWS housekeeping and operational tasks such as:
  • Cleanup of unused EBS Volumes, snapshots, old AMIs
  • Rotation of secrets and credentials using secrets manager
  • Log retention enforcement using S3 Lifecycle policies and CloudWatch Log groups
  • Perform incident response, disaster recovery planning, and post-mortem analysis for operational outages.
  • Collaborate with cross-functional teams including Data Scientists, Data Engineers, and other stakeholders to gather, implement the infrastructure and data requirements.

Required Skills and Experience:

  • 8+ years of experience in DataOps / CloudOps / DevOps roles, with strong focus on infrastructure automation, data pipeline operations, observability, and cloud administration.
  • Strong proficiency in at least one Scripting language (e.g., Python, Bash) and one infrastructure-as-code tool (e.g., Terraform, CloudFormation) for building automation scripts for AWS resource cleanup, tagging enforcement, monitoring and backups.
  • Hands-on experience integrating and operationalizing LLMs in production pipelines, including prompt management, caching, token-tracking, and post-processing.
  • Deep hands-on experience with AWS Services, including
    • Core: EC2, S3, RDS, CloudWatch, IAM, Lambda, VPC
    • Data Services: Athena, Glue, MSK, Redshift
    • Security: KMS, IAM, Config, CloudTrail, Secrets Manager
    • Operational: Auto Scaling, Systems Manager, CloudFormation/Terraform
    • Machine Learning/AI: Bedrock, SageMaker, OpenSearch serverless
  • Working knowledge of Databricks, including:
    • Cluster and workspace management, job orchestration
    • Integration with AWS Storage and identity (IAM passthrough)
  • Experience deploying and managing CI/CD workflows using GitHub Actions, GitLab CI, or AWS CodePipeline.
  • Strong understanding of cloud networking, including VPC Peering, Transit Gateway, security groups, and private link setup.
  • Familiarity with container orchestration platforms (e.g., Kubernetes, ECS) for deploying platform tools and services.
  • Strong understanding of data modeling, data warehousing concepts, and AI/ML Lifecycle management.
  • Knowledge of cost optimization strategies across compute, storage, and network layers.
  • Experience with data governance, logging, and compliance practices in cloud environments (e.g., SOC2, HIPAA, GDPR)
  • Bonus: Exposure to LangChain, Prompt Engineering frameworks, Retrieval Augmented Generation (RAG), and vector database integration (AWS OpenSearch, Pinecone, Milvus, etc.)

Preferred Qualifications:

  • AWS Certified Solutions Architect, DevOps Engineer or SysOps Administrator certifications.
  • Hands-on experience with multi-cloud environments, particularly Azure or Google Cloud Platform, in addition to AWS.
  • Experience with infrastructure cost management tools like AWS Cost Explorer, or FinOps dashboards.
  • Ability to write clean, production-grade Python code for automation scripts, operational tooling, and custom CloudOps Utilities.
  • Prior experience in supporting high-availability production environments with disaster recovery and failover architectures.
  • Understanding of Zero Trust architecture and security best practices in cloud-native environments.
  • Experience with automated cloud resources cleanup, tagging enforcement, and compliance-as-code using tools like Terraform Sentinel.
  • Familiarity with Databricks Unity Catalog, access control frameworks, and workspace governance.
  • Strong communication skills and experience working in agile cross-functional teams, ideally with Data Product or Platform Engineering teams.

Alchemy: Transforming Your Professional Vision into Reality

Since our inception in 2013, Alchemy has been dedicated to reshaping organizational performance through innovative IT services. With a vision to empower businesses seeking a transformative edge, we ve positioned ourselves at the forefront of digitization and software modernization.

Our name reflects our mission: to transmute technology into gold-standard solutions for our esteemed clients. We proudly serve a diverse range of sectors, including IT and ITES, BFSI, Telecom and Media, Automotive, Manufacturing, Energy, Oil and Gas, Real Estate, Retail, Healthcare, and more.

With a global footprint spanning the USA, India, Europe, Canada, Singapore, Japan, and parts of Central and West Africa, we harness a unique blend of competencies, frameworks, and cutting-edge technologies. Together, we drive growth and innovation across industries, helping organizations turn their visions into reality.

Alchemy Connecting Talent with Opportunities (Diversity, Equity and Inclusion)

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.