Senior Data & AI Platform Architect/Engineer

Overview

On Site
Depends on Experience
Contract - W2
Contract - 10 Month(s)
No Travel Required
Unable to Provide Sponsorship

Skills

Senior Data & AI Platform Architect
Databricks
DataHub
Progress semaphore
GitLab CI/CD
MinIO/Alluxio storage solution
Ignition
HashiCorp Vault
Splunk
PointGuard AI
and large-scale data architectures.
ETL
Delta lake

Job Details

Overview

We are seeking an experienced Senior Data & AI Platform Architect (Contractor) to design, build, and deliver an enterprise-grade, secure, end-to-end Data & AI Platform. The contractor will implement ingestion, secure data transfer, object storage, governance, ML engineering, metadata, observability, and AI security layers across the full modern data stack based on Databricks over a 9 month roadmap.

The ideal candidate has deep hands-on experience with Databricks, DataHub , Progress semaphore, GitLab CI/CD , MinIO/Alluxio storage solution, Ignition,HashiCorp Vault,  Splunk,  PointGuard AI, and large-scale data architectures.

Responsibilities

·      Architecture & Platform Deployment/Development

·      Architect and implement a full-scale, cloud-native Data & AI platform based on Databricks and Datahub integrating ingestion, storage, processing, ML, security, and governance components.

·      Design and deploy Ignition-based ingestion/orchestration pipelines for batch, streaming, and event-driven data.

·      Implement secure, compliant data transfer mechanisms (accelerated transfer, API-based ingestion, encryption in transit).

·      Deploy and operationalize storage/iceberg table format storage as the core object store integrated with Databricks and feature pipelines.

·      Build scalable ETL/ELT data pipelines using Databricks, Delta Lake, Unity Catalog, and Lakehouse best practices.

Machine Learning & MLOps

·      Deploy Databricks Feature Store, MLflow Model Registry, and automated ML pipelines.

·      Implement model governance, lineage, monitoring, rollback, and lifecycle management.

·      Deliver end-to-end pipelines for both traditional ML and LLM/GenAI (Agent Bricks, RAG, model serving).

Security & Compliance

·      Integrate PointGuard AI for AI posture, anomaly detection, compliance, and secure model operations.

·      Implement HashiCorp Vault for secrets management, transit encryption, and certificate workflows.

·      Build zero-trust RBAC/ABAC frameworks aligned with enterprise and federal standards.

·      Integrate audit logging to Splunk across all platform components.

Governance & Metadata

·      Deploy and configure DataHub for metadata ingestion, data lineage, dataset classification, and lifecycle management.

·      Integrate DataHub with Databricks, Vault, storage solution and orchestration pipelines.

·      Implement governance policies including data classification, tagging, and access workflows.

CI/CD & DevOps

·      Build and maintain GitLab CI/CD pipelines for data pipelines, ML pipelines, IaC, and model deployments.

·      Implement automated code scanning, quality gates, DAG submission workflows, and artifact management.

·      Use Terraform for IaC, resource deployment, and environment reproducibility.

Observability & Monitoring

·      Design and implement observability dashboards for data pipelines, ML pipelines, and infrastructure components.

·      Leverage Datadog with Databricks jobs,cluster,workflow etc monitoring.

·      Integrate real-time audit, compliance logs, and anomaly alerts into Splunk.

Collaboration & Delivery

·       Work closely with internal engineering team, data governance, and security teams.

·       Provide knowledge transfer, documentation, and hands-on training.

·       Deliver project milestones on a structured 9-month roadmap.

Required Qualifications

Technical Skills

  • 10+ years building enterprise Data platforms, ML pipelines, or cloud-native architectures
  • Strong hands-on experience with:

o   Databricks (Delta Lake, Workflows, Feature Store, MLflow, Model Serving, Unity Catalog, Declarative pipeline, Clean Room)

o   Ingestion and Orchestration -Prefect (or Apache Airflow/Dagster)

o   Familiarity with MinIO Aistor object storage  or Alluxio and secure Data transfer and Data lineage

o   HashiCorp Vault secrets and AWS KMS for secret management(Experience connecting Vault to Databricks & CI/CD pipelines)

o   GitLab CI/CD pipelines (Pipeline-as-code,DAG submission automation,Automated code scanning (SAST/DAST),Infrastructure-as-code (Terraform)

o   ML model deployment automation

o   Splunk observability integration (logs from storage access, Vault, Databricks, orchestration engines etc)

o   Ability to integrate Datahub with Databricks and ML ecosystem (Automated dataset classification,lineage tracking etc)

o   PointGuard AI or similar AI/security posture/runtime tools,AI model risk management framework,RBAC/ABAC architecture

·       Proficiency in Python, SPARK, SQL, REST APIs,Terraform

·       Experience with Kubernetes, Docker

·       Cloud expertise in AWS/Azure/Google Cloud Platform (AWS preferred)

·       Strong understanding of distributed systems, security architecture, governance, and data compliance

Preferred Qualifications

·       Databricks Certified Data Engineer Professional or Databricks Certified Machine learning Professional.

·       Experience implementing AI governance frameworks (NIST AI RMF, DOE cybersecurity frameworks)

·       Experience in public sector, government, or regulated environments

·       Experience with homomorphic encryption or advanced secure computation

·       Experience integrating Denodo or similar data virtualization layers (optional)

Soft Skills

·       Excellent documentation and diagramming skills

·       Ability to translate business requirements into technical designs and technical execution

·       Strong communication skills with technical & non-technical stakeholders

·       Ability to work independently and deliver long-term projects with minimal oversight

Duration & Engagement

·       9-month contract with possible extension(01/12/2026-09/30/2026)

·       Full-time engagement -40 hours per week

·       Hybrid with few days onsite depending on requirements

 

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.