Senior Databricks Solution Engineer

Hybrid in Dallas, TX, US • Posted 8 hours ago • Updated 8 hours ago
Contract W2
Contract Independent
Contract Corp To Corp
No Travel Required
Hybrid
Depends on Experience
Company Branding Image
Fitment

Dice Job Match Score™

🔗 Matching skills to job...

Job Details

Skills

  • Senior Databricks Solution Engineer - Enterprise Data Platform (EDP) & Customer Data Platform (CDP) / Azure Databricks
  • Lakeflow
  • Delta Live Tables
  • ADF |
  • Kafka |
  • MongoDB

Summary

Role : Senior Databricks Solution Engineer - Enterprise Data Platform (EDP) & Customer Data Platform (CDP) / Azure Databricks | Lakeflow | Delta Live Tables | ADF | Kafka | MongoDB

Location: United States (Hybrid / Remote)

Experience: 10 + Years

 

 

 

This role blends solution architecture, hands‑on data engineering, and enterprise enablement, with a strong focus on retail, customer, marketing, and digital data. You will define and enforce enterprise data standards, build reusable frameworks, and deliver secure, governed data pipelines that support analytics, personalization, marketing activation, and executive reporting.


Key Responsibilities


1. Enterprise Lakehouse Architecture & Standards

  • Define and enforce Client’s Lakehouse architecture standards using Azure Databricks, aligned to scalability, security, and cost efficiency.
  • Implement and operationalize the Medallion Architecture (Bronze / Silver / Gold) as the enterprise standard:
    • Bronze – raw, immutable, audit‑ready ingestion
    • Silver – cleansed, conformed, validated, and privacy‑compliant datasets
    • Gold – curated, analytics‑ready, semantic‑aligned data products (including standardized current and history tables where required, e.g., composite + composite_hist / SCD Type 2)
  • Establish reference architectures, design patterns, and guardrails that enable consistent adoption across Stores, Digital, Marketing, Supply Chain, and Corporate domains.
  • Standardize on Unity Catalog for all new and migrated workloads (minimize/retire Hive Metastore usage), including consistent catalog/schema conventions, data ownership, and access controls.

2. Declarative Pipelines & Lakeflow (Databricks‑native)

  • Build declarative data pipelines using Databricks Lakeflow and Delta Live Tables (DLT) as the preferred enterprise pattern.
  • Define data quality expectations, freshness SLAs, and validation rules directly within pipelines.
  • Leverage DLT capabilities for:
    • automated dependency management
    • data quality enforcement
    • lineage and observability
    • operational simplicity at scale

3. Inbound Data Ingestion Standards (Kafka, Lakeflow Connectors, Auto Loader, Databricks Streaming)

  • Standardize inbound ingestion using:
    • Apache Kafka for event-driven and streaming ingestion (pub/sub, CDC fanout, operational events)
    • Databricks Auto Loader for scalable, incremental file ingestion from cloud object storage with schema inference/evolution
    • Lakeflow Connectors for managed ingestion from SaaS applications and databases (connector-based patterns with governed landing into Bronze)
    • Databricks Structured Streaming (and streaming tables where applicable) for continuous ingestion and low-latency processing into Delta
  • Design resilient ingestion frameworks that support high‑volume customer, marketing, and operational data with schema evolution and fault tolerance.
  • Apply consistent ingestion controls across internal systems and external vendors.

4. CDP, Marketing & Digital Data Integrations

  • Design and deliver customer and marketing data integrations supporting Client’s CDP and activation ecosystem.
  • Build and manage pipelines integrating data from platforms including:
    • Acxiom / BRIDGE
    • LiveRamp
    • Blueshift
    • Ovative
    • Google Analytics
    • Google Ads
    • Meta Ads (Facebook / Instagram)
  • Enable enterprise use cases such as:
    • customer identity resolution
    • audience segmentation and activation
    • marketing attribution and performance analytics
    • personalization and lifecycle analysis

5. Enterprise Data Modeling Methodology (Explicit Standard)

  • Define and enforce Client’s Enterprise Data Modeling methodology, including:
    • Canonical data models for shared enterprise entities (customer, store, product, transaction, vendor)
    • Dimensional modeling (Star / Snowflake schemas) for analytics and reporting
    • Semantic modeling aligned to downstream BI and analytics tools
  • Ensure conformed dimensions and consistent metric definitions across domains.
  • Partner with analytics and business teams to validate business meaning and usability.

6. Semantic Enablement & Consumption (MSTR, Power BI, Unity Catalog)

  • Provision trusted, analytics‑ready datasets for:
    • MicroStrategy (MSTR)
    • Power BI
  • Leverage Unity Catalog as the system of record for:
    • data governance and access control
    • lineage and discovery
    • semantic consistency and certification
    • data classification via Unity Catalog tags (e.g., PII/sensitivity, domain, certification) to drive masking, policy enforcement, and controlled publishing
  • Promote enterprise metric definitions via governed semantic models (e.g., Unity Catalog metrics/semantic layer where adopted) to ensure consistent KPIs across MSTR, Power BI, and downstream consumers.
  • Ensure Gold‑layer datasets are optimized, documented, and certified for enterprise consumption.

7. Orchestration & Integration (ADF, Airflow, Databricks Workflows)

  • Orchestrate end‑to‑end pipelines using Azure Data Factory (ADF) and/or Apache Airflow, integrated with Databricks Workflows.
  • Define dependency management, retry patterns, alerting, and operational ownership for production workloads.

8. PII Protection & 3‑Layer Encryption (Client Standard)

Establish PII protection as a non‑negotiable enterprise standard, including mandatory 3‑layer encryption:

  1. In‑Transit Encryption
    • TLS‑based secure transport for all internal and external transfers.
  2. File‑Level Encryption at Rest
    • Encrypted files and objects for vendor, marketing, and partner exchanges.
  3. Record‑Level / Element‑Level Encryption & Hashing
    • Attribute‑level protection for PII used in CDP, marketing, and analytics workflows.
    • Enforce protections using Unity Catalog controls where applicable (e.g., masking policies and fine‑grained access controls) to ensure governed use across analytics and activation.

Ensure full auditability, regulatory compliance (GDPR, CCPA), and consistent enforcement across platforms and vendors.


9. Secure Vendor & Partner Data Exchange

  • Design and operate secure, high‑volume data exchanges with advertising, marketing, and data partners.
  • Validate keys, credentials, service accounts, and secure repositories (SFTP, cloud object storage).
  • Provide technical direction to vendors to ensure compliant, end‑to‑end delivery under tight timelines.

10. Performance, Reliability & Cost Optimization

  • Optimize Spark, DLT, and SQL workloads for performance, reliability, and cost efficiency.
  • Contribute to production support, incident analysis, and continuous platform improvements.
  • Implement production operational standards using the enterprise toolchain (e.g., New Relic monitoring, PagerDuty incident response/on-call, ServiceNow ticketing), including alerting, runbooks, and SLAs.

11. Leadership & Enablement

  • Mentor engineers and lead architecture reviews across platform, analytics, and marketing teams.
  • Drive adoption of enterprise patterns through documentation, reviews, and enablement sessions.

Required Qualifications

  • 8+ years in data engineering, solution architecture, or platform engineering.
  • Deep experience with Azure Databricks, Spark, Delta Lake, Lakeflow, Delta Live Tables, Python/PySpark, SQL.
  • Experience with KafkaAuto Loader / Auto StreamingADF, and/or Airflow.
  • Strong experience in enterprise data modeling, governance, and BI enablement.
  • Proven delivery of secure, compliant, enterprise‑scale data platforms.

 

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: RTL327349
  • Position Id: 8954615
  • Posted 8 hours ago

Company Info

About Cyma Systems Inc

Cyma Systems is a technology solutions firm based in Connecticut serving mid-sized and Fortune 1000 businesses. We work with a range of corporate commercial clients as well as with the government sector. As a full-service consulting firm, we engage with our clients to share ideas on how we can help their businesses to become more profitable, to get a jump start on competitors by bringing products to market quickly, reducing costs, maximizing the use of existing technologies, driving more sales through the channel, and to bring additional value by forming a mutually beneficial partnership.

We think that the essence of a successful business is its ability to adapt and transform keeping ground realities in sight. At Cyma we work on leveraging the right synergies between technology and people that work towards transforming the outlook and results of your business.

About_Company_OneAbout_Company_Two
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

Remote

Today

Easy Apply

Contract, Third Party

Depends on Experience

Remote

Today

Easy Apply

Contract

Depends on Experience

Remote

Today

Easy Apply

Contract

Depends on Experience

Remote

Today

Easy Apply

Contract

Depends on Experience

Search all similar jobs