Data Engineer

Remote • Posted 2 hours ago • Updated 2 hours ago
Contract Independent
Contract W2
No Travel Required
Able to Sponsor
Remote
Depends on Experience
Fitment

Dice Job Match Score™

🛠️ Calibrating flux capacitors...

Job Details

Skills

  • Python
  • Apache Kafka
  • Apache Parquet
  • Apache Spark
  • Artificial Intelligence
  • Bloomberg
  • Business Intelligence
  • Cloud Computing
  • Collaboration
  • Conflict Resolution
  • DAX
  • Data Domain
  • Data Engineering
  • Data Governance
  • Data Integrity
  • Data Modeling
  • Data Structure
  • Data Validation
  • Database
  • Decision-making
  • Fixed Income
  • Fluency
  • Front Office
  • Generative Artificial Intelligence (AI)
  • ELT
  • Equities
  • Extract, Transform, Load
  • Analytics
  • Distribution
  • Docker
  • RBAC
  • RESTful
  • Real-time
  • Reasoning
  • Reference Data
  • Pricing
  • Privacy
  • Problem Solving
  • Regulatory Compliance
  • Ontologies
  • Orchestration
  • Portfolio Management
  • Modeling
  • Natural Language
  • Normalization
  • Microsoft Azure
  • Microsoft Excel
  • Microsoft Power BI
  • Machine Learning (ML)
  • Meta-data Management
  • Analytical Skill
  • Documentation
  • Kubernetes
  • System Integration
  • Terraform
  • Visualization
  • Scala
  • Semantics
  • Storage
  • Streaming
  • Microsoft
  • API
  • Java
  • Reporting
  • Research
  • SQL
  • SSO
  • Work Ethic
  • Workflow

Summary

We are seeking a strong Data Engineer who can own and evolve the data foundations supporting our investment processes. You will engineer the data pipelines, models, and systems that power decision-making across portfolio management, risk, research, and attribution. This role is ideal for someone who excels at data engineering in a highly analytical, investment-focused environment and wants to work directly with investment professionals and quantitative teams.

Role:

You will design, build, and maintain scalable, cloud-ready data pipelines and infrastructure enabling investment business teams—research analysts, quantitative modelers, traders, risk managers, and portfolio managers—to access accurate, timely, and high-quality data. The role blends engineering rigor with fluency in investment data, helping power analytics, models, and mission-critical investment applications.

Key Responsibilities:

  • Data Pipeline & Engineering Development

    • Build and optimize end-to-end data pipelines to support ingestion, scrubbing, transformation, and distribution of investment data.

    • Implement robust ETL/ELT processes for structured and unstructured datasets across fixed income, Equities, and Multi asset domains.

  • Data Modeling, Architecture & Quality Control

    • Develop, maintain, and document data models, schemas, and database structures supporting investment analytics and operational workflows.

    • Implement data validation and quality frameworks aligned with investment teams’ data integrity standards.

  • Investment Data Domain Expertise:

    • Build fluency across various investment data domains and design data access patterns that deliver clean, consistent, and efficient access to consumers.

    • Understand how key model inputs (terms and conditions, pricing, index attributes, reference data) influence analytical outputs and decision-making.

  • Systems Integration & API Development

    • Integrate data with internal platforms and external vendors; streamline and standardize access through APIs, warehouses, or curated data layers.

    • Support enhancements to data streams and contribute to cloud-native data engineering initiatives aligned with machine learning and analytics engineering trends.

    • Enable seamless data access through analytics and visualization tools (e.g., Power BI, Excel, and BQL), ensuring that curated datasets, semantic models, and governed views are optimized for consumption by investment teams and downstream reporting workflows.

  • Collaboration & Cross-Functional Partnership

    • Partner with investment teams to understand requirements and deliver high-quality data that supports research, risk, and portfolio-management workflows.

    • Contribute to documentation, data governance, and best practices across the technology and investment organization.

  • Make Data AI-Ready

    • Design and structure datasets and documents so AI assistants (e.g., Copilot, ChatGPT) can seamlessly interpret and generate on-the-fly access requests for citizen developers.

    • Standardize metadata, tagging, and ontologies to reduce ambiguity in natural-language queries and improve AI-driven retrieval accuracy.

    • Implement data normalization, consistent schemas, and well-defined relationships to support LLM-driven reasoning across investment and operational domains.

    • Build and maintain semantic layers or knowledge catalogs that allow AI models to map user intent to underlying data assets.

    • Partner with architecture, security, and business teams to ensure AI-facing datasets meet governance, privacy, and compliance requirements.

    • Develop automated pipelines that prepare and refresh AI-ready data—including embeddings, vector indexes, and domain-specific feature stores.

    • Evaluate and optimize data structures for compatibility with enterprise GenAI tools (e.g., Microsoft Copilot, RAG systems, internal LLMs).

Required Skills & Qualifications:

Technical Skills:

  • Languages & Processing

    • Python, SQL; Scala/Java for Apache Spark (batch & streaming)

  • Pipelines & Lakehouse

    • Databricks, Spark; orchestration with Azure Data Factory / Airflow

    • Formats & storage: Delta Lake, Parquet; serving on Synapse / Fabric / Snowflake

  • Streaming

    • Event ingestion and near-real-time processing with Kafka / Azure Event Hubs

  • Modeling, Quality & Catalog

    • Dimensional & domain modeling, data contracts, and data validation frameworks

    • Catalog/lineage & governance with Microsoft Purview (or equivalent)

  • Analytics Access (BI & Front-Office)

    • Power BI (DAX, Power Query, XMLA, RLS/OLS, Deployment Pipelines, REST API)

    • Excel (Power Pivot/Power Query); Bloomberg BQL add-ins/templates

  •  AI-Ready Data

    • Ontologies/semantic layer, knowledge catalogs; embeddings & vector search

    • RAG patterns and alignment with Microsoft Copilot / enterprise LLMs

  • Cloud, Security & Operations

    • Azure services, Terraform/Bicep, Docker/Kubernetes, monitoring/observability

    • Secure access patterns (SSO/AAD, gateways), RBAC/ABAC, secrets management

Data & Investment Knowledge:

  • Nice to understand various investment data domains

Soft Skills

  • Intellectual curiosity, strong work ethic, and professional presence.

  • Excellent problem-solving skills and the ability to work across diverse technical and business teams.

  • Ability to communicate complex data concepts to investment stakeholders.

We look forward to reviewing your profile.
Talent Team @ SIALTP

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 10432398
  • Position Id: 6001
  • Posted 2 hours ago
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

Remote or Wellesley, Massachusetts

Today

Full-time

USD 64,890.00 per year

Remote or Saginaw, Michigan

Today

Easy Apply

Full-time

USD 140,000.00 - 150,000.00 per year

Remote or Walnut Creek, California

Today

Full-time

USD 145,204.80 - 217,796.80 per year

Remote

Today

Full-time

USD 119,000.00 - 161,000.00 per year

Search all similar jobs