Senior AI Engineer - Data Scientist

Remote • Posted 13 hours ago • Updated 13 hours ago
Full Time
No Travel Required
Remote
130000 - 150000/yr
Company Branding Image
Fitment

Dice Job Match Score™

👤 Reviewing your profile...

Job Details

Skills

  • Generative AI (GenAI)
  • Large Language Models (LLMs)
  • LLMOps / MLOps / AIOps
  • Retrieval-Augmented Generation (RAG)
  • Agentic AI / Autonomous Agents
  • Prompt Engineering / Prompt Optimization
  • LLM Evaluation & Safety (Hallucination Detection
  • Bias Mitigation)
  • Fine-Tuning / Instruction Tuning
  • Foundation Models
  • AI Model Deployment / Production AI
  • Knowledge Graphs (KG)
  • Semantic Search / Semantic Layer
  • Healthcare Ontologies (CMS
  • FHIR
  • HL7)
  • Ontology Engineering / Semantic Modeling
  • Protégé Ontology Development
  • Graph Databases (Neo4j
  • Neptune
  • TigerGraph)
  • Semantic Retrieval
  • AWS (Bedrock
  • SageMaker
  • Lambda
  • S3)
  • Databricks (Unity Catalog
  • Feature Store)
  • Data Pipelines / ETL / ELT
  • Distributed Data Processing (Spark
  • PySpark)
  • Microservices Architecture
  • API Development (REST / GraphQL)
  • python

Summary

Job Description:

We’re seeking a Senior AI Engineer / Data Scientist to lead the design, deployment, and scaling of enterprise AI capabilities—specifically large language model (LLM) solutions, LLMOps practices, and the development of a healthcare ontology/knowledge graph to enhance a complex data environment.

 

Responsibilities

  • Ontology & Knowledge Graph
    • Design and maintain a healthcare ontology to normalize CMS data across claims, providers, and workflows.
    • Build and manage knowledge graphs (RDF/OWL or property graph) to support semantic search, inference, and RAG augmentation.
    • Develop graph data pipelines for ingestion, transformation, and entity resolution aligned with governance standards.
    • Collaborate with SMEs to define controlled vocabularies and create reusable semantic APIs for analytics and AI.
  • GenAI & LLMOps
    • Architect and operationalize LLMs for production use cases including RAG, agentic workflows, and MCP tools.
    • Build LLM evaluation and safety frameworks (prompt quality, grounding, hallucination detection, bias checks) with automated testing and human-in-the-loop reviews.
    • Design cost- and latency-aware pipelines with observability for performance and reliability.
    • Implement LLMOps best practices: prompt versioning, CI/CD for artifacts, rollout strategies, and A/B testing.
    • Integrate vector databases and optimize chunking, embeddings, and retrieval for high-quality responses.
  • Platform & System Architecture
    • Support productionalization of AI/ML workflows with automated quality checks and lifecycle orchestration.
    • Ensure data security, governance, and CMS compliance.
    • Contribute to high-level system design for integrating new AI capabilities into a cloud-based analytics platform.
    • Maintain documentation and acceptance criteria for system changes.

Basic Qualifications

  • Education/Experience: 10+ years with BS/BA; 8+ years with MS/MA; 5+ years with PhD in Computer Science, Data Science, or related field.
  • Hands-on experience with LLMs and GenAI solutions, including prompt engineering, RAG architecture, and LLMOps practices.
  • Proven experience in ontology design and knowledge graph development for complex data-driven systems.
  • Experience with Databricks, Snowflake, and AWS Cloud Services.
  • Proficiency in Python and SQL (Snowflake SQL).
  • Experience with CI/CD workflows and automated deployments.
  • Familiarity with Scaled Agile Framework (SAFe).
  • Excellent communication skills and ability to work independently.
  • ship required.

Preferred Qualifications

  • Experience with Databricks E2 components (Unity Catalog, Feature Store).
  • Knowledge of CMS systems and Medicare/Medicaid data.
  • Familiarity with LLM/GenAI tooling (LangChain, LlamaIndex, Hugging Face, AWS Bedrock).
  • Experience with vector databases and RAG orchestration.
  • Knowledge graph tools: Neo4j, TigerGraph, AWS Neptune, RDF/OWL, SPARQL, Gremlin, Cypher, Protégé.
  • Model lifecycle & governance: MLflow, Model Registry, feature stores, LLM safety testing.
  • Observability & automation: GitHub Actions/Jenkins, Terraform, Docker/Kubernetes, PrometheGrafana.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 91140539
  • Position Id: 8922669
  • Posted 13 hours ago

Company Info

About Tria Federal

Tria Federal (Tria) is the premier middle-market IT and Advisory services provider delivering digital transformation solutions to Civilian, Defense, and Intelligence agencies across the federal sector. With a future-forward vision and a mission rooted in service, we bridge capability gaps to help government agencies work faster, grow smarter, and stay nimble in the face of change. Our capabilities are far-reaching and expansive, spanning the lifecycle of digital transformation from end to end. Regardless of agency, whatever the mission, at any stage of the modernization journey, we supercharge organizational governance, business processes, and data-driven decision-making to transform the business of government.

We maintain a Prime seat across multiple lanes on our legacy multiple-award procurement vehicles - to include Best-in-Class Government-Wide Acquisition Contracts (GWACs), GSA Federal Schedules, and Agency-Specific Indefinite Delivery/ Indefinite Quantity (IDIQ) and Blanket Purchase Agreements (BPAs)

About_Company_OneAbout_Company_Two
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

It looks like there aren't any Similar Jobs for this job yet.

Search all similar jobs