Agentic AI QA Engineer

Dallas, TX, US • Posted 2 days ago • Updated 2 days ago
Contract Independent
Contract W2
On-site
Depends on Experience
Company Branding Image
Fitment

Dice Job Match Score™

🛠️ Calibrating flux capacitors...

Job Details

Skills

  • Generative Artificial Intelligence (AI)
  • Testing
  • Test Strategy
  • Orchestration
  • Software Development Methodology
  • System Testing
  • Leadership
  • Machine Learning (ML)
  • Machine Learning Operations (ML Ops)
  • LangChain
  • Incident Management
  • Amazon Web Services
  • Artificial Intelligence
  • GitHub
  • Data Science
  • Message Queues
  • Microsoft Windows
  • Mentorship
  • TypeScript
  • Semantics

Summary

Note: Need Only Local to TX, Who can attend F2F Interview.

Position: Agentic AI QA Engineer Location: Dallas, TX - (Onsite) Duration: 6+ Months Contract Interview Mode: F2F Required

Job Description:
Agentic QA Engineer Generative AI & Agentic Systems (Agent, MultiAgent Testing)

Summary:
We are seeking a hands-on AI Engineer to design and execute end-to-end testing strategies for agentic AI solutions, including multi-agent systems in production-grade environments. This role partners with the Agentic Operations Team to ensure resiliency, reliability, accuracy, latency, orchestration, correctness, and scale. You will establish QA frameworks, build reusable test artifacts, drive macro-level validations across complex workflows, and lead the QA function for Agentic AI from Dev to Prod.

Key Responsibilities: Quality Strategy & Leadership Agentic & MultiAgent Testing Reliability, Resiliency, and Latency Accuracy & Macro-Level Validations Scale & Orchestration Dev Prod Readiness
Define and own the QA strategy for agentic/multi-agent AI systems across dev, staging, and prod.
Mentor a team of QA engineers; establish testing standards, coding guidelines for test harnesses, and review practices.
Partner with Agentic Operations, Data Science, MLOps, and Platform teams to embed QA in the SDLC and incident response.
Design tests for agent orchestration, tool calling, planner-executor loops, and inter-agent coordination (e.g., task decomposition, handoff integrity, and convergence to goals).
Validate state management, context windows, memory/knowledge stores, and prompt/graph correctness under varying conditions.
Implement scenario fuzzing (e.g., adversarial inputs, prompt perturbations, tool latency spikes, degraded APIs).
Create resilience testing suites: chaos experiments, failover, retries/backoff, circuit-breaking, and degraded mode behavior.
Establish latency SLOs and measure end-to-end response times across orchestration layers (LLM calls, tool invocations, queues).
Ensure reliability through soak tests, canary verifications, and automated rollbacks.
Define ground-truth and reference pipelines for task accuracy (exact match, semantic similarity, factuality checks).
Build macro validation frameworks that validate task outcomes across multi-step agent workflows (e.g., complex data pipelines, content generation + verification agent loops).
Instrument guardrail validations (toxicity, PII, hallucination, policy compliance).
Design load/stress tests for multi-agent graphs under scale (concurrency, throughput, queue depth, backpressure).
Validate orchestrator correctness (DAG execution, retries, branching, timeouts, compensation paths).
Engineer reusable test artifacts (scenario configs, synthetic datasets, prompt libraries, agent graph fixtures, simulators).
Integrate tests into CI/CD (pre-merge gates, nightly, canary) and production monitoring with alerting tied to KPIs.
Define release criteria and run operational readiness (performance, security, compliance, cost/latency budgets).
Build post-deployment validation playbooks and incident triage runbooks.

Required Qualifications:
7+ years in Software QA/Testing, with 2+ years in AI/ML or LLM-based systems; hands-on experience testing agentic/multi-agent architectures.
Strong programming skills in Python or TypeScript/JavaScript; experience building test harnesses, simulators, and fixtures.
Experience with LLM evaluation (exact/soft match, BLEU/ROUGE, BERTScore, semantic similarity via embeddings), guardrails, and prompt testing.
Expertise in distributed systems testing latency profiling, resiliency patterns (circuit breakers, retries), chaos engineering, and message queues.
Familiarity with orchestration frameworks (LangChain, LangGraph, LlamaIndex, DSPy, OpenAI Assistants/Actions, Azure OpenAI orchestration, or similar).
Proficiency with CI/CD (GitHub Actions/Azure DevOps), observability (OpenTelemetry, PrometheGrafana, Datadog), and feature flags/canaries.
Solid understanding of privacy/security/compliance in AI systems (PII handling, content policies, model safety).
Excellent communication and leadership skills; proven ability to work cross-functionally with Ops, Data, and Engineering.

Preferred Qualifications:
Experience with multi-agent simulators, agent graph testing, and tooling latency emulation.
Knowledge of MLOps (model versioning, datasets, evaluation pipelines) and A/B experimentation for LLMs.
Background in cloud (AWS), serverless, containerization, and event-driven architectures.
Prior ownership of cost/latency/SLAs for AI workloads in production.


Thanks & Regards
Mahesh

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 91159334
  • Position Id: 8904450
  • Posted 2 days ago

Company Info

About Solios Corp

Solios Corp. is a premier information technology company providing professional staffing solutions and project-based IT consulting services. We specialize in providing highly skilled professionals with current technological knowledge. Our goal is to pay attention to both human capital and technical skill sets so that together we can deliver the best products and services for our clients. We are M/WBE Partner.

We connect top minds in various areas of technology:

✓ UI/UX development

✓ Salesforce Solutions

✓ Full-stack Web & Mobile Development

✓ QA automation

✓ DevOps Tools

Our team has over 35 years of experience in the information technology industry. We utilize cutting-edge software engineering practices coupled with an agile methodology to achieve customer satisfaction by providing the best solution—whenever and wherever it is needed.

Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

It looks like there aren't any Similar Jobs for this job yet.

Search all similar jobs