AI QA ENGINEER (AGENTIC & GENERATIVE)
Dallas, TX, US • Posted 17 hours ago • Updated 17 hours ago
Staffingine LLC
Dice Job Match Score™
🔗 Matching skills to job...
Job Details
Skills
- workflows
- triage
- continuous integration
- INCIDENT RESPONSE
- TYPESCRIPT
- MICROSOFT AZURE
- GITHUB
- PROMETHEUS
- QUALITY MANAGEMENT
- DATA SCIENCE
- Leadership
- Communication Skills
- AMAZON WEB SERVICES
- SIMULATIONS
- SYSTEMS DEVELOPMENT LIFE CYCLE
- NETWORK PERFORMANCE
- circuit breakers
- Reliability
- Data Pipelines
- Distributed Systems
- Telemetry
- Artificial Intelligence
- Computer Programming
- Grafana
- JavaScript (Programming Language)
- Region Management
- Testing Skills
- Serverless Computing
- Application Programming Interfaces (APIs)
- Safety Principles
- Budgeting Skills
- Testing (Software)
- Machine Learning Operations
- Large Language Models
- Multi-Agent Systems
- Containerisation
- Perseverance
- Semantics
- Management of Software Versions
- Datadog
- Execution of Experiments
- Queueing Systems
- Concurrency
- Failover
- Quality Strategy
Summary
Job Title: AI QA ENGINEER (AGENTIC & GENERATIVE)
Job Location: Dallas, TX
Job Type: Contract
Job Description:
- Quality Strategy & Leadership
- Agentic & Multi Agent Testing
- Reliability, Resiliency, and Latency
- Accuracy & Macro-Level Validations
- Scale & Orchestration
- Dev Prod Readiness
- Define and own the QA strategy for agentic/multi-agent AI systems across dev, staging, and prod.
- Mentor a team of QA engineers; establish testing standards, coding guidelines for test harnesses, and review practices.
- Partner with Agentic Operations, Data Science, MLOps, and Platform teams to embed QA in the SDLC and incident response.
- Design tests for agent orchestration, tool calling, planner-executor loops, and inter-agent coordination (e.g., task decomposition, handoff integrity, and convergence to goals).
- Validate state management, context windows, memory/knowledge stores, and prompt/graph correctness under varying conditions.
- Implement scenario fuzzing (e.g., adversarial inputs, prompt perturbations, tool latency spikes, degraded APIs).
- Create resilience testing suites: chaos experiments, failover, retries/backoff, circuit-breaking, and degraded mode behavior.
- Establish latency SLOs and measure end-to-end response times across orchestration layers (LLM calls, tool invocations, queues).
- Ensure reliability through soak tests, canary verifications, and automated rollbacks.
- Define ground-truth and reference pipelines for task accuracy (exact match, semantic similarity, factuality checks).
- Build macro validation frameworks that validate task outcomes across multi-step agent workflows (e.g., complex data pipelines, content generation + verification agent loops).
- Instrument guardrail validations (toxicity, PII, hallucination, policy compliance).
- Design load/stress tests for multi-agent graphs under scale (concurrency, throughput, queue depth, backpressure).
- Validate orchestrator correctness (DAG execution, retries, branching, timeouts, compensation paths).
- Engineer reusable test artifacts (scenario configs, synthetic datasets, prompt libraries, agent graph fixtures, simulators).
- Integrate tests into CI/CD (pre-merge gates, nightly, canary) and production monitoring with alerting tied to KPIs.
- Define release criteria and run operational readiness (performance, security, compliance, cost/latency budgets).
- Build post-deployment validation playbooks and incident triage runbooks.
Required Qualifications
- 7+ years in Software QA/Testing, with 2+ years in AI/ML or LLM-based systems; hands-on experience testing agentic/multi-agent architectures.
- Strong programming skills in Python or TypeScript/JavaScript; experience building test harnesses, simulators, and fixtures.
- Experience with LLM evaluation (exact/soft match, BLEU/ROUGE, BERTScore, semantic similarity via embeddings), guardrails, and prompt testing.
- Expertise in distributed systems testing latency profiling, resiliency patterns (circuit breakers, retries), chaos engineering, and message queues.
- Familiarity with orchestration frameworks (LangChain, LangGraph, LlamaIndex, DSPy, OpenAI Assistants/Actions, Azure OpenAI orchestration, or similar).
- Proficiency with CI/CD (GitHub Actions/Azure DevOps), observability (OpenTelemetry, PrometheGrafana, Datadog), and feature flags/canaries.
- Solid understanding of privacy/security/compliance in AI systems (PII handling, content policies, model safety).
- Excellent communication and leadership skills; proven ability to work cross-functionally with Ops, Data, and Engineering.
Preferred Qualifications
- Experience with multi-agent simulators, agent graph testing, and tooling latency emulation.
- Knowledge of MLOps (model versioning, datasets, evaluation pipelines) and A/B experimentation for LLMs.
- Background in cloud (AWS), serverless, containerization, and event-driven architectures.
- Prior ownership of cost/latency/SLAs for AI workloads in production.
- Dice Id: 91165639
- Position Id: 2026-16985
- Posted 17 hours ago
Company Info
Staffingine LLC specializes in enhancing organizational performance through the effective and competent application of technology and outsourcing solutions for our clients. Our objective is to not only solve short term business and technology needs, but to create next-generation of competitive advantages that drive future growth and success.
With the state of the art development center in Noida and associates working in India and US, we provide services to businesses globally. Our broad resource of associates ensures that our clients have easy access to resources that they need in their specific business and technical domain. Staffingine's resources, coupled with Staffingine Group widely spread channel & infrastructure, gives client unique scalability to meet the needs of projects of any size, at any location.
Similar Jobs
It looks like there aren't any Similar Jobs for this job yet.
Search all similar jobs