Stefanini Group is hiring!Stefanini is looking for Backend AI Engineer-HybridFor quick apply, please contact Akash Gupta; Ph: 248 728 Only! Key Responsibilities:Design, develop, and maintain robust backend services and APIs to support AI/ML applications across the organization.Actively participate in Agile rituals and follow Scaled Agile processes as set forth by the CDP Program team.Deliver high-quality backend services following Safe Agile PracticesProactively identify and resolve issues with AI services, APIs, and model serving infrastructure.Deploy comprehensive monitoring and alerting for backend AI systems, implementing auto-remediation where possible to ensure system availability and reliability.Employ a security-first, testing, and automation strategy, adhering to backend engineering and MLOps best practices.Collaborate with cross-functional teams, including data scientists, front-end engineers, data engineers, and business stakeholders, to understand requirements and deliver robust backend solutions.Keep up with the latest trends and technologies, evaluating and recommending new tools, frameworks, and architectures to improve backend AI capabilities. What You'll Bring:Backend API Development (50%):Design and develop robust, scalable RESTful APIs and GraphQL services for AI/ML applications.Build backend services for LLM-powered applications, including RAG systems, document processing pipelines, and knowledge bases.Implement secure API endpoints for model inference, prompt orchestration, and AI service integration.Develop asynchronous processing workflows for long-running AI tasks (document analysis, batch predictions).Create middleware and service layers to abstract complex AI functionality for front-end consumption.Design and implement caching strategies, rate limiting, and API versioning for production AI services.Build event-driven architectures using message queues (SQS, SNS, EventBridge) for scalable AI workflows.Ensure APIs meet security, authentication, and authorization requirements for Federal environments.Implement comprehensive error handling, logging, and observability for AI services. MLOps & Infrastructure (30%)Build and maintain ML model serving infrastructure using AWS SageMaker, Lambda, and containerized deployments.Integrate AWS AI services (Bedrock, Textract, Comprehend) into backend pipelines and APIs.Develop CI/CD pipelines for automated testing, deployment, and rollback of AI services.Implement model versioning, A/B testing frameworks, and canary deployment patterns.Create data preprocessing and feature engineering pipelines using PySpark and Databricks.Build orchestration workflows for multi-step AI processes (ingestion preprocessing inference post-processing).Develop monitoring and alerting systems for model performance, latency, cost, and availability.Collaborate with data scientists to produce ML models and transition from prototype to production.Implement vector databases and semantic search capabilities for RAG architecture.Manage prompt templates, model configurations, and AI service parameters as code. Performance & Support (20%)Optimize backend performance for high-throughput AI workloads and real-time inference.Monitor and troubleshoot production issues including latency, errors, and cost optimization.Implement automated remediation and self-healing capabilities for backend services.Conduct performance testing and capacity planning for AI infrastructureActively participate in Agile rituals and follow Scaled Agile processes as set forth by the CDP Program team.Provide technical support and act as escalation point for backend AI service issues.Create comprehensive technical documentation for APIs, architecture patterns, and deployment procedures.Stay current on backend technologies, AI infrastructure trends, and Federal regulatory requirements.Collaborate with security teams to ensure compliance with data protection and privacy regulations. #LI-AG#LI-HYBRID
Minimum Qualifications:Education: Bachelor's degree in Computer Science, Software Engineering, Information Systems, or related technical field, or equivalent experienceExperience: 4+ years in backend development, with at least 2+ years building and deploying AI/ML or LLM-powered services.Programming: Strong Python proficiency; experience with frameworks like FastAPI, Flask, or DjangoAPI Design: Proven experience designing and implementing RESTful APIs, with understanding of API security and best practicesLLM Integration: Hands-on experience building backend services for LLM applications, including prompt orchestration, RAG architectures, and AI service integrationCloud Platforms: Working knowledge of AWS services including Lambda, API Gateway, SageMaker, Bedrock, S3, and related AI/ML toolsDatabase Experience: Proficiency with both relational (PostgreSQL, MySQL) and NoSQL databases (DynamoDB, MongoDB); experience with vector databases (Pinecone, Weaviate, pgvector) preferredContainerization: Experience with Docker and container orchestration (ECS, EKS, or similar)ML Model Deployment: Demonstrated ability to deploy and serve ML models in production environmentsAsynchronous Programming: Experience with async/await patterns, message queues, and event-driven architecturesTesting: Strong understanding of unit testing, integration testing, and test automation practicesCommunication: Ability to collaborate effectively with cross-functional teams and translate business requirements into technical solutions Preferred Qualifications:3+ years' experience with PySpark and distributed computing frameworksExperience with Databricks, Collibra, and StarburstKnowledge of MLOps tools and frameworks (MLflow, Kubeflow, SageMaker Pipelines)Experience with Infrastructure as Code (Terraform, CloudFormation)Familiarity with streaming data platforms (Kafka, Kinesis)Experience building end-to-end data pipelines and ETL processesUnderstanding of microservices architecture and service mesh patternsExperience with observability tools (DataDog, New Relic, CloudWatch)Background working in regulated industries (financial services, healthcare, government)Knowledge of data governance, lineage, and compliance frameworksExperience with GraphQL API design and implementationFamiliarity with econometric models and statistical computing environments (R, Stata) Stefanini takes pride in hiring top talent and developing relationships with our future employees. Our talent acquisition teams will never make an offer of employment without having a phone conversation with you. Those face-to-face conversations will involve a description of the job for which you have applied. We also speak with you about the process including interviews and job offers. About Stefanini Group:The Stefanini Group is a global provider of offshore, onshore and near shore outsourcing, IT digital consulting, systems integration, application, and strategic staffing services to Fortune 1000 enterprises around the world. Our presence is in countries like the Americas, Europe, Africa, and Asia, and more than four hundred clients across a broad spectrum of markets, including financial services, manufacturing, telecommunications, chemical services, technology, public sector, and utilities. Stefanini is a CMM level 5, IT consulting company with a global presence. We are CMM Level 5 company
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
- Dice Id: 10106616
- Position Id: 63271
- Posted 4 hours ago