Overview
Remote
$65
Full Time
Accepts corp to corp applications
Contract - 27 day((s))
Skills
ML Ops & AI Inference Kubernetes / OpenShift CI/CD & Automation (Git
GitHub Actions
Jenkins
Terraform) Cloud (AWS / GCP / Azure) Python (Expert) LLMOps DevOps & Monitoring
Job Details
Welcome to Tektrnd, where innovation meets excellence. We are a dynamic and growing team dedicated to provide solutions that enhance operational performance for businesses and government agencies. At Tektrnd, we believe in fostering a collaborative and inclusive environment where creativity thrives and ideas come to life.
Why Choose Tektrnd?
- Innovative Culture: Be part of a forward-thinking company that encourages new ideas and embraces cutting-edge technologies.
- Career Growth: We invest in our employees' growth and development, offering opportunities for advancement and learning through mentorship programs and training initiatives.
- Impactful Work: Join us in making a difference in [industry/sector] by contributing to projects that matter and solutions that shape the future.
Why Work With Us?
At Tektrnd, we value our employees and offer competitive benefits, including:
- Comprehensive health and wellness plans
- Retirement savings options
- Flexible work schedules
- Employee discounts and perks
- Company-sponsored events and social activities
Current Opportunities
Explore exciting career opportunities at Tektrnd. Whether you're an experienced professional or just starting your career, we have a place for you:
Role: Principal ML Ops Engineer, AI Inference
Location: San Francisco, CA ( Remote )
Duration: Long Term Contract
Visa : USC
Visa : USC
We are seeking an experienced ML Ops engineer to lead the architecture and implementation of scalable, production-grade AI inference solutions. You will work closely with our product and research teams to scale SOTA deep learning products and software, focusing on building and releasing high-performance AI runtimes.
Responsibilities:
Architect and manage scalable model training and deployment pipelines for enterprise clients.
Lead the strategy for managing and releasing upstream and midstream AI product builds.
Design and implement automated testing frameworks to ensure model correctness, responsiveness, and efficiency.
Troubleshoot, debug, and upgrade mission-critical Dev & Test pipelines.
Define and deploy cybersecurity measures, including continuous vulnerability assessment and risk management for AI systems.
Collaborate with cross-functional teams to define market requirements and establish best practices for LLMOps.
Stay at the forefront of AI technologies and standards, driving innovation within our practice.
Qualifications:
5+ years of experience in ML Ops, DevOps, and Automation, with a focus on enterprise software deployment.
Expertise with Git, Github Actions, Terraform, Jenkins, Ansible, and modern automation/monitoring technologies.
Extensive experience administering Kubernetes/OpenShift in production environments.
Deep understanding of Agile development methodologies.
Proven experience with at least one major cloud provider: AWS, Google Cloud Platform, Azure, or IBM Cloud.
Expert-level Python programming skills.
Advanced troubleshooting and systems-thinking skills.
Experience contributing to open-source AI/ML projects (e.g., vLLM) is a strong plus.
Bachelor's degree or higher in Computer Science o
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.