Overview
Remote
Depends on Experience
Contract - Independent
Contract - W2
Contract - 12 Month(s)
No Travel Required
Skills
Amazon Web Services
Artificial Intelligence
Cloud Computing
Collaboration
Computer Science
Continuous Delivery
Continuous Integration
Data Science
Docker
FOCUS
Generative Artificial Intelligence (AI)
Good Clinical Practice
Google Cloud Platform
Hosting
Kubernetes
LangChain
Large Language Models (LLMs)
Machine Learning (ML)
Machine Learning Operations (ML Ops)
Mentorship
Microsoft Azure
Natural Language Processing
Open Source
Orchestration
Prompt Engineering
PyTorch
Python
Research
Scalability
Software Engineering
Technical Drafting
TensorFlow
Transformer
Vector Databases
RAG
Job Details
Role:-: Generative AI Architect
Location: Remote (100%)
Duration: Long-Term
Experience Level: 15+ Years
Job Summary:
We are seeking a highly experienced and innovative Generative AI Architect to lead the design, development, and deployment of AI/ML solutions with a focus on Generative AI technologies. The ideal candidate will have deep expertise in Large Language Models (LLMs), transformer-based architectures, and real-world experience building and scaling Gen AI systems.
Key Responsibilities:
- Architect and implement scalable Generative AI solutions using LLMs (e.g., GPT, PaLM, LLaMA, Claude).
- Collaborate with data scientists, ML engineers, and product teams to integrate Gen AI into applications and platforms.
- Design retrieval-augmented generation (RAG) pipelines and fine-tuning strategies for LLMs.
- Evaluate and optimize AI models for performance, scalability, and reliability in production environments.
- Define best practices, frameworks, and architectural blueprints for Gen AI solutions.
- Stay updated on the latest Gen AI research and tools; guide adoption of new technologies.
- Lead technical design discussions and mentor junior AI engineers.
Required Qualifications:
- 10+ years of experience in AI/ML architecture or software engineering.
- 3+ years of hands-on experience with Generative AI/LLMs.
- Proficiency in Python, TensorFlow, PyTorch, and Gen AI frameworks (e.g., LangChain, Hugging Face, Transformers).
- Strong understanding of NLP, prompt engineering, and model fine-tuning.
- Experience deploying AI models using cloud platforms (AWS, Azure, or Google Cloud Platform).
- Knowledge of containerization and orchestration (Docker, Kubernetes).
- Familiarity with vector databases (e.g., FAISS, Pinecone, Weaviate).
Preferred Qualifications:
- Master's or PhD in Computer Science, Machine Learning, Data Science, or a related field.
- Experience with RAG, embeddings, and multimodal models (text + images).
- Exposure to MLOps and CI/CD for ML pipelines.
- Experience working with open-source LLMs or private model hosting.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.