LLM Research Engineer

  • Mountain View, CA
  • Posted 7 hours ago | Updated 7 hours ago

Overview

On Site
$90 - $140
Accepts corp to corp applications
Contract - W2
Contract - Independent
Contract - 1 Year(s)
100% Travel

Skills

NLP
Pytorch
generative AI
MLOPS
Artificial Intelligence
A/B Testing
Generative Artificial Intelligence (AI)
TensorFlow
Machine Learning Operations (ML Ops)
Large Language Models (LLMs)
Natural Language Processing
Data Science
Deep Learning
Computer Science
Kubernetes
Machine Learning (ML)
Research
Language Models
Docker

Job Details

Key Responsibilities: Design, train, and fine-tune large language models (e.g., GPT, LLaMA, PaLM) for various applications. Conduct research on cutting-edge techniques in natural language processing (NLP) and machine learning to improve model performance. Explore advancements in transformer architectures, multi-modal models, and emergent AI behaviors. Collect, clean, and preprocess large-scale text datasets from diverse sources. Develop and implement data augmentation techniques to improve training data quality. Ensure data is free from bias and aligned with ethical AI standards. Optimize model architecture to improve accuracy, efficiency, and scalability. Implement techniques to reduce latency, memory footprint, and inference time for real-time applications. Collaborate with MLOps teams to deploy LLMs into production environments using Docker, Kubernetes, and cloud Develop robust evaluation pipelines to measure model performance using key metrics like accuracy, perplexity, BLEU, and F1 score. Continuously test for bias, fairness, and robustness of language models across diverse datasets. Conduct A/B testing to evaluate model improvements in real-world applications. Stay updated with the latest advancements in generative AI, transformers, and NLP research. Contribute to research papers, patents, and open-source projects. Present findings and insights at conferences and internal knowledge-sharing sessions. Qualifications: Advanced degree in Computer Science, Artificial Intelligence, Data Science, or a related field. Strong programming skills. Proficiency with deep learning frameworks such as TensorFlow, PyTorch, or JAX. Hands-on experience with transformer-based models (e.g., GPT, BERT, RoBERTa, LLaMA). Expertise in natural language processing (NLP) and sequence-to-sequence models. Familiarity with Hugging Face libraries and OpenAI APIs. Experience with MLOps tools like Docker, Kubernetes, and CI/CD pipelines. Strong understanding of distributed computing and GPU acceleration using CUDA. Knowledge of reinforcement learning and RLHF (Reinforcement Learning with Human Feedback).

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.