LLM Research Engineer IV

Overview

On Site
$121 - $121
Contract - W2

Skills

A/B Testing
Artificial Intelligence
BERT
CUDA
Cloud Computing
Collaboration
Computer Science
Continuous Delivery
Continuous Integration
Data Quality
Data Science
Deep Learning
Distributed Computing
Docker
Evaluation
Generative Artificial Intelligence (AI)
Knowledge Sharing
Kubernetes
Language Models
Large Language Models (LLMs)
GPU
JAX
Machine Learning (ML)
Machine Learning Operations (ML Ops)
Natural Language Processing
Open Source
Patents
PyTorch
Real-time
Research
Scalability
TensorFlow
Training
Transformer

Job Details

Duties: Key Responsibilities:

Design, train, and fine-tune large language models (e.g., GPT, LLaMA, PaLM) for various applications. Conduct research on cutting-edge techniques in natural language processing (NLP) and machine learning to improve model performance. Explore advancements in transformer architectures, multi-modal models, and emergent AI behaviors. Collect, clean, and preprocess large-scale text datasets from diverse sources. Develop and implement data augmentation techniques to improve training data quality. Ensure data is free from bias and aligned with ethical AI standards. Optimize model architecture to improve accuracy, efficiency, and scalability. Implement techniques to reduce latency, memory footprint, and inference time for real-time applications. Collaborate with MLOps teams to deploy LLMs into production environments using Docker, Kubernetes, and cloud Develop robust evaluation pipelines to measure model performance using key metrics like accuracy, perplexity, BLEU, and F1 score. Continuously test for bias, fairness, and robustness of language models across diverse datasets. Conduct A/B testing to evaluate model improvements in real-world applications. Stay updated with the latest advancements in generative AI, transformers, and NLP research. Contribute to research papers, patents, and open-source projects. Present findings and insights at conferences and internal knowledge-sharing sessions. Qualifications: Advanced degree in Computer Science, Artificial Intelligence, Data Science, or a related field. Strong programming skills. Proficiency with deep learning frameworks such as TensorFlow, PyTorch, or JAX. Hands-on experience with transformer-based models (e.g., GPT, BERT, RoBERTa, LLaMA). Expertise in natural language processing (NLP) and sequence-to-sequence models. Familiarity with Hugging Face libraries and OpenAI APIs. Experience with MLOps tools like Docker, Kubernetes, and CI/CD pipelines. Strong understanding of distributed computing and GPU acceleration using CUDA. Knowledge of reinforcement learning and RLHF (Reinforcement Learning with Human Feedback).

Skills: Recommend 7-10 years experience

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.