Overview
Skills
Job Details
Role: Senior AI/ML Engineer
Location: Bay Area, CA
About the Role
We are seeking a Senior AI/ML Engineer to design, build, and deploy large-scale, production-grade machine learning systems powering our Personal AI & Human Connection intelligence platform. You will architect and implement end-to-end ML pipelines, from data ingestion and feature engineering to model training, deployment, and monitoring ensuring reliability, scalability, and compliance with Personal data standards.
The ideal candidate combines strong software engineering fundamentals with deep experience in machine learning, MLOps, and cloud-native systems (AWS/Google Cloud Platform). You ll work closely with product, data, and executive teams to translate real-world challenges into impactful AI solutions for our users.
Responsibilities
Design and implement end-to-end ML pipelines data ingestion, preprocessing, model training, validation, and deployment.
Develop and productionize ML models for prediction, classification, and recommendation tasks across complex workflows.
Build automated retraining and evaluation pipelines using tools such as AWS SageMaker, Vertex AI, Kubeflow, or MLflow.
Develop and maintain feature stores and data transformation pipelines using Spark, PyTorch, or TensorFlow.
Deploy and serve models via REST/gRPC endpoints or Lambda-based inference APIs with high availability and low latency.
Collaborate with backend engineers to integrate models into production microservices and event-driven workflows.
Implement model observability and drift detection, logging key metrics (accuracy, latency, bias, cost).
Ensure all pipelines and models meet security, privacy, and compliance standards (HIPAA, SOC 2).
Write clean, testable code; create unit, integration, and performance tests for all critical ML components.
Partner with product and data science teams to refine model goals, define success metrics, and optimize inference performance.
Qualifications
Bachelor s or Master s degree in Computer Science, Machine Learning, or a related field.
6+ years of hands-on experience building, training, and deploying ML models in production.
Expert in Python and ML libraries/frameworks: PyTorch, TensorFlow, scikit-learn.
Strong experience with AWS AI/ML stack (S3, SageMaker, Lambda, Step Functions, ECS, DynamoDB, RDS).
Proficiency in data pipelines (Airflow, Spark, Glue) and feature engineering for large, structured/unstructured datasets.
Familiarity with Docker/Kubernetes and deploying ML services via CI/CD.
Deep understanding of ML lifecycle management and versioning (MLflow, DVC, Weights & Biases).
Knowledge of prompt engineering, RAG pipelines, and LLM integration is a plus.
Experience handling sensitive data (PII), with strong focus on security, encryption, and compliance.
Excellent debugging and optimization skills across model performance, latency, and infrastructure cost.
Preferred Skills
Experience building LLM applications using OpenAI, Anthropic, or Hugging Face APIs.
Experience with vector databases (Pinecone, FAISS, or Weaviate) and retrieval-augmented generation (RAG).
Familiarity with MLOps tools like SageMaker Pipelines, Vertex AI Pipelines, or Kubeflow.
Working knowledge of data governance, ethical AI, and model explainability (SHAP, LIME)