Job Title: Data Engineer – AI & Agentic Systems
Location: Sunnyvale, CA (Onsite)
Duration: 12+ Month
Primary skills - Must have: Snowflake, SQL
Work experience - AI integration, Agentic AI
Secondary: Tableau
About the Role
Apple is seeking a highly skilled Data Engineer with strong AI integration and Agentic AI experience to help build next‑generation intelligent data systems. This role combines modern data engineering (Snowflake, SQL) with AI/LLM orchestration, Agentic workflows, and scalable distributed systems, contributing to Apple’s mission to deliver transformative experiences across products and services.
You will work closely with cross‑functional AI/ML teams to architect, build, and optimize data infrastructure supporting LLM‑driven agentic automation, multi‑agent orchestration, and high‑volume data workflows powering intelligence across Apple platforms.
Key Responsibilities
- Design, build, and optimize large‑scale distributed data pipelines using Snowflake, SQL, and cloud‑based data frameworks.
- Integrate LLMs, RAG pipelines, Agentic workflows, and multi‑agent execution patterns into data systems.
- Build infrastructure supporting agentic orchestration, including context handling, memory persistence, vector search integration, and multi‑agent communication (A2A/MCP).
- Collaborate with AI, software engineering, and ML teams to deliver end‑to‑end AI‑powered data solutions.
- Ensure scalable, secure, highly available data systems operating at Apple‑scale.
- Support data ingestion, transformation, and evaluation frameworks powering foundation models and AI workflows.
- Build dashboards/insights using Tableau as needed for cross‑functional visibility. (Secondary Skill)
Minimum Qualifications
- Strong expertise in Snowflake & SQL, including data warehousing, distributed data processing, and performance optimization.
- Experience integrating AI/LLM systems, especially with Agentic frameworks (LangGraph, ADK, LangChain, LlamaIndex or similar).
- Hands‑on experience building LLM‑powered or agentic systems, including RAG pipelines and vector databases (FAISS, Milvus, Weaviate, Vespa, etc.).
- Strong programming experience in Python, with knowledge of distributed data systems and ML data flows.
- Understanding of large‑scale system architecture, data structures, concurrency, and multi‑threaded design.
- Experience processing data for ML applications at scale.
Preferred Qualifications
- Prior Apple experience or experience building systems at Apple‑scale.
- Experience working with Agent‑to‑Agent (A2A) protocols, context engineering, non‑deterministic loop handling, and agentic memory.
- Experience with vector search, feature stores, and ML data pipelines.
- Experience integrating analytics tools such as Tableau for visual storytelling.
- Knowledge of secure data handling, prompt‑injection defense, and privacy‑by‑design principles.
- Familiarity with cloud data ecosystems (Google Cloud Platform, AWS), Apache Beam, Kafka, Spark, or similar technologies.
Soft Skills
- Excellent communication and ability to work with multi‑disciplinary business, research, and engineering teams.
- Strong ownership mentality, creativity, and ability to thrive in fast‑paced environments.