Data Engineer at Dallas, TX
3 days/week in client office – Only locals
Candidate should have past Palantir client exp.
JD-
Data Integration & ETL: Build and manage scalable data pipelines to ingest data from diverse sources (ERP, CRM, APIs, S3, SQL databases) into Foundry.
• Ontology Modeling: Define and maintain the "Ontology"—the platform’s semantic layer—which maps technical data to real-world business objects (e.g., "Aircraft," "Customer," or "Invoice").
• Pipeline Development: Write and optimize data transformations using PySpark, SQL, or Java within Foundry's Code Repositories.
• Application Building: Develop front-end operational applications and interactive dashboards using low-code/pro-code tools like Workshop and Slate.
• AIP Integration: Implement Artificial Intelligence Platform (AIP) features, such as LLM-backed functions and agents, to automate workflows.
• Data Governance & Security: Configure granular access controls, data health monitors, and lineage tracking to ensure compliance and reliability.
## Core Technical Skills
• Languages: High proficiency in Python (PySpark) and SQL is mandatory. Knowledge of Java, TypeScript, or JavaScript is often required for front-end customization.
• Big Data: Understanding of distributed computing (Spark), data warehousing concepts, and schema design (Star, Snowflake, etc.).
• DevOps: Experience with Git-based version control, CI/CD practices, and debugging complex data workflows.
• Cloud Architecture: Familiarity with AWS, Azure, or Google Cloud Platform environments where Foundry is typically hosted.