Location: Austin, TX (5x/ week onsite)- LOCALS ONLY
Duration: 6 months
Data Engineer
Submission must have LinkedIn profile, Preference will be for candidates local to Austin, TX
MUST HAVE:
- Hands-on experience with writing Complex queries using Joins, Self Joins, Views, Materialized Views, Cursor also Recursive, use of GROUP BY, PARTITION BY functions / SQL Performance tuning
- Hands-on experience with ETL and Dimensional Data Modelling Slowly Changing Dimensions (SCD Type 1, 2, 3)
o Good understanding of concepts like schema types, table types - fact-dimension etc. like how to design a dimension vs fact, design considerations factored etc.
- Proficiency in Python scripting/programming using Pandas, PyParsing, Airflow.
o Pandas, Tableau server modules, Numpy, Datetime, Apache Airflow related modules, APIs
o Setting up Python scripts on DataLab, scheduling processes, connecting with DataLake (S3 etc )
o Data Pipeline automation
- Good understanding on Snowflake Architecture - experience with designing and building solutions.
o Architecture, design aspects, performance tuning, time travel, warehouse concepts - scaling, clustering, micro-partitioning
o Experience with SnowSQL, Snowpipe
- Must Have - Experience with Snowflake performance optimization techniques
- Own project delivery collaborating with Offshore.
- Actively participating in discussions with business to understand requirements and provide suitable solutions.
- Experience with AI (very beneficial) and advanced AI integration
o Good experience with Gen AI and LLM Integration which includes:
Having good understanding of RAG
Prompt and Context Engineering structure, query and manage data context fed to LLMs
Vector Data Management handling and storing data (also unstructured) in vector databases, indices for semantic search and RAG
Experience with LLM Orchestration frameworks LangChain, LlamaIndex