Overview
On Site
Full Time
Skills
SQL
Spark
Python
AWS
Data Archiect
Job Details
Job Title: Data Architect
Location: Santa Clara, CA (Onsite)
Fulltime/Permanent
Job Description:
What you'll be doing:
- Build data pipelines to transport data from a data source to the data lake.
- Build data systems and pipelines ensuring that data sources, ingestion components, validation functions, transformation functions, and destination are well understood for implementation.
- Develop and implement the new End-to-End Data systems for our Planning, Logistics and Services initiatives.
- Prepare data for prescriptive and predictive modeling by making sure that the data is complete, has been cleansed, and has the necessary rules in place.
- Analyze and organize raw operational data including structured and unstructured data.
- Lead discussions with stakeholders and IT to identify and implement the right data strategy given data sources, data locations, and use cases.
- Interpret trends and patterns by performing complex data analysis.
- Build/develop algorithms, prototypes, and analytical tools that enable the Ops teams to make critical business decisions.
What we need to see:
- Master's or Bachelor's degree in Computer Science or Information System, or equivalent experience.
- 8+ years of relevant experience including programming knowledge (i.e Python, SQL).
- 5+ years of relevant experience in big data technologies and cloud platforms (i.e Spark, AWS).
- 5+ years of relevant experience in data lake technologies (i.e Iceberg, Delta, Huidi).
- 5+ years of experience in development best practices like CICD, Unit testing, Integration testing
- 5+ years of experience grabbing data from source systems like REST APIs, other databases using JDBC, ODBC, SFTP servers.
- Differentiating skill sets:
- 2+ years of experience with kubernetes and docker
- Experience in developing required infrastructure for optimal extraction, transformation, and loading of data from various sources using AWS, Azure, SQL or other technologies.
- Experience architecting, designing, developing, and maintaining data warehouses/data lakes for complex data ecosystems.
- Experience working with large datasets, databases and the software used to analyze the data.
- Expert in data and database management including data pipeline responsibilities in replication and mass ingestion, streaming, API and application and data integration.
- Strong analytical skills with the ability to collect, organize, and disseminate significant amounts of information with attention to detail and accuracy.
- Highly independent, able to lead key technical decisions, influence project roadmap and work effectively with team members.
Ways to stand out from the crowd:
- Knowledge in operational processes in chips, boards, systems, and servers with a view of data landscape.
- Self-starter, positive mindset with integrity and accountability, highly motivated, driven, and high-reaching.
- Solid ability to drive continuous improvement of systems and processes.
- A consistent record to work in a fast-paced environment where good interpersonal skills are crucial.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.