ML ETL Architect

Overview

On Site
Hybrid
Depends on Experience
Accepts corp to corp applications
Contract - W2
Contract - Independent

Skills

MFG
SQL
Data Warehouse
BI Testing
Python
Snowflake
ETL
BigQuery
Databricks
Apache
Airflow
AWS
MWAA

Job Details

ML Data Architect
Location: Culver City, CA
Duration: 6 months
___________________________________________________________________________

Seeking a talented Machine Learning / ETL Engineer to join our IT team and take on the challenge of advancing our Business Intelligence initiatives. In this role, you will be responsible for designing, developing and optimizing ETL / ELT processes for large-scale data processing and leveraging Machine Learning frameworks to perform data quality and anomaly detection across 1st and 3rd party datasets.

Responsibilities-
Architecting and implementing robust, high-performance data pipelines across various data sources (databases, APIs, web services) to efficiently process and transform large datasets.
Enable and ensure data collection from various sources and loading it to a data lake, data warehouse, or any other form of storage.
Proficiently utilizing cloud platforms Snowflake, AWS, Google Cloud Platform to build scalable data processing solutions, including managing data lakes and data warehouses.
Developing and implementing data quality checks and monitoring systems to ensure data accuracy and reliability throughout the data pipeline.
Contribute to data documentation and governance strategies, including data access controls, data lineage tracking, and data retention policies.

Qualifications:-
10+ years of experience with data warehouse technical architectures, ETL/ ELT, reporting/analytic tools, and scripting.
Extensive knowledge and understanding of data modeling, schema design, and data lakes.
5+ years of data modeling experience and proficiency in writing advanced SQL and query performance tuning with on Snowflake in addition to Oracle, and Columnar Databases SQL optimization experience.
Experience with AWS services including S3, Lambda, Data-pipeline, and other data technologies.
Experience implementing Machine Learning algorithms for data quality, anomaly detection and continuous monitoring, etcs.
Required skills and experience:
Strong proficiency in Python, SQL for data manipulation and processing.
Experience with data warehouse solutions for Snowflake, BigQuery, Databricks.
Ability to design and implement efficient data models for data lakes and warehouses.
Familiarity with CI/CD pipelines and automation tools to streamline data engineering workflows
Deep understanding of principles in data warehousing and cloud architecture for building very efficient and scalable data systems.
Experience with Apache Airflow and/or AWS MWAA

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.