Remote: Sr. ETL Developer (Talend Big Data Platform)

  • Posted 2 hours ago | Updated 2 hours ago

Overview

Remote
Depends on Experience
Contract - Independent
Contract - W2
Contract - 3 Month(s)
No Travel Required
Able to Provide Sponsorship

Skills

ETL
Talend
Big Data
MDM
Advanced SQL

Job Details

Role Overview

We are seeking a seasoned Senior ETL Developer with specialized expertise in the Talend Real-Time Big Data Platform (v6.4.x or higher). You will be responsible for designing, developing, and maintaining complex data integration pipelines that bridge legacy systems with modern big data environments. The ideal candidate is not just a "tool user" but an architect who understands how Talend generates code, manages memory, and interacts with Apache Spark and Hadoop clusters.

Key Responsibilities

Pipeline Development: Design and develop high-volume ETL/ELT jobs using Talend Studio, focusing on Big Data components (tSparkConfiguration, tMap, tHDFSOutput, etc.).

Architecture & Optimization: Optimize Talend job performance by analyzing generated Java code and tuning Spark/MapReduce parameters.

Data Management: Implement data quality rules, metadata management, and Master Data Management (MDM) workflows within the Talend ecosystem.

Advanced SQL: Write and optimize complex SQL queries for relational databases (Oracle, PostgreSQL, Snowflake) and Big Data environments (Hive, Impala).

Real-Time Integration: Build and support real-time streaming interfaces using Talend and Apache Kafka.

Collaboration: Act as a technical bridge between Data Architects and Business Stakeholders, translating complex requirements into scalable technical designs.

Technical Requirements

Talend Expertise: 5+ years of experience with Talend Real-Time Big Data. Must be proficient in Talend Administration Center (TAC) for job scheduling, monitoring, and project management.

Big Data Ecosystem: Hands-on experience with the Hadoop stack (HDFS, Hive, YARN) and Apache Spark (Spark Streaming and MLlib).

Core ETL/SQL: Expert-level understanding of ETL patterns (Slowly Changing Dimensions, Change Data Capture, Error Handling) and advanced SQL (Window functions, CTEs, Query Tuning).

Programming: Strong proficiency in Java (as Talend is Java-based) for creating custom routines and components.

Data Modeling: Solid understanding of Star/Snowflake schemas and Data Vault modeling.

Soft Skills

Communication: Ability to explain technical debt and architectural trade-offs to non-technical leadership.

Problem Solving: A "forensic" approach to debugging—identifying whether a failure is in the Talend logic, the Java runtime, or the Hadoop cluster.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.