DivIHN (pronounced “divine”) is a CMMI ML3-certified Technology and Talent solutions firm. Driven by a unique Purpose, Culture, and Value Delivery Model, we enable meaningful connections between talented professionals and forward-thinking organizations. Since our formation in 2002, organizations across commercial and public sectors have been trusting us to help build their teams with exceptional temporary and permanent talent.
Visit us at to learn more and view our open positions.
Please apply or call one of us to learn more
For further inquiries regarding the following opportunity, please contact our Talent Specialist, Deepak at
Title: Junior Data Engineer (ETL)
Location: On-site at Corning, NY
Duration: 12 Months
Work Schedule: Typical 40 hours per week. May require working weekends/holidays or longer days to support projects.
Travel: Limited to no travel required, and no on-call requirements.
The manager is open to non-local candidates willing to relocate at their own expenses
Only W2 candidates are eligible for this position. Third-party or C2C candidates will not be considered
Description:
Education and Experience:
This position focuses on Data pipelines & workflows
• Bachelor’s degree in computer science, information systems, data engineering, or related field, or equivalent practical experience. May consider an Associate if the candidate has an additional 3-5 years of experience than what is required.
• 2+ years of professional experience in data engineering, ETL development, or related work, or equivalent hands-on experience
• Experience or interest in scientific software, materials science, research environments, or technically complex domains is a plus
Scope of Position:
1. Embed within a cross-functional Agile team, participating in sprint planning, stand-ups, backlog refinement, and technical discussions.
2. Design, build, troubleshoot, and maintain ETL/ELT workflows that support application functionality, analytics, reporting, and scientific workflows.
3. Develop and manage data pipelines using Apache Airflow, ensuring reliable orchestration, scheduling, monitoring, and recovery of data processes.
4. Work with stakeholders including software developers, scientists, and engineers, to understand data sources, workflow requirements, and downstream data needs.
5. Extract, transform, validate, and load data across systems, including relational databases such as Postgres SQL and Oracle.
6. Write, optimize, and maintain complex SQL queries, scripts, and transformation logic to support operational and analytical use cases.
7. Troubleshoot data quality issues, ETL failures, pipeline bottlenecks, and schema inconsistencies; identify root causes and implement durable solutions.
8. Support database exploration, data validation, and troubleshooting using tools such as DBeaver and related database utilities.
9. Evaluate and help adopt new data tools and technologies, including lightweight analytics and transformation solutions (e.g. DuckDB) where appropriate.
10. Collaborate with engineering teams to support reliable integration between data pipelines, applications, APIs, and downstream consumers.
11. Assist with schema evolution, data modeling, migration planning, and data consistency across systems.
12. Document pipeline logic, data dependencies, transformation rules, and operational procedures to support maintainability and team knowledge sharing.
13. Help improve data engineering standards, observability, testing practices, and operational reliability across the team.
14. Regularly interact with scientists and engineers to understand research and technical workflows; experience in scientific or research environments is a strong plus.
Technical Skills – 2+ years (or Commensurate Experience):
1. Experience designing, building, and troubleshooting ETL/ELT pipelines
2. Hands-on experience with workflow orchestration tools, preferably Apache Airflow
3. Strong experience writing and optimizing SQL
4. Experience working with relational databases, especially PostgreSQL and Oracle
5. Ability to develop and maintain data transformations, validation steps, and pipeline logic across multiple systems
6. Experience with database tools such as DBeaver or similar for query development, exploration, and troubleshooting
7. Familiarity with modern data processing and analytical tools such as DuckDB or interest in evaluating emerging data technologies
8. Understanding of data modeling, schema design, data integrity, and performance tuning
9. Experience troubleshooting pipeline failures, performance issues, and inconsistent or incomplete datasets
10. Familiarity with scripting or programming for pipeline development and automation; Python experience is strongly preferred
11. Understanding of version control and collaborative development workflows
12. Experience supporting production data systems with an emphasis on reliability, maintainability, and clear documentation
Team Skills:
1. Confident collaborating with developers, scientists, analysts, and product stakeholders
2. Able to gather and clarify technical and data requirements and translate them into scalable data solutions
3. Strong communication skills around pipeline status, data quality issues, dependencies, and tradeoffs
4. Comfortable handling ambiguity, improving incomplete processes, and helping define best practices
5. Proactive in identifying opportunities to improve data workflows, tooling, performance, and operational stability
Soft Skills:
1. Strong analytical and problem-solving skills
2. High attention to detail and commitment to data quality, consistency, and reliability
3. Demonstrated initiative in troubleshooting issues and improving pipeline robustness
4. Curiosity and willingness to evaluate and adopt new tools, technologies, and approaches
5. Ability to balance immediate operational needs with long-term maintainability and scalability
6. Comfortable proposing improvements, collaborating across teams, and building trust through reliable execution
Interview Process: Phone screen, then either an onsite interview for local candidates or a Teams Meeting for non-local candidates
About us:
DivIHN, the ''IT Asset Performance Services'' organization, provides Professional Consulting, Custom Projects, and Professional Resource Augmentation services to clients in the Mid-West and beyond. The strategic characteristics of the organization are Standardization, Specialization, and Collaboration.
DivIHN is an equal opportunity employer. DivIHN does not and shall not discriminate against any employee or qualified applicant on the basis of race, color, religion (creed), gender, gender expression, age, national origin (ancestry), disability, marital status, sexual orientation, or military status.