Senior Data Engineer

Overview

Hybrid
Depends on Experience
Contract - W2

Skills

Agile Development
Cloudera
Data Integration
Data Warehouse
ETL
Hadoop
OLAP
NoSQL
shell scripting
Impala
Change Data Capture
Data Capture
Linux
Java
Unix

Job Details

Location: Reston VA. Must go to office 1 time a week

Duration: 12

The Senior Data Engineer is responsible for orchestrating, deploying, maintaining, and scaling cloud OR on-premise infrastructure targeting big data and platform data management (Relational and NoSQL, distributed and converged) with emphasis on reliability, automation, and performance. This role will focus on developing solutions and helping transform the company's platforms to deliver data-driven, meaningful insights and value to the company.

ESSENTIAL FUNCTIONS:

1) Work with Business Analysts and Product team to gather data requirements

2) Design and Build Ab Initio data graphs and data pipelines to extract the data from various databases/flat files/message queues

3) Transform the data to create a consumable data layer for various application uses

4) Support Data pipeline with bug fixes, and additional enhancements

5) Document Technical design, Operational Runbook, etc.

Qualifications

To perform this job successfully, an individual must be able to perform each essential duty satisfactorily. The requirements listed below are representative of the knowledge, skill, and/or ability required. Reasonable

accommodations may be made to enable individuals with disabilities to perform the essential functions.

Education Level: Bachelor's Degree

Education Details: Computer Science, Information Technology or Engineering or related field

Experience:

Total of 7+ Years of IT Experience predominantly in Data Integration/ Data Warehouse area

Must have at least 3 years of ETL Design and Development experience using Ab Initio

1-2 years of Data Integration project experience on Hadoop Platform, preferably Cloudera

At least one project implementation of Ab Initio CDC ( Change Data Capture ) in a Data Integration/ETL project

Working knowledge of HDFS, Hive, Impala, and other related Hadoop technologies

Sound understanding of SQL and ability to write well-performing SQL queries

Good knowledge of OLTP and OLAP data models and other data warehouse fundamentals

Rigor in high code quality, automated testing, and other engineering best practices, ability to write reusable code components

Ability to unit test the code thoroughly and troubleshoot issues in production environments

Must have some working experience with Unix/Linux shell scripting

Must be able to work independently and support other junior developers as needed

Some Java development experience is nice to have

Knowledge of Agile Development practices is required