SME Data Engineer

  • Posted 50 days ago | Updated 14 hours ago

Overview

Remote
Depends on Experience
Full Time
No Travel Required

Skills

Oracle PL/SQL
MongoDB
Cassandra
MYSQL
ETL
S3
Redshift
EMR
Scala
Cloud
AWS

Job Details

Become an integral part of a diverse team while working at an Industry Leading Organization, where our employees come first.  At ADTECH, you’ll help protect our national security while working on innovative projects that offer opportunities for advancement.

 

Currently, ADTECH is seeking a motivated, career and customer-oriented SME Data Engineer to join our team.

 

Each day U.S. Customs and Border Protection (CBP) oversees the massive flow of people, capital, and products that enter and depart the United States via air, land, sea, and cyberspace.  The volume and complexity of both physical and virtual border crossings require the application of solutions to promote efficient trade and travel.  Further, effective solutions help CBP ensure the movement of people, capital, and products is legal, safe, and secure.  In response to this challenge, ADTECH, as a trusted mission partner of CBP seeks capable, qualified, and versatile SME Data Engineers to help develop complex data analytical solutions for law enforcement personnel to assess risk of potential threats entering the country.

 

SME Data Engineer

Remote

Full time position

 

Responsibilities include, but are not limited to:

  • Design, develop, and maintain scalable data pipelines and architectures to support data extraction, transformation, and loading (ETL/ELT) processes. Utilize strong SQL skills to perform complex data transformations and optimize database queries, ensuring high performance and efficiency.
  • Building comprehensive datasets by aggregating data sourced from various relational databases, facilitating data analysts and data scientists in creating machine learning models, reports, and dashboards.
  • Collaborate with cross-functional teams (data analysts, data scientists, and business stakeholders) to understand business requirements and translate them into technical solutions.
  • Assist with the implementation of data migration/pipelines from on-prem to cloud/non-relational storage platforms.
  • Leverage distributed computing frameworks like Apache Spark to process large volume of data efficiently.
  • Utilizing data analysis, problem-solving, investigative, and creative thinking skills to handle extremely large datasets, transforming them into various formats for diverse analytical products.
  • Respond to data queries/analysis requests from various groups within an organization. Create and publish regularly scheduled and/or ad hoc reports as needed.
  • Troubleshoot data-related issues, identify root causes, and implement solutions to ensure data integrity and accuracy.
  • implement best practices for data governance, security, and quality supporting the core business applications.
  • Responsible for data engineering source code control using GitLab.

Basic Qualifications:

  • Experience with relational databases and knowledge of query tools and/or BI tools like Power BI or OBIEE and data analysis tools
  • Extensive experience with SQL and proficiency in writing complex queries.
  • Solid understanding of data warehousing concepts and platforms such as Oracle and cloud-based solutions.
  • Strong experience in automating ETL jobs via UNIX/LINUX shell scripts and CRON jobs.
  • Demonstrate a strong practical understanding of data warehousing from a production relational database environment.
  • Strong experience using analytic functions within Oracle or similar tools within non-relational (MongoDB, Cassandra etc.) database systems.
  • Strong understanding of distributed computing principles and experience with frameworks like Apache Spark
  • Hands-on-experience with data lake architectures and technologies in a cloud environment.
  • Experience with Atlassian suite of tools such as Jira and Confluence
  • Knowledge of Continuous Integration & Continuous Development tools (CI/CD)
  • Must be able to multitask efficiently and progressively and work comfortably in an ever-changing data environment.
  • Must work well in a team environment as well as independently.
  • Excellent verbal/written communication and problem solving skills; ability to communicate information to a variety of groups at different technical skill levels.

 

Preferred Qualifications:

  • 5+ years of experience in developing, maintaining and optimizing complex Oracle PL/SQL packages to aggregate transactional data for consumption by data science/machine learning applications.
  • 10+ years of experience in working in data engineering, with a focus on building and optimizing data pipelines and architectures. Must have full life cycle experience in design, development, deployment and monitoring.
  • Experience with one or more relational database systems such as Oracle, MySQL, Postgres, SQL server, with heavy emphasis on Oracle.
  • Extensive experience with cloud platforms (e.g. AWS, Google Cloud, etc) and cloud-based ETL/ELT tools.
  • Experience with Amazon services such as S3, Redshift, EMR and Scala.
  • Experience with migrating on-prem legacy database objects and data to the Amazon S3 cloud environment.
  • Experience or familiarity with data science/machine learning and development experience for supervised and unsupervised learning with structure and unstructured datasets.
  • Certifications in relevant technologies (e.g. AWS Certified Big Data, Google Professional Data Engineer) are a plus.

 

Security Clearance Requirements:

     

     

    Kalyan Ponnam Technical Recruiter

    | , Ext: 102 |
    11130 Fairfax Boulevard | Suite 200 | Fairfax | VA 22030
    Contract Vehicles: SBA 8(a) Certified | GSA 8(a) Stars III | GSA Multiple Award Schedule (MAS) 070