Data Engineer

  • St. Louis, MO
  • Posted 10 days ago | Updated 11 hours ago

Overview

Remote
On Site
Hybrid
BASED ON EXPERIENCE
Contract - Independent
Contract - W2

Skills

DATA ENGINEER
DATA ANALYST
SQL
PYTHON

Job Details

Data Engineer

Connecting People with Their Purpose

At TalentBridge, we know success begins with personal connections, and our goal is to fully understand your values, motivations, and passions, then to connect you with companies whose values and motivations match your own.

Whether you're looking for temporary employment or a full-time career, the TalentBridge family is here to help.

About the Role

We're looking for a talented Data Engineer to join one of our Global 500 clients and be part of a growing team that plays a critical role in the organization.

In this role, your responsibilities will include:

  • Ingest, store, transform, and curate data from various internal and external sources including databases, flat files, and API endpoints
  • Query databases and data warehouses to access data using efficient SQL queries, PySpark, and ETL tools
  • Optimize solutions with a focus on usability, performance, flexibility, testability, and standardization
  • Demonstrate commitment to software development best practices, including documentation, reusable code design, test driven implementation, and integrated software testing
  • Document and demonstrate each solution by producing documentation, process flow diagrams, code comments, and clear code
  • Troubleshoot and analyze data inconsistencies, bugs, and performance related issues, often shifting priorities in a rapidly changing environment

Required Skills, Education, and Experience
  • Bachelor's degree required. Concentration in computer science, IT/IS, engineering, statistics, supply chain, business, or related field preferred but not required if experience in these fields can be substituted.
  • At least 2 years of experience in data engineering or related field (software engineering, data analytics, business intelligence, data science)
  • Ability to adapt quickly to an existing, complex environment
  • Ability to take initiative, handle multiple tasks, shift priorities, and meet deadlines
  • Ability to quickly learn new concepts and software as necessary, with little oversight
  • Strong interpersonal, written, and verbal communication skills
  • Technical proficiency in the following:
    • Strong SQL and Python skills required (including Pandas or PySpark)
    • Experience with ETL/ELT processes, data pipelines, orchestration, and workflow management required
    • Knowledge of relational database management systems and data warehousing concepts required
    • Experience with Git or knowledge of Azure DevOps, GitHub, or similar tools (e.g. Jira or GitLab) required
    • Experience with Databricks and/or Snowflake preferred
    • Experience with Azure ecosystem a plus - ADF, ADLS, Azure SQL DB
    • Experience with REST APIs a plus
    • Experience with Microsoft Excel, SAS, Oracle, and/or Microsoft SQL Server a plus
Specific Technical Skills:
  • SQL - Should be able to write SQL code very comfortably
    • DQL/DML statements, including a mastery of joins, conditionals/case when, aggregation, and filtering on datasets
    • DDL/defining objects like creating tables and views
  • Python
    • Should know the fundamentals, including the standard library and functions/OOP (i.e. should know what a class is, passing arguments into a method, etc)
      • Data engineering/data science experience preferred but also open to strong Python experience in web development or another related area
    • Need to be comfortable with lists, tuples, sets, and dictionaries
    • Databricks/PySpark and/or Pandas required
      • If you do not have experience with Databricks/PySpark, you must have experience with Pandas
      • You need to be familiar with the concept of a dataframe - getting data into it, manipulating it, running analysis on it, managing multiple dataframes together (joins, unions, etc), and so forth
Nice to Haves:
  • Familiarity with other libraries such as Requests (used commonly in this role for API calls)
  • Other advanced data science libraries such as NumPy and Matplotlib are always nice but not required for this role
  • Familiarity with advanced Data Structures