Senior Databricks Data Engineer

Overview

On Site
Full Time

Skills

Root cause analysis
Agile
Management information systems
Data architecture
Operational efficiency
Data engineering
Design review
Microsoft Azure
Stored procedures
Microsoft SQL Server
Data Lake
Code refactoring
Design patterns
Real-time
Dimensional modeling
Data governance
Data warehouse
Unstructured data
Transact-SQL
Relational databases
Problem solving
Effective communication
Parallel computing
Computer science
Information Technology
Business management
Databricks
Data
Management
Design
Extract
transform
load
Collaboration
Mentorship
Analytical skill
Documentation
Coaching
Python
Scala
ADF
SQL
Database
Continuous integration
Continuous delivery
Boomi
MapReduce
Apache Spark
Training

Job Details

Senior Databricks Data Engineer
- Direct Hire
- 100% Remote
- Must be authorized to work in the U.S.

Description:
  • Design, develop, optimize, and maintain data architecture and pipelines that adhere to ETL principles and business goals.
  • Collaborate with data engineers, data consumers, and other team members to come up with simple, functional, and elegant solutions that balance the data needs across the organization.
  • Solve complex data problems to deliver insights that helps the organization achieve its goals.
  • Create data products that will be used throughout the organization.
  • Advise, consult, mentor and coach other data and analytic professionals on data standards and practices.
  • Foster a culture of sharing, re-use, design for scale stability, and operational efficiency of data and analytic solutions.
  • Develop and deliver documentation on data engineering capabilities, standards, and processes; participate in coaching, mentoring, design reviews and code reviews.
  • Partner with business analysts and solutions architects to develop technical architectures for strategic enterprise projects and initiatives.
  • Deliver awesome code.
Technical Qualifications:
  • 7+ years relevant and progressive data engineering experience.
  • Deep Technical knowledge and experience in Databricks, Python, Scala, Microsoft Azure architecture and platform including Synapse, ADF (Azure Data Factory) pipelines and Synapse stored procedures.
  • Hands-on experience working with data pipelines using a variety of source and target locations (e.g., Databricks, Synapse, SQL Server, Data Lake, file-based, SQL and No-SQL database).
  • Experience in engineering practices such as development, code refactoring, and leveraging design patterns, CI/CD, and building highly scalable data applications and processes.
  • Experience developing batch ETL pipelines; real-time pipelines are a plus.
  • Knowledge of advanced data engineering concepts such as dimensional modeling, ETL, data governance, data warehousing involving structured and unstructured data.
  • Thorough knowledge of Synapse and SQL Server including T-SQL and stored procedures.
  • Experience working with and supporting cross-functional teams in a dynamic environment.
  • A successful history of manipulating, processing and extracting value from large disconnected datasets.
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
  • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
  • Knowledge and understanding of Boomi is a plus.
Additional Qualifications and Experience:
  • Excellent problem-solving skills and experience.
  • Effective communication skills.
  • Strong collaboration skills.
  • "Self-starter" attitude and the ability to make decisions with minimal guidance from others.
  • Innovative and passionate about your work and the work of your teammates.
  • Ability to comprehend and analyze operational systems and ask appropriate questions to determine how to improve, migrate or modify the solution to meet business needs.
  • Experience with data ingestion and engineering, specifically involving large data volumes.
  • Knowledge of CI/CD release pipelines is a plus.
  • Understanding of Python and knowledge of parallel processing frameworks like MapReduce, Spark, Scala.
  • Knowledge of the Agile development process.
Working Conditions:
  • Remote.
  • Occasional time outside of core hours depending on priorities and deliverables.
  • Some travel (with advance notice) may be required for training, team meetings or other events.
Education:
  • Bachelor's degree in computer science, information technology, business management information systems, or equivalent experience.