Lead Data Engineer

Agile, Analysis, Architecture, Automated, Bash, Customer Service, Data Analysis, Data Modeling, Data Warehouse, Development, Ecommerce, Java, Linux, Management, Micro, Modeling, NoSQL, Python, Scrum, SQL, XML
Full Time
Work from home not available Travel not required

Job Description


For over a century, Neiman Marcus Group has served the unique needs of our discerning customers by staying true to the principles of our founders: to be the premier omni-channel retailer of luxury and fashion merchandise dedicated to providing superior service and a distinctive shopping experience in our stores and on our websites. Neiman Marcus Group is comprised of the Specialty Retail Stores division, which includes Neiman Marcus and Bergdorf Goodman, and our international brand, mytheresa.com. Our portfolio of brands offers the finest luxury and fashion apparel, accessories, jewelry, beauty, and home dcor. The Company operates more than 40 Neiman Marcus full-line stores in the most affluent markets across the United States, including U.S. gateway cities that draw an international clientele. In addition, we operate 2 Bergdorf Goodman stores in landmark locations on Fifth Avenue in New York City. We also operate more than 40 Last Call by Neiman Marcus off-price stores that cater to a value oriented, yet fashion minded customer. Our upscale eCommerce and direct-to-consumer division includes NeimanMarcus.com, BergdorfGoodman.com Horchow.com, LastCall.com, and CUSP.com. Every day each of our 15,000 NMG associates work towards the goal of enabling our customers to shop any of our brands "anytime, anywhere, and on any device." Whether the merchandise we sell, the customer service we offer, or our investments in technology, everything we do is to enhance the customer experience across all channels and brands.

Data engineer will have the unique combination of business acumen needed to interface directly with key stakeholders to understand the problem along with the skills and vision to translate the need into a world-class technical solution using the latest technologies

This person will be a hands-on role who is responsible for building data engineering solutions for NMG Enterprise using cloud based data platform. They will provide day-to-day technical leadership and active oversight for technical design, development and support for data engineering workloads. In this role, you need to be equally skilled with the whiteboard and the keyboard.

Job Duties

  • Work with business partners and data science teams to understand business context and craft best-in-class solutions to their toughest problems

  • Develop deep understanding of business knowledge of data and Retail domain

  • Create robust and automated pipelines to ingest and process structured and unstructured data from source systems into analytical platforms using batch and streaming mechanisms leveraging cloud native toolset

  • Lead building a scalable next gen Cloud platform to so support data transformation, ML and AI Strategy including migration of current platform to AWS.

  • Work with data scientists to operationalize and scale machine learning training and scoring components by joining and aggregating data from multiple datasets to produce complex models and low-latency feature store

  • Lead and guide the evolution of various data management functions (Data Catalog, Data Prep, Data Lineage & Metadata Management functions)

  • Provide hands-on technical leadership in all aspects of data engineering design and implementations including data ingestion, data models, data structures, data storage, data processing, and data monitoring at scale

  • Develop data engineering best practices with considerations for high data availability, computational efficiency, cost, and quality

  • Lead, Design and develop automation to optimize data platform compute and storage resources

  • Evaluate and make recommendations on new technologies, tools and guide the team on up-to-date technologies, standards, and practices

  • Coach and mentor more junior resources

  • Contribute to the internal knowledge base to build expertise and awareness within the organization

  • Assist in our recruiting and interviewing process

  • Participate in developing projects plan, timelines and providing estimates

  • Participate in daily scrum calls and provide clear visibility to work products

Job Requirements

  • BS in Computer Science or related field

  • 6+ years of experience in the data and analytics space

  • Certification preferably AWS Certified Big Data or any other cloud data platforms, big data platforms

  • 4+ years experience developing and implementing enterprise-level data solutions utilizing Python (Scikit-lean, Scipy, Pandas, Numpy, Tensorflow) , Java, Spark, and Scala, Airflow , Hive and Python.

  • 3+ years in key aspects of software engineering such as parallel data processing, data flows, REST APIs, JSON, XML, and micro service architectures.

  • 2+ year of experience working on Big Data Processing Frameworks and Tools Map Reduce, YARN, Hive, Pig, Oozie, Sqoop, and good knowledge of common big data file formats (e.g., Parquet, ORC, etc.)

  • 6+ years of RDBMS concepts with Strong Data analysis and SQL experience

  • 3+ years of Linux OS command line tools and bash scripting proficiency

Nice to have:

  • Kubernetes and Docker experience a plus

  • Prior working experience on data science work bench

  • Cloud data warehouse experience - Snowflake is a plus

  • Data Modeling experience a plus

Knowledge, Skills and Abilities:

  • A passion for technology and data analytics with a strong desire to constantly be learning and honing skills

  • Ability to deliver independently without oversight

  • Be productive even with ambiguity and highly fluid requirements during initial stages of projects

  • Flexibility to work in matrix reporting structure

  • Experienced in implementing large scale event based streaming architectures

  • Strong communication and documentation skills

  • Working knowledge of NoSQL, in-memory databases

  • Background in all aspects of software engineering with strong skills in parallel data processing, data flows, REST APIs, JSON, XML, and micro service architecture

  • Experienced in collaborating with cross-functional IT teams and global delivery teams

  • Solid Programing experience in Python - needs to be an expert in this 4/5 level

  • Working knowledge of data engineering aspects within machine learning pipelines (e.g., train/test splitting, scoring process, etc.)

  • Experience working in a scrum/agile environment and associated tools (Jira)

Dice Id : RTX1854dc
Position Id : 10020
Have a Job? Post it