Charles Schwab is currently seeking a mid level Data Engineer / Big Data ETL Engineer with a passion for hands-on design/development and collaboration with our business partners. The ideal candidate is adept at using large data sets to find opportunities for product and process optimization to test the effectiveness of different courses of action. This candidate must have deep experience in Enterprise Data Warehouse, Data Mart design, development and end to end execution of data solutions. Prior experience in developing design patterns for Big Data and Cloud solutions is a big plus.
This role will require hands on development with wide range of technical skills which include data analysis, ETL (Talend / Informatica, Unix, Big Data Technologies) and modeling skills (SQL, ERwin).
What you are good at
This position is part of the Data and Rep Technology (DaRT) organization that governs the strategy and implementation of the enterprise data warehouse and emerging data platforms.
This opportunity will allow you to design, build and support data processing pipelines to transform data using Hadoop technologies. You'll be designing schemas, data models and data architecture for Hadoop and HBase environments. You'll be implementing data flow scripts using Unix / Hive QL / Pig scripting. You'll be designing, building data assets in MapR-DB (HBASE), and HIVE. You'll be leveraging ETL tools like Talend and Informatica for data movement from multiple data sources and across disparate platforms. You'll be developing and executing quality assurance and test scripts. You'll be work with product owners and business analysts to understand business requirements and use cases to design solutions. You'll have the opportunity to grow in responsibility, work on exciting and challenging projects, train on emerging technologies and help set the future of the Data Solution Delivery team. You'll lead investigation and resolution efforts for critical/high impact problems, defects and incidents. Provides technical guidance to team members.
What you have
Bachelor's degree in Computer Science or related discipline
Experience with a structured application development methodology, using any industry standard Software Development Lifecycle, in particular Agile Methodologies is required
3+ years of overall experience in I.T. with strong understanding of best practices for building and designing ETL code, Strong SQL experience with the ability to develop, tune and debug complex SQL applications is required
2+ years of experience in ETL tools. Specific expertise in implementing Informatica / Talend in an Enterprise environment is a plus.
Experience architecting the whole process of consuming all the data from all the systems that are of interest
Experience in schema Analysis, design, developing data models and proven ability to work with complex data is required
Hands-on experience in Java object oriented programming (At least 2 years)
Hands-on experience with Hadoop, MapReduce, Hive, Pig, Flume, STORM, SPARK, Kafka and HBASE (At least 2 years)
Understanding Hadoop file format and compressions is required
Familiarity with MapR distribution of Hadoop is preferred
Understanding of best practices for building Data Lake and analytical architecture on Hadoop is required
Strong scripting / programming with UNIX, Java, Python, Scala etc. is required
Strong SQL experience with the ability to develop, tune and debug complex SQL applications is required
Expertise in schema design, developing data models and proven ability to work with complex data is required
Experience in real time data ingestion into Hadoop is required
Experience in or deep understanding of cloud based data technology Google Cloud Platform/AWS is preferred
Proven experience in working in large environments such as RDBMS, EDW, NoSQL, etc. is preferred
Knowledge of Big Data ETL such as Informatica BDM and Talend tools is preferred
Understanding security, encryption and masking using Kerberos, MapR-tickets, Vormetric and Voltage is preferred
Experience with Test Driven Code Development, SCM tools such as GIT, Jenkins is preferred
Experience with Graph database is preferred
Strong with SQL Server, Oracle and Mango DB preferred
Experience in Active Batch Scheduling , control M preferred
Excellent analysis, debugging and trouble-shooting skills, and problem solving skills
Good verbal and written communication skills
Ability to thrive in a flexible and fast-paced environment across multiple time zones and locations
Experience in Financial Services industry a plus.
Why work for us?
Own Your Tomorrow embodies everything we do! We are committed to helping our employees ignite their potential and achieve their dreams. Our employees get to play a central role in reinventing a multi-trillion-dollar industry, creating a better, more modern way to build and manage wealth.
Benefits: A competitive and flexible package designed to empower you for today and tomorrow. We offer a competitive and flexible package designed to help you make the most of your life at work and at home-today and in the future. Explore further .
Schwab is committed to building a diverse and inclusive workplace where everyone feels valued. As an Equal Opportunity Employer, our policy is to provide equal employment opportunities to all employees and applicants without regard to any status that is protected by law. Please click here to see the policy.
Schwab is an affirmative action employer, focused on advancing women, racial and ethnic minorities, veterans, and individuals with disabilities in the workplace. If you have a disability and require reasonable accommodations in the application process, contact Human Resources at or call.
TD Ameritrade, a subsidiary of Charles Schwab, is an Equal Opportunity Employer. At TD Ameritrade we believe People Matter. We value diversity and believe that it goes beyond all protected classes, thoughts, ideas, and perspectives.