Overview
Skills
Job Details
Hadoop Admin (Hortonworks & Big Data - Cloudera)
-Understand, monitor, and evaluate production Hadoop job streams in support of Bank Operations.
-Serve as architect and technical lead on medium scale to large scale projects as required
-Accountable for software design and code quality
-Utilizes development languages and tools to create and implement design
-Demonstrates continuous professional and technical development
-Task management ensures on time delivery
-Creates high level technical design for use by Software Engineer
-Participates in review meetings as necessary
-Actively participates in the continuous improvement of software development process
-Recommendations for improvement
-Follows the defined development / implementation process
-Ability to advocate ideas and to objectively participate in design critique
-3 years Architect level contributions with Hadoop distributions (Horton, Cloudera, Pivotal)
-Experience with architecting developing applications on Hortonworks (HIVE, SPARK), Apache NiFi, Talend
-Experience in building and maintaining Hadoop/spark clusters
-Programming languages Spark, Scala, Java/J2EE, Linux, Wakari, Anaconda, R on Hadoop, PHP, Python, Hadoop, Hive, HBase, Pig, MapReduce and other Hadoop eco-system components
-Good knowledge of Unix/Linux, Ruby, Python, Perl or other Scripting languages
-Design and development of data models for a new HDFS Master Data Reservoir and relational or object Current Data environments
-Ability to provide guidance, based on real experience, on the structure of a multi-tenant Hadoop infrastructure
-Experience with data analysis tools such SAS, SSAS, or Tableau, Starburst, Trino
-Experience with data reporting tools such as Tableau or Datameer
-Hands-on experience with the overall Hadoop ecosystem
-Hands-on experience with related/complementary open source software platforms and languages
-Hadoop architect provides technical leadership and expertise within the Hadoop environment using technologies such as HDFS, MapReduce, Hive, Impala, HBase, Flume, ZooKeeper, Spark/Spark Streaming, Kafka and other Big Data technologies.
-Should be able to identify immediate solution and permanent fix based on the root cause analysis.
-Must have experience dealing with complex inquiries raised by customers in a large, multi-tenant, Hadoop environment.
-Should have efficient communication skills to deliver the details to tenants and drive the triage sessions. Architect and design data patterns, compute and design strategies for enterprise applications to support client business processes and functional requirements as it relates to complex data driven solutions