BigData / Big Data / Infrastructure Engineer

YARN, Spark Core, Spark SQL, Spark Streaming, Scala, Map Reduce, Hive 2.3, Pig 0.17, Zookeeper 3.4.11, Sqoop 1.4, Oozie 4.3, Bedrock, Apache Flume 1.8, Kafka 2.0, Impala 3.0, Nifi, MongoDB, HBase, Python, PL/SQL, Java, HiveQL, Pig Latin, UNIX shell scripting, Hortonworks, Azure, Amazon Web services (AWS), Redshift, EMR, EC2, S3, RDS, Cloud Search, Data Pipeline, Lambda, Oracle 12c, MS-SQL Server 2017, MySQL, PostgreSQL, NoSQL (HBase, Cassandra 3.11, MongoDB), Teradata r14, Eclipse 4.8, NetBeans 9.0, Informatica, IBM DataStage, Talend, Maven, Jenkins 2.12, GitHub, SVN, CVS., DataWare House & Business Intelligence, Linux systems and shell scripting, Ab Initio, Cloudera, Active Directory/LDAP, NoSQL stores, Hbase, Avro, Kafka, Spark, Kudu, Big Data, Hadoop, MapReduce, Yarn, Hive, HDFS, Oozie
Contract Corp-To-Corp, Contract Independent, Contract W2, 12 Months
Depends on Experience
Work from home not available Travel not required

Job Description

  • 12 + Contract (Extendable up to 36 months)
  • Corp-to-Corp, W2, 1099 are most welcome
  • In-Person Interview -- No Exception

Responsibilities:

  • Provide support for successful installation and configuration of new Big Data clusters across various environments
  • Provide support for the successful expansion of Big Data clusters across various environments
  • Provide day to day Big Data administration support
  • Work with Big Data team, Infrastructure Architect, Change Management team to support configuration, code migrations of Big Data Deliverables
  • Looks to leverage reusable code modules to solve problems across the team, including Data Preparation and Transformation and Data export and synchronization
  • Act as a Big Data administration liaison with Infrastructure, Security, Application Development , Project Management
  • Keep current on latest Big Data technologies and products, including hands-on evaluations and in-depth research
  • Works with Big Data lead/architect to perform detailed planning, risk/issue management

Required Qualifications

    • 5+ years of administrator experience working with batch processing and tools in the Hadoop technical stack (e.g. MapReduce, Yarn, Hive, HDFS, Oozie)
    • 5+ years of administrator experience working with tools in the stream processing technical stack (e.g. Kudu, Spark, Kafka, Avro)
    • Administrator experience with NoSQL stores (e.g. Hbase)
    • Expert knowledge on Active Directory/LDAP security integration with Big Data
    • Hands-on experience with Cloudera Hadoop Distribution
    • Hands-on experience monitoring and reporting on Hadoop resource utilization
    • 5+ years of doing data related benchmarking, performance analysis and tuning
    • Expert level scripting skills
    • Hands-on experience supporting code deployments (Spark, Hive, Ab Initio, etc.) into the Hadoop cluster
    • 4+ years of experience with SQL and at least two major RDBMS s
  • 6+ years as a systems integrator with Linux systems and shell scripting
  • Bachelor s degree in Computer Science, Information Systems, Information Technology or related field and 8+ years of software development/DW & BI experience
  • Excellent verbal and written communication skills

Preferred Qualifications

    • Healthcare industry experience
    • Master of Science in Mathematics, Engineering, or Computer Science
    • ETL solution experience, preferably on Hadoop
    • Experience with industry leading Business Intelligence tools

Posted By

Samuel Johnson

Dice Id : 10201174
Position Id : 5935119
Have a Job? Post it