Confluent Kafka Admin SME at Tampa, FL/Irving, TX/New Castle, DE

Overview

On Site
Depends on Experience
Full Time
No Travel Required

Skills

Big Data
Confluent Kafka
Spark
Scala
Admin

Job Details

Key Skills:

BigData, Confluent Kafka Administration, Spark, Scala

Detailed Job Description for Kafka Admin SME at Tampa, FL/Irving, TX/New Castle, DE:

  • Big Data Confluent KAFKA Administrator Tech Lead
  • Creation of key performance metrics, measuring the utilization, performance, and overall health of the cluste Perform high level, day-to-day administration and support functions.
  • Capacity planning and implementation of new/upgraded hardware and software releases as well as for storage infrastructure.
  • Research and recommend innovative ways to maintain the environment and where possible, automate key administration tasks.
  • Ability to work with various infrastructure, administration, and development teams across business units.
  • Document and share design, build, upgrade and standard operating procedures. Conduct knowledge transfer sessions and workshops for other members in the team. Provide technical expertise and guidance to new and junior members in the team.
  • Implemented and supported any enterprise product such as any well known ERP products, Data warehouse, Middleware etc.
  • Minimum 6 years of experience in BigData Administration using KAKFA and related tools like Spark etc.
  • Minimum 4 years of experience in setting up the environment for solutions that collects and process data using Kafka Spark, Scala
  • Expert on Kafka Big Data solutions along with sound understanding of Kafka architecture.
  • Manage single and multi-node Kafka cluster deployed on VM, Docker and Kubernetes Container platform. Experience with Confluent Platform running on prem.
  • Perform Kafka Cluster build, MRC, including Design, Infrastructure planning, High Availability and Disaster Recovery
  • Implementing wire encryption using SSL, authentication using SASL LDAP authorization using Kafka ACLs in Zookeeper, Broker Client, Connect cluster connectors, Schema Registry, REST API, Producers Consumers, KsqlDevelop and maintain Unix scripts to perform day to day Kafka Admin and Security related functions using Confluent REST Proxy server
  • Setting up monitoring tools such as Prometheus, Grafana to scrape metrics from various Kafka cluster components Broker, Zookeeper, Connect, REST proxy, Mirror Maker, Schema Registry and other endpoints such as webservers, databases, logs etc. and configure alerts for Kafka Cluster and supporting infrastructure to measure availability and performance SLAs.
  • Experience with Confluent ksql to query and process Kafka streams
  • Knowledge of Kafka Producer and Consumer APIs, Kafka Stream Processing, Confluent Ksql
  • Security related config for above listed software or any other tools in SSL for wire encryption, integration with AD for authentication and RBAC for authorizations
  • Database administration skills in Oracle, MSSQL, SAP HANA, DB2, Aerospike, Postgres.
  • Exposure to SaaS based observability platform like New Relic
  • Deployment of container images and pods using CI/CD pipelines using Jenkins or comparable tools.
  • Experience in building Kafka deployment pipelines using Terraform, Ansible, Cloud formation templates, shells etc.
  • Availability to work in shifts, extended hours and to provide on call support as required. There will be work over weekends at times depending on the project needs.
  • Worked in Public cloud environment such as Azure or AWS or Google Cloud Platform, preferably in Azure.
  • Must have excellent communications and interpersonal skills Preferred but Optional skills.
  • Experience in a Financial Services or large complex and or global environment preferred.
  • Effective written and verbal communication skills
  • Effective analytic diagnostic skills
  • Good presentation skills