Kafka Admin

  • New York, NY
  • Posted 5 hours ago | Updated 5 hours ago

Overview

On Site
$40 - $45
Contract - W2
Contract - Independent

Skills

Kafka
Admin
AWS/Azure/GCP

Job Details

Role: Kafka Admin

Location: NYC, NY (Onsite)

Key Responsibilities

  • Install, configure, and manage Apache Kafka clusters (on-prem and/or cloud) and Confluent Platform components.

  • Perform capacity planning, scaling, and cluster upgrades with minimal downtime.

  • Manage Kafka topics, partitions, ACLs, and replication policies.

  • Implement and enforce security standards including authentication (SASL/SCRAM, Kerberos) and authorization (RBAC).

  • Monitor Kafka infrastructure using tools like Prometheus, Grafana, Confluent Control Center, and ELK.

  • Troubleshoot performance and latency issues across producers, consumers, and brokers.

  • Automate routine operational tasks using Python, Shell, or Ansible scripts.

  • Manage disaster recovery configurations, data replication, and multi-data-center synchronization.

  • Collaborate with DevOps, Cloud, and Application Engineering teams to integrate Kafka into CI/CD pipelines.

  • Maintain documentation of architecture, topology, configurations, and standard operating procedures.

  • Provide 24x7 support for critical production incidents on a rotational basis.


Required Skills & Experience

  • 7+ years of overall IT experience with at least 3 5 years of hands-on Kafka administration.

  • Strong knowledge of Kafka internals brokers, zookeepers/KRaft, topics, producers, consumers, offsets, replication, and retention policies.

  • Experience with Confluent Kafka, Kafka Connect, Schema Registry, and KSQL.

  • Experience with monitoring and alerting tools Prometheus, Grafana, Splunk, Control Center, etc.

  • Proficiency in Linux system administration and performance tuning.

  • Scripting experience in Python, Bash, or Ansible for automation.

  • Hands-on experience with containerization (Docker) and orchestration (Kubernetes/OpenShift).

  • Knowledge of AWS, Google Cloud Platform, or Azure Kafka deployments (MSK, Confluent Cloud) is a plus.

  • Familiarity with CI/CD pipelines and DevOps tools such as Jenkins, GitLab, or Ansible Tower.

  • Understanding of high availability, disaster recovery, and data replication strategies.

  • Excellent communication, analytical, and problem-solving skills.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.