Overview
Skills
Job Details
Pay Range: $ 65 - $ 70
Job Description:
Responsibilities:
Manage daily operations of big data and containerized platforms including Hadoop, Spark, Kafka, Kubernetes/OpenShift, Docker/Podman, and Jupyter Notebook.
Support, maintain, and troubleshoot AI/ML and data analytics platforms through upgrades, incident response, and cluster management.
Implement monitoring, alerting, and capacity forecasting to ensure platform reliability and performance.
Automate system tasks and deployment pipelines using tools like Ansible, Shell scripting, and Python.
Collaborate with development teams and stakeholders in an agile environment to support rapid software delivery.
Drive initiatives for platform resilience, disaster recovery, and performance optimization.
Must-Haves:
Strong understanding of Big Data platforms: Hadoop, Spark, Kafka, HBase, Impala.
Hands-on experience with Cloudera Hadoop stack and familiarity with HDFS internals (fsimage, blocks, etc.).
Scripting and automation experience using Python, Shell, and Ansible.
Proficiency in SQL and experience with databases such as Cassandra and Postgres.
Familiarity with monitoring and logging tools like ELK Stack or Splunk.
Solid experience in Unix environments and agile development practices.
Nice-to-Haves:
Exposure to AI/ML platforms such as DataRobot, C3 AI, or similar.
Strong containerization and orchestration skills (Docker, Podman, Kubernetes/OpenShift).
Background in financial or enterprise-scale environments.