Develop Java-based tools and applications to implement solutions that enhance data processing, automated processes, automate manual tasks, streamline data workflows, improve overall operational efficiency, scripts to validate data integrity, identify anomalies, and generate timely reports for stakeholders.
• Develop and maintain tools and applications to enhance data processing, automate tasks, and streamline workflows, improving overall operational efficiency.
• Collaborate with team members to understand complex product requirements and translate them into data-driven solutions, leveraging frameworks and libraries for complex data models.
• Implement development processes and coding best practices, conduct code reviews, and ensure high-quality data solutions, including data validation, anomaly identification, and timely reporting for stakeholders.
• Set up and manage Kubernetes services, deployments, and Nginx ingress configurations for deploying microservices and orchestrating applications on cloud environments.
• Utilize CI/CD tools like Jenkins, Maven, SonarQube, and Docker for continuous integration and deployment, and manage version control with YAML configurations and shell scripts.
• Optimize Kafka consumer configurations for performance and reliability, handle real-time events and data processing using Databricks and AWS DynamoDB, and publish data using REST APIs on Kafka topics.
• Collect, clean, and transform data from various sources using ETL processes, execute batch processing, and design and query databases to support robust data governance practices.
• Develop and maintain RESTful API endpoints for seamless data integration and management and implement REST APIs to integrate workflow functionalities into external systems.
• Communicate insights effectively through data visualization and reporting tools, interpret data using statistical tools, and execute real-time analytics on stored and live datasets to facilitate data-driven decision-making.