Hybrid in Irving, Texas
•
11d ago
Key Responsibilities: Design, develop, and optimize scalable data pipelines using Apache Spark, Python, and Hadoop. Implement robust data ingestion, transformation, and storage solutions for large-scale datasets. Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions. Manage and deploy Big Data tools and frameworks including Kafka, Hive, HBase, and Flink. Ensure data quality, integrity, and availability across distributed systems.
Easy Apply
Full-time
Depends on Experience