Data Engineer

Overview

On Site
$50 - $60
Accepts corp to corp applications
Contract - W2
Contract - 12 Month(s)

Skills

Apache Kafka
Scala
RESTful
Java
Apache Spark
Apache Storm
API
Amazon Web Services
Cloud Computing
Vertex

Job Details

Preferred location is Sunnyvale, CA although will accept candidates from Bentonville, AR

Job Description:

We are seeking a skilled Data Engineer with strong expertise in Big Data, Cloud platforms, and distributed data systems. The ideal candidate will have hands-on experience in designing, building, and optimizing data pipelines, API integrations, and real-time stream-processing systems.


Key Responsibilities:

  1. Design, develop, and optimize large-scale data pipelines and ETL workflows.

  2. Work with Java, Python, and Scala to build scalable data solutions.

  3. Develop APIs and integrate with systems using Node.js, GQL, and RESTful services.

  4. Implement big data solutions leveraging Hadoop, Hive, Spark (Scala), Presto/Trino, and Data Lake architectures.

  5. Deploy and manage workflows with Airflow, Luigi, Automic, and similar orchestration tools.

  6. Build and maintain real-time data streaming systems using Storm, Spark-Streaming, and Kafka.

  7. Utilize Vertex AI and Cloud services (AWS/Google Cloud Platform/Azure) for advanced analytics and ML integration.

  8. Ensure system reliability, scalability, and performance in distributed environments.

  9. Collaborate with cross-functional teams (data scientists, analysts, and engineers) to deliver high-quality data solutions.

  10. Apply best practices in CI/CD, Kubernetes-based deployments, and monitoring.


Required Skills:

  • Strong programming skills in Java, Python, and Scala.

  • Expertise in big data frameworks: Hadoop, Hive, Spark (Scala).

  • Hands-on with API development (REST, GraphQL, Node.js).

  • Experience with stream-processing tools: Kafka, Storm, Spark-Streaming.

  • Proficiency with workflow orchestration: Airflow, Luigi, Automic.

  • Knowledge of Presto/Trino and distributed SQL query engines.

  • Cloud experience (AWS, Google Cloud Platform, or Azure), with exposure to Vertex AI.

  • Strong understanding of Data Lake and data warehousing concepts.

  • Experience with Kubernetes for container orchestration.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

About Purple Drive Technologies LLC