Data Architect

  • Jersey City, NJ
  • Posted 30 days ago | Updated 30 days ago

Overview

On Site
$100,000 - $160,000
Full Time
10% Travel
Able to Provide Sponsorship

Skills

SQL
NoSQL
PostgreSQL
Architecture
Hadoop
HDFS
IMPALA
kafka
Hive
cloudera
oozie
nifi
Jenkins
GitLab
Hbase
Git
Java
Scala
Database
Data Warehouse
Cloudera Impala
Extract
Transform
Load
ETL

Job Details

Designation: Data Architect
Location: Jersey City, NJ / Tampa, FL - Onsite
Job Type: Fulltime

Role Overview

Key Responsibilities:

  • Architect and design large-scale, distributed big data solutions using Java and big data technologies to handle high-volume data processing and analytics.
  • Develop, test, and deploy data processing applications using Apache Spark and Scala.
  • Optimize and tune Spark applications for better performance on large-scale data sets.
  • Work with the Cloudera Hadoop ecosystem (e.g., HDFS, Hive, Impala, HBase, Kafka) to build data pipelines and storage solutions.
  • Collaborate with data scientists, business analysts, and other developers to understand data requirements and deliver solutions.
  • Design and implement high-performance data processing and analytics solutions.
  • Ensure data integrity, accuracy, and security across all processing tasks.
  • Troubleshoot and resolve performance issues in Spark, Cloudera, and related technologies.
  • Implement version control and CI/CD pipelines for Spark applications.
Required Skills & Experience:
  • Minimum 15+ years of experience in application development.
  • Strong hands on experience in Apache Spark, Scala, and Spark SQL for distributed data processing.
  • Hands-on experience with Cloudera Hadoop (CDH) components such as HDFS, Hive, Impala, HBase, Kafka, and Sqoop.
  • Familiarity with other Big Data technologies, including Apache Kafka, Flume, Oozie, and Nifi.
  • Experience building and optimizing ETL pipelines using Spark and working with structured and unstructured data.
  • Experience with SQL and NoSQL databases such as HBase, Hive, and PostgreSQL.
  • Knowledge of data warehousing concepts, dimensional modeling, and data lakes.
  • Ability to troubleshoot and optimize Spark and Cloudera platform performance.
  • Familiarity with version control tools like Git and CI/CD tools (e.g., Jenkins, GitLab).
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.