Overview
On Site
$50 - $60
Accepts corp to corp applications
Contract - W2
Contract - Independent
Contract - 12 Month(s)
Skills
data engineer
airflow
flink
nifi
delta lake
Job Details
Job Title: Data Engineer
Location: Boston, MA (Onsite 4 days a week, Local Candidates Only)
Type: C2C / Onsite
Experience: 5+ Years
Key Responsibilities
- Develop and optimize data flows, analytics pipelines, and ML-driven systems for cybersecurity data.
- Design and build petabyte-scale distributed systems ensuring high availability, throughput, data security, and privacy.
- Create data models, ELT workflows, and automation for raw, rationalized, correlated, and reporting data (graph, timeseries, structured & semi-structured).
- Leverage modern big data technologies (Spark, Presto/Trino, Flink, Kafka, Delta Lake, Iceberg, Databricks, Snowflake, Neo4j, MemGraph).
- Integrate security tooling (Splunk, Datadog, Elastic, Panther, Arcsight) into analytics workflows.
- Build data APIs and integrations (GraphQL, Apache Arrow, gRPC, ProtoBuf).
- Partner with data scientists, analysts, and cybersecurity stakeholders to deliver scalable, compliant, and innovative data products.
Required Skills & Qualifications
- Bachelor s or Master s in computer science, Engineering, or related field (or equivalent work experience).
- 5+ years building large-scale distributed systems and data analytics processes.
- Strong expertise in Spark (GraphX/GraphFrames, Structured Streaming) and scalable graph/time-series operations.
- Hands-on experience with data lakes/warehouses (Snowflake, Redshift, Postgres, Cassandra, BigQuery, Athena, Presto).
- Strong SQL and ETL/ELT development skills.
- Proficiency in Python, Java, or similar languages with cloud platforms (AWS, Google Cloud Platform, or Azure).
- Experience with stream & batch processing, MLOps, feature stores, and iterative model deployment cycles.
- Knowledge of data governance, lineage, security, compliance, and DataSecOps platforms (Immuta, Satori) including RBAC/ABAC, data masking, and tokenization.
- Familiarity with Airflow, Prefect, Flink, NiFi, Delta Lake, Iceberg, LakeFS.
Preferred Experience
- Building globally distributed federated data systems.
- Implementing data quality monitoring, history, and time-travel in data pipelines.
- Experience in highly regulated environments with security, governance, and compliance requirements.
- Strong collaboration and communication skills, particularly with cross-functional teams.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.