Job ID: 2601247
Location: REMOTE WORK, VA, US
Date Posted: 2026-02-09
Category: Software
Subcategory: SW Engineer
Schedule: Full-time
Shift: Day Job
Travel: No
Minimum Clearance Required: Public Trust
Clearance Level Must Be Able to Obtain: None
Potential for Remote Work: Yes
Description SAIC is looking for a Java Developer who will be responsible for converting existing PySpark codebases into optimized Java-based Spark applications. This role includes developing, refactoring, and maintaining scalable data processing solutions developed on the Databricks platform (or similar Spark execution environments).
Key Responsibilities:
Convert existing PySpark applications into equivalent, efficient Java Spark implementations
Design, develop, and maintain scalable Spark-based data pipelines
Implement data processing logic using Java 8+ with best practices in OOP and functional programming
Integrate solutions with IRS datasets including IRMF, BMF, and IMF
Optimize Spark jobs for performance, maintainability, and cost-efficiency
Collaborate across development, data engineering, and architecture teams
Troubleshoot and debug Spark workloads in distributed environments
Ensure compliance with IRS data handling, security, and governance policies
Qualifications Required QualificationsRequired:
- Bachelor's degree in Computer Science, Information Systems, or a related field.
- Active MBI Clearance
- 5+ years of professional experience in a data engineering or software development role.
- Advanced expertise in:
IRS datasets (IRMF, BMF, IMF) and tax system data structures.
Java 8+ (experience with functional programming, Streams API, Lambdas).
Apache Spark (Spark Core, Spark SQL, DataFrame APIs, performance tuning).
Big data ecosystems (HDFS, Hive, Kafka, S3).
Working with batch and streaming ETL pipelines for data processing.
- Proficient with Git, Maven/Gradle, and DevOps tools.
- Expertise in debugging Spark transformations and ensuring performance.
Preferred Qualifications:
- Hands-on experience converting PySpark workloads into Java Spark.
- Familiarity with ecosystems such as Databricks, Google Dataproc, or similar.
- Knowledge of Delta Lake or Apache Iceberg.
- Proven experience in big data performance modeling and tuning.
Target salary range: $80,001 - $120,000. The estimate displayed represents the typical salary range for this position based on experience and other factors.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
- Dice Id: 10111346
- Position Id: 2601247
- Posted 2 hours ago