offergenie_white
SAIC

Java Developer

SAIC

Ashburn, VA 20147Remote$80,001 - $120,000 a yearFull Time
Data EngineerRemote
Apply with AI Cover Letter

Job Description

Job ID: 2601247

Location: REMOTE WORK, VA, US

Date Posted: 2026-02-09

Category: Software

Subcategory: SW Engineer

Schedule: Full-time

Shift: Day Job

Travel: No

Minimum Clearance Required: Public Trust

Clearance Level Must Be Able to Obtain: None

Potential for Remote Work: Yes

Description

SAIC is looking for a Java Developer who will be responsible for converting existing PySpark codebases into optimized Java-based Spark applications. This role includes developing, refactoring, and maintaining scalable data processing solutions developed on the Databricks platform (or similar Spark execution environments).

Key Responsibilities:

· Convert existing PySpark applications into equivalent, efficient Java Spark implementations

· Design, develop, and maintain scalable Spark-based data pipelines

· Implement data processing logic using Java 8+ with best practices in OOP and functional programming

· Integrate solutions with IRS datasets including IRMF, BMF, and IMF

· Optimize Spark jobs for performance, maintainability, and cost-efficiency

· Collaborate across development, data engineering, and architecture teams

· Troubleshoot and debug Spark workloads in distributed environments

· Ensure compliance with IRS data handling, security, and governance policies

Qualifications

Required QualificationsRequired:

Bachelor’s degree in Computer Science, Information Systems, or a related field.

Active MBI Clearance

5+ years of professional experience in a data engineering or software development role.

Advanced expertise in:

· IRS datasets (IRMF, BMF, IMF) and tax system data structures.

· Java 8+ (experience with functional programming, Streams API, Lambdas).

· Apache Spark (Spark Core, Spark SQL, DataFrame APIs, performance tuning).

· Big data ecosystems (HDFS, Hive, Kafka, S3).

· Working with batch and streaming ETL pipelines for data processing.

Proficient with Git, Maven/Gradle, and DevOps tools.
Expertise in debugging Spark transformations and ensuring performance.

Preferred Qualifications:

Hands-on experience converting PySpark workloads into Java Spark.
Familiarity with ecosystems such as Databricks, Google Dataproc, or similar.
Knowledge of Delta Lake or Apache Iceberg.
Proven experience in big data performance modeling and tuning.

Target salary range: $80,001 - $120,000. The estimate displayed represents the typical salary range for this position based on experience and other factors.