Big Data Engineer- Remote

Remote • Posted 16 hours ago • Updated 16 hours ago
Contract Corp To Corp
Contract W2
Remote
$55/hr
Company Branding Image
Fitment

Dice Job Match Score™

✨ Finding the perfect fit...

Job Details

Skills

  • AWS
  • Redshift
  • spark
  • Scala

Summary

SkillSets:
Total EXP:
US EXP:
Spark:
Scala:
AWS:
Redshift:
ETL:
API/Webservice:
Java
Fully remote. Long term contract. Healthcare Experience needed!
Top Skills' Details
1. Senior level resource with Scala and Spark for data processing and ETL
2. 5+ years experience with AWS Cloud services like Spark, Scala, AWS EMR, RedShift and Airflow
3. Experience with REST APIs and web services
4. Software Engineering Experience/Advanced in Java.
Secondary Skills - Nice to Haves
  • Sql
  • python
Job Description
Reporting to Application Development Manager, the Big Data Software Engineer IV will work on a dedicated team of engineers developing, enhancing, and maintaining Availity's high transactional Provider Data Management platform.

Location: Remote US
Why work on this team:
This team supports a high transactional platform that directly impacts patient experience
This team is working to continually improve process and enhance platform capabilities
What you will be doing:
Develop a scalable and resilient cloud data platform and scalable data pipelines.
Ensure industry best practices around data pipelines, metadata management, data quality, data governance, and data privacy.
Build highly scalable AWS Infrastructure (from scratch or through 3rd party products) to enable Big Data Processing in the platform.
Find optimization within cloud resource usage to minimize costs while maintaining system reliability, including leveraging reserved instances and spot instances effectively.
Find performance sensitive considerations within development best practices, as well as, troubleshooting across the data platform utilizing tools (e.g., Splunk and New Relic, Cloud Watch, etc.) to ensure performance measurement and monitoring.
Participate in coding best practices, guidelines and principles that help engineers write clean, efficient, and maintainable code.
Participate in code reviews to catch issues, improve code quality, and provide constructive feedback to individuals within the team during code reviews.
Working on ETL transformation which includes gathering raw data and files from the client, transforming it into Availity's format and sending down the ETL pipeline for further processing
Working on a team following Agile Scrum principles
Incorporating development best practices
Ensuring your code is efficient, optimized, and performant
Collaborating on programming or development standards
Maintaining technical debt and applying security principles
Innovating with ideas and products to the organization
Performing unit testing and complex debugging to ensure quality
Learning new things & sharing your knowledge with others
Requirements:
Bachelor's degree preferably Computer Science, Engineering, or other quantitative fields
6+ years of related experience in designing and implementing enterprise applications using big data
5+ years of experience in a senior level engineering role mentoring other engineers, which includes engineering best practices, unblocking, code reviews, unit testing, managing deployments, technical guidance, system design, etc.
5+ years of experience working with large-scale data and developing SQL queries
Advanced experience with scripting languages (e.g., Python, Bash, node.js) and programming languages (e.g., SQL, Java, Scala) to design, build, and maintain complex data processing, ETL (Extract, Transform, Load) tasks, and AWS automation.
5+ years of hands-on experience with AWS cloud services, such as Apache Spark, with Scala, AWS EMR, Airflow, RedShift
4+ years of experience with RESTFul APIs and web services
Excellent communication skills including discussions of technical concepts, soft skills, conducting peer-programming sessions, and explaining development concepts
In-depth understanding of Spark framework, scripting languages (e.g., Python, Bash, node.js) and programming languages (e.g., SQL, Java, Scala) to design, build, and maintain complex data processing, ETL (Extract, Transform, Load) tasks, and AWS automation.
A firm understanding of unit testing.
Possess in-depth knowledge of AWS services and data engineering tools to diagnose and solve complex issues efficiently, specifically AWS EMR for big data processing.
In-depth understanding of GIT or other distributed version control systems.
Excellent communication. Essential to performing at maximum efficiency within the team.
Collaborative attitude. This role is part of a larger, more dynamic team that nurtures collaboration.
Strong technical, process, and problem-solving proficiency.
Thorough understanding of complex data structures and transformations, such as nested JSON, XML, Avro, or Parquet, into structured formats suitable for analysis and large datasets (100 gigs or more).
Advance skills in data cleansing, deduplication, and quality validation to maintain high-quality data in the transformed datasets.
Experience in the healthcare industry or another highly regulated field is a plus
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 10527831
  • Position Id: 2026-327
  • Posted 16 hours ago

Company Info

About CogniSoft Technologies

CogniSoft Technologies focuses on providing niche Business Intelligence & Data Analytic solutions to businesses across various industries. We are Tableau Alliance Partners. We are also an SAP certified Company.

We are a team of experienced, dedicated, hardworking, and innovative professionals who have had proven their expertise in different business verticals such as Banking and Financial Industry, Healthcare, Supply chain, Insurance, Software Development, IT consulting, Business consulting etc.

Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

Remote

3d ago

Easy Apply

Contract

Depends on Experience

Remote

Today

Easy Apply

Contract

105 - 112

Remote or Hybrid in Houston, Texas

3d ago

Easy Apply

Contract

Depends on Experience

Remote or Wisconsin

Today

Easy Apply

Contract, Third Party

$80

Search all similar jobs