Data Engineer
Hybrid in Columbus, OH, US • Posted 7 hours ago • Updated 7 hours ago

MARKS IT SOLUTIONS LLC
Dice Job Match Score™
🫥 Flibbertigibetting...
Job Details
Skills
- Amazon Redshift
- Amazon Kinesis
- Amazon S3
- Apache Spark
- Electronic Health Record (EHR)
- Extract, Transform, Load
Summary
Job Title: Data Engineer
Location: Columbus, OH
Work Arrangement:
This position is based in Columbus, OH. Candidates must be local or willing to relocate.
We are seeking a skilled Data Engineer to design, develop, and maintain scalable data pipelines and distributed data processing systems. This role focuses on building robust ETL workflows, optimizing big data processing, and supporting cloud-based data platforms in a highly regulated enterprise environment. The ideal candidate will bring strong Python and Spark expertise, hands-on AWS experience, and a solid understanding of data warehousing and large-scale data systems.
Key Responsibilities:
Data Engineering & Pipeline Development
· Design, build, and maintain scalable data pipelines using Python and PySpark.
· Develop and optimize ETL workflows for structured and unstructured data.
· Process and transform large-scale datasets in distributed environments.
Cloud Data Platform & AWS Services
· Implement and manage data solutions using AWS services such as S3, Glue, EMR, Redshift, Athena, and Lambda.
· Optimize data storage, retrieval, and performance within AWS ecosystems.
· Support serverless and big data architectures in cloud environments.
Data Modeling & Warehousing
· Design and implement data models aligned with business and reporting requirements.
· Apply data warehousing concepts including star and snowflake schemas.
· Develop and optimize complex SQL queries for relational and non-relational databases.
DevOps & Version Control
· Collaborate using Git and version control best practices.
· Support CI/CD pipelines for automated deployment of data solutions.
· Ensure code quality, documentation, and adherence to enterprise standards.
Required Skills & Experience:
· Strong proficiency in Python for data processing and pipeline development.
· Hands-on experience with Apache Spark (PySpark preferred).
· Solid experience with AWS services including S3, Glue, EMR, Redshift, Athena, and Lambda.
· Strong SQL skills with experience in relational and non-relational databases.
· Knowledge of data modeling, data warehousing concepts, and ETL frameworks.
· Experience working with large-scale, distributed data systems.
· Familiarity with CI/CD pipelines and Git-based version control.
· Strong analytical, problem-solving, and communication skills.
Preferred / Nice-to-Have Skills:
· Experience with Airflow or other workflow orchestration tools.
· Knowledge of Kafka, Kinesis, or other streaming data platforms.
· Experience with Docker and Kubernetes.
· Exposure to Delta Lake, Iceberg, or Apache Hudi.
- Dice Id: 91171094
- Position Id: 8888606
- Posted 7 hours ago
Company Info
MARKS IT Solutions is a trusted partner in delivering agile and scalable workforce solutions across Technology and Business domains. We specialize in Recruitment Process Outsourcing (RPO), MSP/VMS staffing, International Talent Solutions, and comprehensive Managed Services, helping top employers build and manage high-performing teams worldwide.
Similar Jobs
It looks like there aren't any Similar Jobs for this job yet.
Search all similar jobs