Overview
Skills
Job Details
Job Summary:
We are looking for an experienced Data Engineer to design, build, and maintain scalable data pipelines and infrastructure. The ideal candidate will have strong expertise in ETL processes, data warehousing, cloud platforms, and big data technologies, ensuring reliable and efficient data flow across systems.
Key Responsibilities:
Design, develop, and maintain data pipelines and ETL workflows to collect, process, and transform large datasets.
Collaborate with data analysts, scientists, and business stakeholders to understand data requirements and deliver high-quality data solutions.
Implement and optimize data models, data lakes, and data warehouses.
Ensure data quality, integrity, and governance through validation, monitoring, and documentation.
Work with large-scale structured and unstructured data from multiple sources.
Optimize data processing for performance and scalability.
Deploy and manage data pipelines using CI/CD practices.
Troubleshoot data issues and enhance existing data workflows.
Required Skills & Qualifications:
Bachelor s or Master s degree in Computer Science, Information Systems, Engineering, or related field.
9+ years of experience in data engineering or a related role.
Strong experience with SQL and database technologies (e.g., Snowflake, Redshift, BigQuery, Synapse, or PostgreSQL).
Hands-on experience with ETL tools such as Informatica, Talend, AWS Glue, Azure Data Factory, or dbt.
Proficiency in Python, Scala, or Java for data processing.
Experience with big data frameworks (e.g., Apache Spark, Hadoop, Kafka).
Strong understanding of cloud data platforms (AWS, Azure, or Google Cloud Platform).
Familiarity with data warehousing concepts, star/snowflake schemas, and dimensional modeling.
Experience with version control (Git) and CI/CD pipelines.
Good understanding of data governance, security, and compliance standards.