Overview
Skills
Job Details
Experience: 7 10 Years
Employment: W2 Only (H1B Transfer / EAD)
We are looking for an experienced Data Engineer to build, maintain, and optimize large-scale data pipelines and cloud-based data solutions. The ideal candidate should have strong hands-on experience with PySpark, SQL, and modern cloud platforms.
Key Responsibilities:-
Design, build, and manage scalable ETL/ELT pipelines for large datasets.
-
Develop data processing workflows using PySpark, Spark, and distributed systems.
-
Optimize data ingestion, transformation, and storage for high performance.
-
Work on cloud-native data solutions using AWS/Azure/Google Cloud Platform.
-
Collaborate with data scientists, analysts, and application teams to deliver reliable data products.
-
Ensure data quality, governance, and security standards.
-
Implement monitoring, logging, and automation for data pipelines.
-
Strong experience with PySpark, Spark, and distributed data processing.
-
Expertise in SQL and performance tuning.
-
Experience with cloud services (AWS Glue, EMR, Lambda / Azure Data Factory / Google Cloud Platform DataFlow).
-
Hands-on experience with data warehousing concepts (Snowflake, Redshift, BigQuery, etc.).
-
Strong understanding of ETL/ELT design and best practices.
-
Familiarity with version control, CI/CD, and DevOps workflows.
-
Experience with Databricks.
-
Experience with Kafka, streaming pipelines, or real-time data processing.
-
Knowledge of Python data libraries (Pandas, NumPy).