Overview
Hybrid
$55 - $65
Contract - W2
Contract - 12 Month(s)
Skills
AWS
Job Details
Role: Lead/ Sr. AWS Data Engineer
Address: Houston, TX (Hybrid)
Duration: Long Term
JOB SUMMARY :
As a Senior Data Engineer within Data and Analytics organization, you will play a crucial role in architecting, implementing, and managing robust, scalable data infrastructure. This position demands a blend of systems engineering, data integration, and data analytics skills to enhance TFG s data capabilities, supporting advanced analytics, machine learning projects, and real-time data processing needs.
QUALIFICATIONS:
- Bachelor's or master s degree in computer science, MIS, or other business discipline or equivalent combination of education and/or experience.
- 7+ years of experience in data engineering, with a proven track record in designing and operating large-scale data pipelines and architectures.
- Expertise in developing ETL/ELT workflows.
- Fluent in Infrastructure-as-code paradigms (e.g. Terraform)
- Comprehensive knowledge of platforms and services like Databricks, Dataiku, and AWS native
- data offerings
- Solid experience with big data technologies (Databricks, Apache Spark, Hadoop, Kafka) and cloud services (AWS, Azure) related to data processing and storage.
- Strong experience in AWS (preferred) and Azure cloud services, with hands-on experience in integrating cloud storage and compute services with Databricks.
- Proficient in SQL and programming languages relevant to data engineering (Python, Java, Scala).
- Hands on RDBMS experience (data modeling, analysis, programming, stored procedures)
- Familiarity with machine learning model deployment and management practices is a plus.
- Fluent with CI/CD workflows & automation
- Strong communication skills, capable of collaborating effectively across technical and non-technical teams.
CERTIFICATES, LICENSES, REGISTRATIONS:
Preferred: Databricks Certified Associate Developer for Apache Spark, AWS Certified Solutions Architect, or other relevant certifications.
ESSENTIAL FUNCTIONS:
- Design and implement scalable and reliable data pipelines to ingest, process, and store diverse data at scale, using technologies such as Databricks, Apache Spark, Hadoop, and Kafka.
- Work within cloud environments like AWS (preferred) or Azure to leverage services including but not limited to EC2, RDS, S3, Lambda, and Azure Data Lake for efficient data handling and processing.
- Develop and optimize data models and storage solutions (SQL, NoSQL, Data Lakes) to support operational and analytical applications, ensuring data quality and accessibility.
- Utilize ETL tools and frameworks (e.g., Apache Airflow, Talend) to automate data workflows, ensuring efficient data integration and timely availability of data for analytics.
- Implement pipelines with a high degree of automation
- Collaborate closely with data scientists, providing the data infrastructure and tools needed for complex analytical models, leveraging Python or R for data processing scripts.
- Ensure compliance with data governance and security policies, implementing best practices in data encryption, masking, and access controls within a cloud environment.
- Monitor and troubleshoot data pipelines and databases for performance issues, applying tuning techniques to optimize data access and throughput.
- Stay abreast of emerging technologies and methodologies in data engineering, advocating for and implementing improvements to the data ecosystem.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.