Required Education
A 4-year degree and/or Master s degree from an accredited college or university, or equivalent experience.
Technical Skills (Must-Have)
5 7 years experience in data management, data engineering, or data operations (data design, data quality, metadata, governance, etc.)
Experience with development and delivery of microservices using serverless AWS services (S3, RDS, Aurora, DynamoDB, Lambda, SNS, SQS, Kinesis, IAM)
4+ years of experience in a cloud environment (AWS, Snowflake)
Strong SQL and Python
Soft Skills (Must-Have)
Verbal and written communication skills with ability to clearly communicate complex technical ideas
Problem solving skills, customer service, and interpersonal skills
Ability to work collaboratively in a complex, rapidly changing, and culturally diverse environment
Desired Skills (Nice to Have)
Strong AWS experience, including SageMaker, S3, RDS, CloudWatch and related services
Strong Snowflake experience
Strong SQL and Python development skillset
Understanding of logical data domains, primarily Customer & Equipment Domains
Experience in Data Operations, Tier 2 Support, or comparable Data Engineering Support role
Strong knowledge of end-to-end data lifecycle across traditional data warehouses, relational databases, operational data stores, business intelligence reporting, and big data analytics
Knowledge of data technology products and components for Big Data and Cloud (AWS, Data Lakes, and similar)
Job Responsibilities
Perform all necessary data related tasks including data design, data quality, data triage, data governance, or data architecture SQL, Snowflake, AWS
Develop break/fix solutions and address root causes in data pipeline implementation/code Python, AWS
Develop scripts and automation tools to better detect and correct data issues
Develop monitoring and alerting capabilities to proactively detect data issues
Work directly on complex application/technical problem identification and resolution, including responding to off-shift and weekend support calls
Identify, investigate, and obtain resolution commitments for platform and data issues to maintain and improve quality and performance of assigned digital product data
Issue Identification: Review reports from customers, dealers, industry representatives, and subsidiaries
Issue Investigation: Perform statistical analysis, data triage, and infrastructure problem-solving
Issue Resolution: Identify root causes, create SageMaker scripts to fix data, and perform break/fix tasks on data pipeline code
Communicate with end users and internal customers to help direct development, debugging, and testing of application software for accuracy, integrity, interoperability, and completeness
Participate in technical sync-ups and meetings with internal team including US and offshore teams
Liaise with designers, engineers, and support teams to improve data pipeline performance and reliability
Perform other job duties as assigned by Caterpillar management from time to time