Sr Principal Data Analyst

Eden Prairie, MN, US • Posted 30+ days ago • Updated 3 days ago
Contract Independent
Contract W2
On-site
Depends on Experience
Company Branding Image
Fitment

Dice Job Match Score™

🤯 Applying directly to the forehead...

Job Details

Skills

Summary

Maintain existing applications and work on developing new applications supported by FDM on FIN360 platforms for various Finance & Accounting business teams. 

Team:
Teams has around 25 developers (Onshore and Offshore) primarily working on developing and maintaining finance and accounting data processing and reporting applications and automation for UHC controllership. The primary skills in team are SAS, Python, SQL skills for relational databases including Snowflake, Oracle, Teradata, SQL Server etc.

Position Background:
Maintain existing applications and work on developing new applications supported by Finance & Accounting business teams.

Top Responsibilities:
• Support the design, development, testing, and deployment of data analytics programs and processes for supporting various operational data stores using SAS, and Relational Databases.
• Collect, interpret, and aggregate data from traditional and non-traditional data sources for supporting programs and applications utilizing various data analytics purposes.
• Understand data requirements and business need to develop data tools such as dashboards and data visualizations.
• Use business intelligence, data visualization, query, analytic and statistical software to build solutions, perform analysis and interpret data.
• Solves moderately complex problems and can translate concepts into practice.
• Recognize problems and make recommendations for solutions.
• Works under minimal guidance and within tight deadlines for deliverables.
• Has an exploring mindset
• Effectively interact with business users for new projects, enhancement projects, and issue resolution including addressing issues reported regarding data and existing applications.
• Adopts a structured approach focusing on understanding needs of business users, documenting requirements, and clarifying expectations. This involves active listening, utilizing various techniques to gather information, and ensuring clear communication to address any uncertainties or inconsistencies.
• Create and document high level design, detail design, implementation and standard operating procedures guides.
• Perform Production support tasks including job monitoring, addressing production failures, perform data analysis, root cause analysis and issue resolutions.
• Design, develop, and maintain scalable ETL/ELT pipelines using ADF, Python, Apache Spark, and PySpark in Databricks.
• Develop batch and incremental data processing pipelines handling large-scale structured, semi-structured and unstructured datasets.
Implement optimized data transformation logic using Spark SQL and DataFrame APIs.
• Ensure pipelines follow enterprise data engineering best practices for performance, scalability, and maintainability.
• Implement reusable ingestion patterns and transformation templates aligned with enterprise architecture standards.
• Ensure compliance with enterprise metadata management, monitoring, and operational standards.
• Design and manage datasets stored in Apache Iceberg and Delta Lake.
Implement schema evolution, partitioning strategies, and version control for large datasets.
• Optimize data lake storage structures in Azure Data Lake Storage (ADLS) or AWS S3.
Develop scalable pipelines using Databricks notebooks, jobs, and clusters.
• Manage dataset governance and access controls using Unity Catalog.
Optimize Spark performance through partitioning, caching, and cluster tuning.
• Develop and schedule ETL pipelines using Apache Airflow.
Implement dependency management, monitoring, alerting, and failure recovery mechanisms.
• Build pipelines that integrate with Snowflake data warehouse.
• Optimize transformations and data loading using Snowflake SQL and staging techniques.
• Design efficient data models for analytics and reporting.
• Support migration of legacy SAS pipelines to modern Spark-based frameworks and Databricks where applicable.
• Use Unix/Linux commands for common tasks and shell scripting to automate data engineering workflows.
• Support CI/CD deployment processes for ETL pipelines.
• Implement logging, auditing, and monitoring for production pipelines.
Work with data architects, analysts, and business stakeholders to gather requirements and deliver data solutions.
• Participate in design reviews, architecture discussions, and code reviews.
• Mentor junior data engineers and provide technical guidance.
• Be the SME for DBX and do knowledge training for team

Ideal Candidate:
The ideal candidate will have strong experience working with Databricks (Lakehouse, Delta Lake, Workflows, Medallion Architecture, Apache SPARK, Unity Catalog, Delta Sharing, Notebooks, SQL, GIT), PySpark, Python, Snowflake, and ADF frameworks.
This role focuses on building and optimizing large-scale data pipelines using ADF, Apache Iceberg, Delta Lake, cloud data lakes (ADLS/S3), and workflow orchestration tools like Airflow. The analyst will work closely with data architects, and platform teams to build reliable and governed data solutions aligned with enterprise standards.

Must Have:
• Bachelor’s degree in computer science, Computer Applications, Analytics, Data Science, or Information Technology.
• 8+ years of experience in ETL / Data Engineering
• 8+ years of experience with programming using Python
• 8+ years of experience working in Unix/Linux environments
• 8+ years of experience writing Shell scripts
• 6+ years of experience with Databricks ecosystem including Lakehouse, Delta Lake, Workflows, Medallion Architecture, Apache SPARK, PySpark, Unity Catalog, Delta Sharing, Notebooks, SQL, GIT.
• 6+ years of experience with ADF
• 6+ years of experience working with large enterprise datasets
• 4+ years of experience with Snowflake
• Strong analytical and troubleshooting skills
• Excellent communication and collaboration abilities
• Ability to work independently and mentor junior analysts
• Strong documentation and design skills
• Strong SQL skills
• Experience implementing governance using Unity Catalog
• Experience working with Apache Iceberg or other open table formats
• Experience working with Azure Data Lake Storage (ADLS) or AWS S3
• Understanding of cloud data lake architecture
• Hands-on experience with Apache Airflow
• Experience developing pipelines for Snowflake
• Strong understanding of SAS programming, SAS Data step, SAS Macros, PROC SQL

Nice To Have:
• Experience migrating SAS ETL pipelines to Spark and Databricks
• Knowledge of data governance frameworks
• Healthcare experience

Dress Code:
Business casuals


Horizontal facilitates valuable and productive conversations between you and potential employers. We can assist you in growing your career by partnering you with employers that offer challenging assignments. For those that join the team, we offer competitive compensation and benefits including medical, dental, vision, and retirement. Check out all we have to offer and how you can become part of the Horizontal Talent Team. The pay range for this role is $28 - $58 per hour. This is not a guarantee of compensation, as final offer amount may vary based on factors including but not limited to experience and geographic location.

Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
  • Dice Id: 10123528
  • Position Id: 73445
  • Posted 30+ days ago

Company Info

About Horizontal Talent

Horizontal Integration offers two service lines. Our staffing agency places IT, digital and creative talent in both contract and permanent roles. Our digital agency provides digital marketing strategy and technology consulting on a project basis. Many of our clients take advantage of both service lines.

We are a certified Minority Business Enterprise, with more than 200 employees across our offices in Minneapolis, Denver, Dallas and Vadodara, India. We ve been consistently growing and profitable since the company was founded in 2003.

Honors:
Recognized as a "2009 Best Place to Work" by Minneapolis/St.Paul Business Journal.

Fact:
Our people and our organization are deeply rooted in IT and interactive marketing.

Horizontal is proud to be an Equal Opportunity and Affirmative Action Employer. We seek to provide employment opportunities to talented, qualified candidates regardless of race, color, sex/gender including gender identity and/or expression, national origin, religion, sexual orientation, disability, marital status, citizen status, veteran status, or any other protected classification under federal, state or local law.

In addition, Horizontal will provide reasonable accommodations for qualified individuals with disabilities. If you need to request a reasonable accommodation in order to complete the application or interview process, please contact us. All applicants applying must be legally authorized to work in the country of employment.

About_Company_OneAbout_Company_Two
Create job alert
Set job alertNever miss an opportunity! Create an alert based on the job you applied for.

Similar Jobs

Cleveland, Ohio

Yesterday

Easy Apply

Contract

Depends on Experience

Dallas, Texas

4d ago

Easy Apply

Contract

Depends on Experience

Los Gatos, California

Yesterday

Easy Apply

Contract

Depends on Experience

Search all similar jobs