Data Analyst

  • Arlington, VA
  • Posted 4 hours ago | Updated 3 hours ago

Overview

Hybrid
Depends on Experience
Contract - W2
Contract - Independent

Skills

ETL
Qlik Sense

Job Details

Job Title: Data Analyst (Onsite/Hybrid)

Location: Arlington, VA

Duration: 6+ Months

Remote/Onsite Hybrid Model

This position requires the individual to be on-site at Corporate Headquarters in Arlington, VA, and Dulles Airport Corporate office for at least four days a week (i.e., Monday, Tuesday, Wednesday, and Thursday). The remaining day could be remote. Work hours should be between 8:00 am and 6:00 pm.

This position does not require a vehicle to mobilize between campuses.

This position does not require AOA security access badging.

Brief description

The position is responsible for designing and implementing strategic data initiatives across the client, including enterprise data warehousing, metadata management, and machine learning solutions. The candidate must have extensive experience in data architecture, ETL development, and advanced analytics using Qlik Sense.

Job Description and Role

  • Designs and implements strategic data initiatives, such as enterprise data warehouse (EDW), master data, data governance, data quality, metadata management, and data marts.
  • Administers and monitors automated and manual data integration and Data/ETL jobs to verify execution and measure performance.
  • Designs and implements the Airports Authority-wide metadata and information management program, including development of enterprise conceptual, logical, and physical data models. Provides data architecture support to major application development initiatives in an agile environment.
  • Implement program standards and procedures to support data warehouse administration. Provides technical advice on data management and design to Application Development and Enterprise Architecture project teams, end-users, and business stakeholders.
  • Develops design and Data/Extract, Transform, and Load (ETL) coding of Source Dependent Extracts (Relational DB, APIs, flat files etc.,), Source Independent Loads, and Post-Load Processes from source to target systems for operational data stores and dimensional data warehouses into an in-memory or Massively Parallel Processing platform or a Structure Query Language (SQL) Server/Oracle Data Warehouse, using near real-time loads and Change Data Capture.
  • Data mining or extracting usable data from valuable data sources. Using machine learning tools to select features, create and optimize classifiers. Carrying out preprocessing of structured and unstructured data.
  • Enhancing data collection procedures to include all relevant information for developing analytic systems.
  • Processing, cleansing, and validating the integrity of data to be used for analysis.
  • Analyzing large amounts of information to find patterns and solutions. Developing prediction systems and machine learning algorithms.
  • Deliver a ML project from beginning to end, including understanding the business need, aggregating data, exploring data, building & validating predictive models, and deploying completed models with concept-drift monitoring and retraining to deliver business impact to the organization.
  • Research and implement novel ML approaches by using AI services, ML platforms and frameworks (e.g., TensorFlow, PyTorch, OpenNN, H2O.ai SparkML, SageMaker, scikit-learn, MXNet, Azure Synapse Analytics, Google BigQuery). Implementation includes descriptive, predictive, and prescriptive analytics.
  • Performs other duties as assigned.

Qualifications

  • Eight years of progressively responsible experience in data warehousing and integration, including experience applying data modeling techniques and ETL coding.
  • Knowledge of and skill in developing and implementing a corporate data architecture program with responsibility for enterprise conceptual/logical data modeling, data policies, standards and compliance monitoring, metadata mapping, data governance, and as-is/target data architecture to participate in strategic data initiatives.
  • Experience in an ML engineer or data scientist role building and deploying ML models or hands on experience developing deep learning models.
  • Knowledge of and skill in working with database management systems, such as Oracle, Redshift, MySQL, Snowflake and Microsoft Structured Query Language (SQL), and data warehouse solutions, and ETL tools (like Informatica Power Center, IICS) and reporting tools (like Qlik Sense, Tableau and PowerBI) to centrally manage and analyze data originating from disparate source systems.
  • Programming Skills knowledge of statistical programming languages like R, Python, and database query languages like SQL, Hive, Pig is desirable. Familiarity with Scala, Java, or C++ is an added advantage.
  • Statistics Good applied statistical skills, including knowledge of statistical tests, distributions, regression, maximum likelihood estimators, etc. Proficiency in statistics is essential for data-driven companies.
  • Machine Learning good knowledge of machine learning methods like k-Nearest Neighbors, Naive Bayes, SVM, Decision Forests.
  • Strong Math Skills (Multivariable Calculus and Linear Algebra) - understanding the fundamentals of Multivariable Calculus and Linear Algebra is important as they form the basis of a lot of predictive performance or algorithm optimization techniques.
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.