YoVDO

DataOps Methodology

Offered By: IBM via Coursera

Tags

Data Science Courses Artificial Intelligence Courses Big Data Courses Data Engineering Courses Data Pipelines Courses DataOps Courses

Course Description

Overview

DataOps is defined by Gartner as "a collaborative data management practice focused on improving the communication, integration and automation of data flows between data managers and consumers across an organization. Much like DevOps, DataOps is not a rigid dogma, but a principles-based practice influencing how data can be provided and updated to meet the need of the organization’s data consumers.” The DataOps Methodology is designed to enable an organization to utilize a repeatable process to build and deploy analytics and data pipelines. By following data governance and model management practices they can deliver high-quality enterprise data to enable AI. Successful implementation of this methodology allows an organization to know, trust and use data to drive value. In the DataOps Methodology course you will learn about best practices for defining a repeatable and business-oriented framework to provide delivery of trusted data. This course is part of the Data Engineering Specialization which provides learners with the foundational skills required to be a Data Engineer.

Syllabus

  • Establish DataOps - Prepare for operation
    • In this module you will learn the fundamentals of a DataOps approach. You will learn about the people who are involved in defining data, curating it for use by a wide variety of data consumers, and how they can work together to deliver data for a specific purpose:
  • Establish DataOps – Optimize for operation
    • In this lesson you will learn the fundamentals of a DataOps approach. You will learn about how the DataOps team works together in defining the business value of the work they undertake to be able to clearly articulate the value they bring to the wider organization:
  • Iterate DataOps - Know your data
    • In this lesson you will learn about the capabilities that you will need to use to understand the data in repositories across an organization. Data discovery is most appropriately employed when the scale of available data is too vast to devise a manual approach or where there has been institutional loss of data cataloging. It utilizes various techniques to programmatically recognize semantics and patterns in data. It is a key aspect of identifying and locating sensitive or regulated data to adequately protect it, although in general, knowing what stored data means unlocks its potential for use in analytics. Data Classification provides a higher level of semantic enrichment, enabling the organization to raise data understanding from technical metadata to a business understanding, further helping to discover the overlap between multiple sources of data according to the information that they contain:
  • Iterate DataOps – Trust your data
    • In this lesson you will learn that understanding data semantics helps data consumers to know what is available for consumption, but it does not provide any guidance on how good that data is. This module is all about trust, how reliable a data source can be in providing high fidelity data that can be used to drive key strategic decisions, and whether that data should be accessible to those who want to use it; whether the data consumer is permitted to see and use it. This module will address the common dimensions of data quality, how to both detect and remediate poor data quality. And it will look at enforcing the many policies that are needed around data quality, not least the need to respect an individual’s wishes and rights around how their data is used:
  • Iterate DataOps – Use your data
    • In this lesson you will learn that providing useful data in a catalog can often necessitate some transformation of that data. Modifying original data can optimize data ingestion in various use-cases, such as combining multiple data sets, consolidating multiple transaction summaries, or manipulating non-standard data to conform to international standards. This module will examine the choices for data preparation, how visualization can be used to facilitate the human understanding of the data and what needs to be changed, and the various options for single use, optimization of data workflows and ensuring the regular production of transformations for operational use. Furthermore, this module will show you how to plan and implement the data movement and integration tasks that are required to support a business use case. The module is based on a real-world data movement and integration project required to support implementation of an AI-based SaaS analytical system for supply chain management running in the Google cloud. The module will cover the major topics that need to be addressed to complete a data movement and integration project successfully:
  • Improve DataOps
    • In this lesson you will learn about evaluating the last data sprint, observe what worked and what did not, and make recommendations on how the next iteration could be improved.
  • Summary & Final Exam

Taught by

Elaine Hanley

Tags

Related Courses

Hands-On with Dataflow
A Cloud Guru
Azure Data Engineer con Databricks y Azure Data Factory
Coursera Project Network via Coursera
Data Integration with Microsoft Azure Data Factory
Microsoft via Coursera
Azure Data Factory : Implement SCD Type 1
Coursera Project Network via Coursera
MLOps1 (Azure): Deploying AI & ML Models in Production using Microsoft Azure Machine Learning
statistics.com via edX