Wednesday, October 19, 2011

EA Roadmapping

Strategy Roadmap Development




1) Develop a clear and unambiguous understanding of the current state
  • Business Objectives
  • Functional needs
  • High impact business processes
  • Current operating model
  • Cost and complexity drivers
  • Business and technical assets (Artifacts)


2) Define desired end state
  • Performance targets (Cash flow, Profitability, Growth, Customer intimacy)
  • Operating Model Improvements
  • Guiding principals


3) Conduct Gap Analysis
  • Gap closure strategies
  • Organizational
  • Functional
  • Architectural (technology)
  • Process
  • Reward or economic incentives

The following diagram illustrates a sample index or collection of the findings focused across the four architecture domains (Business, Information, Application, and Technology) related to the architecture. 





4) Prioritize
  • Actionable items
  • Relative business value
  • Technical complexity


5) Discover the Optimum Sequence


  • Dependencies of actionable items
  • Capacity for the organization to absorb change? 



6) Develop the Road Map


A Sample MDM Roadmap



Monday, October 17, 2011

Enterprise Information Architecture


Data Quality Management

Data Quality Management entails the establishment and deployment of roles, responsibilities, policies, and procedures concerning the acquisition, maintenance, dissemination, and disposition of data.

Data Quality Dimensions
The dimensions are distinct components that encompass the broader definition of data quality. The quality dimensions are defined as follows: 
  1. Completeness - Level of Data missing or unusable.
  2. Accuracy The accuracy dimension refers to how well information in or derived from the data holding reflects the reality it was designed to measure. 
  3. Timeliness - Timeliness refers primarily to how current or up to date the data is at the time of release, by measuring the gap between the end of the reference period to which the data pertains and the date on which the data becomes available to users. 
  4. Integrity - Degree of referenced data.
  5. Uniqueness - Level of non-duplicates.
  6. Conformity - Degree of data stored in a nonstandard format.
  7. Usability - Usability reflects the ease with which a data holding’s data may be understood and accessed. 
  8. Relevance Relevance reflects the degree to which a data holding meets the current and potential future needs of users. 

Information Life Cycle Phases

The Virtuous Cycle of Data Quality
Data quality management incorporates a “virtuous cycle” in which continuous analysis, observation, and improvement lead to overall improvement in the quality of organizational information across the board.




Five Fundamental Data Quality Management Practices

1. Data Quality Assessment - To understand the impact of poor data quality affects the business and to develop a business case for data quality management.
  • Business Impact Analysis
  • Data Profiling
  • Anomaly Review
  • Define Measure of Data Quality
  • Prepare Data Quality Assessment Report
2. Data Quality Measurement and Metrics - To analyse the assessment result and focus on the data elements that are deemed critical based on the selected business users’ needs.
  • Select Dimensions of Data Quality 
  • Define Data Quality Metrics
  • Define Data Validation Rules
  • Set Acceptability Thresholds
  • Devise Data Quality Scorecard
3.
Integrating Data Quality into the Application
 - To integrate data requirements analysis across the organisation and to engineer data quality into the System Development Life Cycle (SDLC).
  • Data Quality Requirements Analysis
  • Enhance the SDLC for Data Quality
  • Integrate Data Quality Improvement Methods
4. Operational Data Quality Improvement - Data Stewardship procedures are used to manage identified data quality rules and conformance to the acceptability thresholds.
  • Data Standards Management
  • Active Metadata Management
  • Data Quality Inspection and Monitoring
  • Define Data Quality Service Level Agreement
5. Data Quality Incident Management - To review the levels of acceptability, report, log, and track issues, and document the process for remediation and improvement.
  • Data Quality Issue Reporting and Tracking
  • Root Cause Analysis
  • Data Cleansing
  • Process Remediation

Summary of Data Quality Management

Plan
  1. Strategy & Approach
  2. People 
  3. Process
  4. Technology
  5. Priorities
Improve
  1. Analyze
  2. Measure
  3. Solution
  4. Execute
  5. Communicate
Control
  1. Focus
  2. Monitor
  3. Maintain
  4. Manage
  5. Communicate


The following points should be considered when launching a data quality management initiative: 
  1. Identify & Measure Data Quality: This is a key first step, as the ability to understand the up-front level of data quality will form the foundation of the business rules and processes that will be put in place. Without performing an upfront assessment, the ability to effectively implement a data quality strategy will be negatively impacted. From an ongoing perspective, the data quality assessment will allow an organization to see how the data quality procedures put in place have caused the quality of the data to improve.
  2. Define Data Quality Rules & Targets: Once the initial data quality assessment is complete, the second part of the analysis phase involves score-carding the results in order to put success criteria and metrics in place for the data quality management initiative. From an ongoing perspective, this phase will involve performing trend analyses on the data and the rules in place to ensure the data continues to conform to the rules put in place through the data quality management initiative.
  3. Design Quality Improvement Processes: This phase involves the manipulation of the data to align with the established business rules. Examples of potential improvements include: standardization, removing noise, alignment of product attributes, measures or classifications.
  4. Implement Quality Improvement Processes: Once the data has been standardized, the second phase of the enhancement process involves the identification of duplicate data and taking action based upon the business rules that have been identified. Since data quality is an iterative process, the rules to identify and address duplicate data will continue to evolve with an organization.
  5. Monitor Data Quality Versus Targets: The ability to monitor the data quality processes is critical, as it provides the organization with a quick snapshot of the health of the data within the organization. Through analysis of the data quality scorecard results, a data governance committee will have the information needed to confidently make additional modifications to the data quality strategies in place if needed. Conversely, the scorecards and trend analysis results can also provide the peace of mind that data quality is being effectively addressed within the organization.

Data Quality Management Challenges
Deploying a data quality management program is not easy; there are significant challenges that must be overcome.Some of the most significant reasons companies do not pursue a formal data quality management initiative include: 
  • No business unit or department feels it is responsible for the problem.
  • It requires cross-functional cooperation.  
  • It requires the organization to recognize that it has significant problems.
  • It requires discipline. 
  • It requires an investment of financial and human resources. 
  • It is perceived to be extremely manpower-intensive. 
  • The return on investment is often difficult to quantify. 













Enterprise Data Warehousing

Architecture Is Different than Methodology
It is important to recognize that data warehouse architecture identifies component parts, their characteristics, and the relationships among the parts, while methodology identifies the activities that have to be performed and their sequencing. Too often, the architecture and methodology terms are used interchangeably, which creates confusion. The architecture is the end product while a methodology is the process for developing an end product. But while architecture and methodology are different, they should be compatible. It is important to use a methodology that is consistent with the architecture that is being implemented.

Sometimes the hub and spoke architecture (e.g., Corporate Information Factory) is referred to as a top down approach and the bus architecture as bottom up. The reason for this is that the hub and spoke architecture places considerable emphasis on initially putting the infrastructure and processes in place to create an enterprise data warehouse and the bus architecture focuses on delivering a solution that addresses a current business need. These are methodologies rather than architectures because they describe development processes.

Over time, the top down and bottom up approaches have become increasingly similar. Advocates of the top down approach agree on the importance of developing incrementally and delivering early “wins.” The bottom-up proponents recognize the importance of having an enterprise plan for integrating the incrementally developed data marts. As a result, the two methodologies are not as different as many people believe.

Master Data Management

Metadata Management

Metadata is defined as “data about data”. This layer of information is created to help people use raw data as information.


Metadata management is the set of tools and processes that maintains a unified reference to the details on all data, information and knowledge existing within an organization. 


Metadata management is an important component of any Master Data Management (MDM) and Enterprise Data Warehousing (EDW) initiative. It is often overlooked and assumed to be taken care of automatically. 


Metadata management helps to lower the cost of ownership by documenting the entire end-to-end process for master data at the metadata level. Metadata can be documented everywhere from the source systems, MDM hub, data quality tools, data and business modelling tools, and integration tools.


There are 5 categories of Metadata:
  1. Business Metadata
  2. Navigational Metadata
  3. Structural Metadata
  4. Analytic Metadata
  5. Operational Metadata
There are 3 approaches to manage metadata: 
  1. Centralized Approach
  2. Distributed Approach
  3. Federated (Hybrid) Approach





Wednesday, October 12, 2011

Data Governance

Data governance can be initiated without having a MDM practice. However you can't have an effective MDM practice without data governance.


Summary of Data Governance


Plan:
  1. Strategy & Approach
  2. Charter & Process
  3. People Involvement
  4. Quality Management
  5. Control Targets

Implementation
  1. Implementation Plan
  2. Priorities
  3. Measurement
  4. Training & Readiness
  5. Communication

Govern
  1. Manage
  2. Maintain
  3. Improve
  4. Communicate
  5. Mature


Starting a Data Governance Model
  • Initial assessment and groundwork to drive a sound charter and implementation proposal
  • Business driven initiative sponsored at VP level
  • Led the program and initiative by Senior Director/Manager
  • Data Governance Council (DGC) to have sufficient influence to ensure that data standards, validation rules, and quality control expectations are actively involved in business process areas.

Ten mistakes to avoid
  1. Failing to define data governance
  2. Ready, shoot, aim: Failing to design data governance
  3. Prematurely launching a council
  4. Treating data governance as a project
  5. Ignoring existing steering committees
  6. Overlooking cultural considerations
  7. Prematurely pitching data governance
  8. Expecting too much from a sponsor
  9. Relying on the big bang
  10. Being ill-equipped to execute




High Level Data Governance Process Design and Implementation Approach
  • Planning and Design Phase
          1. Establishing the Charter
      • Distinguishing the Charter
      • Agreement on Mission and Objectives
      • Define Scope and Jurisdiction
      • Identify Roles and Responsibilities
      • Set Top Priorities
      • Committed Resources and Budgeting

          2. Policies, Standard, and Control
      • Define Key Policies, Big Rules, and Quality Standards
      • Establish Key Metrics, Monitors, and Improvement Targets
      • Identify Data Entry Points and Team Leads
      • Establish Quality and Service Level Agreement
      • Define Metadata Management Plan


  • Implementation Phase
          3. Process Readiness
      • Communication of Charter and Implementation
      • Readiness of Processes, Tools, and Baseline Measurements
      • Completion of Training and Readiness Plans with Core Teams and sub-teams
      • Launch the Process

              4. Implement
        • Conduct Regular Council Meetings
        • Manage Priorities, New Issues, and Requirements
        • Review Key Metrics and Performance Indicators
        • Communicate Status of Projects and Improvements
        • Keep Sub-Teams and Regional Teams Actively Engaged


              5. Maintain and Improve
        • Complete Key Improvement Projects
        • Identify and Address Negative Quality Trends
        • Monitor and Correct Negative Data Entry Process Behavior
        • Manage New Data Integration Requirements and Quality Impacts