0% found this document useful (0 votes)
8 views

Tech Proposal

Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views

Tech Proposal

Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 7

Project Objectives

Client is requesting the implementation services proposal for Data Governance tools.

This will enhance the Client’s vision towards future Data governance, covering the
following purposes:

- Data Quality

- Data Governance

- Metadata management

Modern Data Platform Architecture


On the scalability of the platform
The next sections will show how Client will be able to leverage the flexibility of the
platform to shape its architecture and make it change over time.

A few points worth noting on that topic:

100% reusability of the assets


In the scenario described below, the deployment of the processing engine changes
quite a lot. It is important to note that no change will be required on the
Integration jobs themselves. No rework will be required as long as the connectivity to
the systems is still available.

100% flexibility of deployment


There is no hard rules or guidance in terms of designing the architecture:

Client might start with 1 Agent covering all their needs and stay in a single location
by adding more agents to that central point over time.

Cloud / Hybrid future proof


Even though the Secure Agents are placed in the Data Canters managed by Client in
the following scenario, it is worth noting that Client will have the flexibility of
provisioning Secure Agents in the Cloud (Public, Private Cloud) at any point in time,
and assign some workload to the Agents in the Cloud.

This could be useful to handle future Cloud-to-Cloud integration jobs that do not
require for internal data to be accessed.

Single Agent
Initially, one Secure Agent will be deployed in a central location and will handle all
the processing, across Client Data Centers and applications. This Secure Agent will be
connected to all the applications.

This setup enables Client to start quickly addressing their Integration needs without
having to engage in a massive architecture change or implementation project. The
platform will feature all the components described in this document and will not
compromise on capability.

Scope of Work
The project implementation scope will be covering integrated, end-to-end data
governance solution with comprehensive data quality, business glossary, metadata
management and landscape data and process governance to provide the capabilities
to provide processes which manage and improve data for the benefit of all
stakeholders.

Below sections explains the different tasks and activities to be carried out as part of
the project and solution delivery.

Technical and Business Analysis


Implementation team will do the required technical and functional analysis to cover
the below areas:

 Identifying key functional and technical stakeholders.


 Source system technical analysis to understand and collect the different
source system technical details.
 Source system model analysis to understand the underlying data model used
to host the information.
 Collecting and analyzing the existing data quality issues and required data
validation required to be applied.
 Identifying the mapping logic between the source system model and target
master model
 Identifying the list automatic data standardization and cleansing actions to be
automatically applied by the system.
 Identifying the list of matching criteria/conditions to detect duplicates
customers and the actions to be applied in case of potential matching
situation by either merging the records or keeping them for further
investigations.
 Documenting the output of the analysis phase and get it discussed, finalized,
and approved by the different stakeholder to be the basis for the solution
implementation, delivery, and acceptance.

Metadata Catalog Implementation


Implementation team will do the following implementation activities:

 Integrate between source systems, across the enterprise to scan data life
cycle and provide the visibility for a full technical metadata in the tool.
 In-scope Source system configuration and metadata integration
 Metadata extraction to build the end-to-end lineage.
 Data domain and classification discovery configuration
 Build the stakeholders and data owners community and leverage the
collaboration between catalog users.
 Enable sources owner notification and enable metadata change history.
 Data Quality to show rules and scorecards linked to the physical data
elements.
 Implement NDMO’s “Data Catalog and Metadata” controls and specifications.

DG Model and Business Glossary Definition


Implementation team will do the following implementation activities:
 Collect the current Data Governance documentation like framework,
operationalization model, business glossary from Client.
 Analyze the document and work on building templates to be imported to
Informatica IDMC.
 Build data governance organization and operation model setup (Data
Stewardship).
 Define DG policies, standards, templates, workflows, processes, and
procedures as collected from Client.
 Developing data governance workflows to manage the above assets (e.g.,
approval workflow, creating of new DQ rule, modifying existing policy, etc.)
 Import business glossary.
 Defining business rules and link to business terms.
 Link all above assets together and build the lineage.

Compliance with NDMO Regulations


The proposed solution should be fully complaint with NDMO regulations including all
data domains such as Data Quality, Data Catalog and Metadata, Data Dictionary,
Data Governance.

In the following sub-sections, we will give more details about how the solution
compliant with two of those data domains as examples:

1) Data Catalog and Metadata Domain

Data Catalog and Metadata domain comprises of 6 controls and 20 specifications.


This domain focuses on enabling an effective access to high quality integrated
metadata. The access to metadata is supported by use of the Data Catalog
automated tool acting as the single point of reference to the organizations' metadata.

Vendor team to stick and comply with the NDMO regulations and specification.
2) Data Quality Domain

Data Quality domain comprises of 4 controls and 13 specifications. This domain Data
Quality focuses on the improvement of the quality of the organization data, ensuring
that data is fit for purpose based on consumers' requirements.

Vendor team to stick and comply with the NDMO regulations and specification.

Data Profiling and Quality Implementation


Vendor team will do the required data quality, profiling, and cleansing activities to
ensure that data is enhanced and enriches as much as possible and to achieve the
highest level of matching accuracy.
The below activities will be delivered as part of the quality implementation scope:

 Source System Data Profiling for the data elements.


 Data profiling will be executed against the key in scope source tables hosting
the business to analyze the data formats, anomalies, frequencies, and
patterns.
 Data profiling will describe and analyze the source data.

Based on the business analysis output and the data profiling results source system
attributes and details suffering data quality issues will be identified.

Solution Testing
Based on the technical and business analysis and the agreed on functional and
technical requirements a detailed test cases will be designed executed to cover the
below main testing aspects:

Unit Testing: A detailed technical oriented test case and test scenario
executed against the developed jobs and processes to make sure that the
solution components are valid and match the execution expectations.

Business Acceptance Testing: Business test cases to be design and


executed to make sure that quality profiling rules, metadata lineage, and data
classifications are valid and meets the business expectations. Implementation
Team to work with Business Users and Data Stewards to conduct Business
Acceptance Testing.

As a result of the above explained testing phase, a detailed report will be shared with
client explaining test case design plus a report showing the execution status of each
of those cases.

Knowledge Transfer and Enablement Workshops


Solution documentation, analysis output, operations guide, and technical know-how
will be handed over to Client team along with practical onsite handover session to
make sure that Client team are able to manage, operate, and expand solution to
cover any future requirements.

Implementation Assumptions
 A maximum of 30 processes.
 A maximum of 30 policies.
 A maximum 150 unique CDEs.
 15 in-scope EDC resources.
 Data profiling will be performed for up to 100 tables.
 A maximum 50 basic rules. A basic rule is a rule that belongs to one field, and
it’s automatically discovered by the tool using basic profiling (Field is not null,
Format of email is not accurate, inconsistent data type, etc.)
 A maximum 50 complex rules (which need complex calculations).
 Data Remediation and Cleansing will be in scope.

Knowledge Transfer and Training Assumptions


 All training sessions will be conducted either at Client premises or online,
depending on the customer's preference.
 For KT and Training that will be availed; Vendor will conduct the training
based on “Train the Trainer” concept.

Project Timeline
 High Level Timeline should be proposed with phases and planned activities.

You might also like