SlideShare a Scribd company logo
Multidimensional Challenges and the
Impact of Test Data Management
To deliver benefits such as reducing test windows, improving test
quality, optimizing data processing/storage costs and maintaining
complex data model/relationships, TDM must be embedded equally
within an IT organization’s QA and development teams.
Executive Summary
Managing costs without compromising quality is
the latest mantra of IT organizations. End-to-end
test data management (TDM),1
as against mere
provisioning of data for test execution, has far-
reaching potential and can help rein in operation-
al and processing costs.
Over the past few years organizations have
focused more on test data management, realizing
its potential impact on multiple aspects of the
quality assurance (QA) process, such as testing
window reduction, improvement in quality of
testing, optimizing data processing and storage
costs and maintaining complex data models and
relationships.
To deliver these multidimensional impacts, TDM
must be embedded within quality assurance and
development teams, alike. Moreover, TDM must
also intersect with other disciplines such as
release management, environment management
and data security (see Figure 1). TDM acts as a
conduit between development and QA teams.
It also needs to interface with the disciplines
shown in the diagram to deliver the larger
impacts discussed in the prior paragraph. These
interfaces with the various disciplines increase
the complexity that must be overcome.
Why is TDM so pertinent? Our survey of the top
250 lines of business served by us makes clear
why:2
•	80% of IT organizations acknowledge that
they either have no tools to query/retrieve the
right data or require improvements to their
in-house tools.
•	61% of IT organizations do not have any TDM
tools.
•	50% of IT organizations do not have a data
reusability strategy in place.
•	40% of IT organizations face data integrity
issues in their test environments.
•	38% of IT organizations have sensitive
production data that can be accessed by
in-house testing or QA teams.
This white paper explores various dimensions of
potential benefits organizations can derive from
effective TDM and what is needed for QA teams to
leverage these elements to achieve critical multi-
dimensional benefits.
• Cognizant 20-20 Insights
cognizant 20-20 insights | july 2014
2
Unleash the Potential of Test Data
Management
By maximizing the potential of TDM, IT organi-
zations can address the above challenges and
deliver several valuable benefits:
•	Increased quality of test conditions and the
leveraging of data integration tools. Based on
our survey, nearly two-thirds of IT organizations
do not have any TDM tools. Many organizations
face a common challenge of ensuring that the
test environment replicates the exact patterns
of production. Some data integration and data
profiling and extraction tools are built with
features that can help identify data patterns in
the production stage. Such tools can be used
to capture and mimic similar patterns in the
test environments. This can help elevate test
data quality to the next level. The tools result
in increased business confidence regarding
test coverage and, eventually, in higher quality
applications.
•	Optimized storage and processing costs.
Typically, the size of a test environment
increases as additional projects are brought
online. This results in increasing storage costs.
Bulky test environments also add indirectly to
processing expenses such as archival and batch
processing costs, because larger volumes of
data are processed.
•	An efficient TDM framework needs to
consider not only making the data available
to testers but also must minimize the
volumes of data in a test environment. A test
data strategy needs to consider a long-term
pipeline of activities and arrive at an optimal
test bed that does not compromise data avail-
ability. In addition, the test data strategy must
be supplemented by an optimal archival,
refresh or restore plan.
•	A shorter batch process window. This is a
natural extension of the resulting impact of
the previous challenge. Not only does it cost
more to process unnecessary data, but also the
batch process itself takes longer than needed.
Today’s TDM approach tries to minimize the
batch window.
•	Improved compliance to data protection
policies and regulatory requirements. Our
survey also revealed that over one-third of
IT organizations let in-house testing teams
access sensitive production data for their
testing needs. IT organizations require test
data architects to devise a solution to comply
with their data protection policy, while simul-
taneously retaining production-like quality test
data. Persistent and dynamic data masking
solutions are highly sought-after, especially in
the context of the dynamic environment of reg-
ulations. This is more pronounced in domains
such as healthcare, banking and financial
services and insurance where data privacy is a
highly sensitive concern.
Also, a small percentage of IT organiza-
tions quoted data privacy as a roadblock for
offshoring. Data masking solutions can enable
cognizant 20-20 insights
TDM’s Connections with Other
Organizational Functions
Figure 1
Tools Strategy
Release
Management
Environment
Management
Program
Management &
Governance
Test Data
Management
Knowledge
Management
Data Security
Office
Tools StTT rategyrr
Release
Management
Environment
Management
Programrr
Management &
Governance
Test DataTT
Management
Knowledge
Management
Data Security
Office
Test Data
Management
Program
Management
& Governance
Data Security
Office
Release
Management
Knowledge
Management
Tool
Strategy
Environment
Management
3cognizant 20-20 insights
offshore test execution in compliance with the
policy and relevant regulatory requirements.
The key here is to ensure data integrity of the
masked data, without exposing sensitive data.
There are also instances in which a synthetic
data generation solution has been leveraged
to overcome data privacy concerns, paving the
way for offshoring of TDM.
•	Better technical support to black-box
testing teams for knowledge management.
Nearly half of IT organizations surveyed face
data integrity issues in their test environments.
Test data teams can act as a bridge between
the white-box develop-
ment teams and black-box
functional testing teams.
While the focus of functional
testing teams is to treat the
application under test as
a black box, the setup of
the prerequisite test data
requires a deep understand-
ing of the underlying data
structure and flow. Thanks
to test data management
tools, the knowledge about
numerous business rules
and the interactions among
interfaces can be captured in
a more elaborate and robust
manner. This can lead to
enhanced knowledge quality,
better retention and easier
maintenance.
•	Availability of more in-process efficien-
cies: data fencing and formal data request
systems. Nearly half of IT organizations polled
spend at least 10% of their testing efforts on
test data preparation. Improvements can also
be made in terms of data fencing. That is, the
ways in which testing teams earmark their data
and avoid data loss and overrides. This process
becomes complex when the testing team’s size
is large (more than 200) and when it is involved
in multiple and parallel projects that can impact
the same module within an application.
Efficiencies can also be extracted based on how
testing teams create and handle their test data
requests. More IT organizations now realize the
impact of a speedy test data request and provi-
sioning system. Moreover, they are also willing
to invest in one. Some IT organizations are even
deploying commercial request management
systems with integrated work flows.
•	Improved test coverage. It is commonly
assumed that production data, though
present in large volumes, only covers about
30% of possible functionalities or scenarios.
To maximize test coverage, testers need to
condition the data to ensure testing of all
possible scenarios. Here, a synthetic data
generation solution can play a crucial role in
creating the much-needed variety of data
with desired distribution patterns. It can also
be extended to improve coverage of perfor-
mance testing by generating volumes that
are sometimes many times bigger than the
current production size. Nearly two-thirds of
IT organizations surveyed have synthetic data
generation needs.
Also, as part of migration/upgrade scenarios,
when large volumes of data are moved from
source to target systems, it is only a small
portion of target data that is typically tested
manually with the help of scripts and queries.
The low coverage of testing leads to the risk of
untested data at the target. Organizations can
leverage validation-related products, either
those in-house or the ones available commer-
cially, to ensure 100% coverage.
•	Optimized data provisioning time: TDM
accelerators. More than two-thirds of the IT
organizations surveyed acknowledge that they
have no tools to either query or retrieve the
right data. To capture the business rules and
understand the information data structures
more efficiently, accelerators can be set
up that readily contain this information for
packages such as SAP or Siebel. Accelerators
are also available for regulatory compliances
such as Payment Card Industry’s Data Security
Standards (PCI DSS) and Protected Health
Information (PHI). These accelerators not only
cut down the configuration time and effort,
but also improve the consistency of how these
rules are applied.
•	Maximized utilization of knowledge and
skills. A small percentage of IT organiza-
tions surveyed have centralized their test
data team services. As such, the benefits of
a managed services model can be applied to
TDM. TDM requires skills in niche areas such
as extract-transform-load (ETL) and online
analytical processing (OLAP)/online transac-
tion processing (OLTP) which may not be nec-
essarily available with functional testing teams.
Having such high-skilled resources functioning
as TDM specialists can help achieve greater
efficiency.
Thanks to test data
management tools,
the knowledge
about numerous
business rules and
the interactions
among interfaces
can be captured in a
more elaborate and
robust manner. This
can lead to enhanced
knowledge quality,
better retention and
easier maintenance.
cognizant 20-20 insights 4
Managed Services TDM in Action
Figure 2
Business Outcomes: Case in Point
Figure 2 highlights engagements where we helped
IT organizations embrace, deploy and benefit from
TDM delivered via a managed services approach.
Looking Ahead
A comprehensive and end-to-end TDM strategy
that addresses the multidimensional challenges
IT organizations face can deliver multifaceted
benefits to enterprises across industries. It can
help reduce operational and processing costs,
while simultaneously compressing the time
needed to achieve the cost savings. It can also
eliminate the risk of exposing sensitive data to
nonproduction users while also improving test
coverage.
TDM, in our experience, furthers better knowledge
retention and management with QA organiza-
tions, and higher data reuse. Applying a managed
services model to TDM provides key benefits to
IT organizations, including wider access to niche
skills and optimal resource utilization. TDM has a
far wider potential than mere provisioning of data
for test execution. The advent of big vendors such
as Informatica and IBM from the data integration
space has only advanced TDM’s ability to deliver
large-scale benefits, while shrinking processing
windows.
Challenges
Improvements and
Changes
Results
A Multibillion Banking
& Financial Services
Organization in the UK
Manually intensive test
data provisioning. Lack
of a formal data request
tracking mechanism.
Lack of consolidated
knowledge on data
conditions.
Centralized team with
internal ownership of
knowledge management.
API-based automation
(data extraction and data
generation). Centralized
data request tracking:
SLAs based on the
turnaround time and data
suitability.
Annual savings of
$2.5 million in test
data provisioning.
40% increase in
productivity.
A U.S.-Based Retail
Conglomerate
Complex application
design. Sensitive
production data used in
test environment. Full
load of data each time,
making it longer than
required.
Deployed the TD MaximTM
framework3
and configured
complex rules removing
manual efforts. TD Maxim
used for applying filters to
get just enough data.
75% reduction in
TDM efforts. Gained
data privacy in
test environments.
Reduction of data
volume by 80%.
A Leading U.S.-Based
Healthcare
Organization
Sensitive protected
health information (PHI)
was unprotected in
the test environment.
Inconsistent and effort-
intensive methods for
generating test data.
Deployed the TD Maxim
framework with necessary
masking algorithms and
policies. The framework
was also utilized for
synthetic data generation.
Achieved legal
compliance as per
PHI regulations.
Consistent and
streamlined data
generation process.
A Global
Biopharmaceuticals
Giant
A large-scale program
facing huge schedule
slippage due to
unavailability of test
data. Simple stubbing
ruled out because of the
need for high-quality
test data.
Deployed the TD Maxim
framework for synthetic
data generation, which
ensures data integrity
rules across all interfaces.
Averted risk of
derailment of the
project.
About Cognizant
Cognizant (NASDAQ: CTSH) is a leading provider of information technology, consulting, and business process out-
sourcing services, dedicated to helping the world’s leading companies build stronger businesses. Headquartered in
Teaneck, New Jersey (U.S.), Cognizant combines a passion for client satisfaction, technology innovation, deep industry
and business process expertise, and a global, collaborative workforce that embodies the future of work. With over 75
development and delivery centers worldwide and approximately 178,600 employees as of March 31, 2014, Cognizant
is a member of the NASDAQ-100, the S&P 500, the Forbes Global 2000, and the Fortune 500 and is ranked among
the top performing and fastest growing companies in the world. Visit us online at www.cognizant.com or follow us on
Twitter: Cognizant.
World Headquarters
500 Frank W. Burr Blvd.
Teaneck, NJ 07666 USA
Phone: +1 201 801 0233
Fax: +1 201 801 0243
Toll Free: +1 888 937 3277
Email: inquiry@cognizant.com
European Headquarters
1 Kingdom Street
Paddington Central
London W2 6BD
Phone: +44 (0) 20 7297 7600
Fax: +44 (0) 20 7121 0102
Email: infouk@cognizant.com
India Operations Headquarters
#5/535, Old Mahabalipuram Road
Okkiyam Pettai, Thoraipakkam
Chennai, 600 096 India
Phone: +91 (0) 44 4209 6000
Fax: +91 (0) 44 4209 6060
Email: inquiryindia@cognizant.com
­­© Copyright 2014, Cognizant. All rights reserved. No part of this document may be reproduced, stored in a retrieval system, transmitted in any form or by any
means, electronic, mechanical, photocopying, recording, or otherwise, without the express written permission from Cognizant. The information contained herein is
subject to change without notice. All other trademarks mentioned herein are the property of their respective owners.
About the Author
Gopinath Mandala leads Cognizant’s QE&A Test Data Management Center of Excellence in India. He
has 17-plus years of IT experience spanning application design and development, test environment
management, test process and automation consulting, program management and enterprise CMMI
assessments. Gopi has mentored and transformed test data management processes for large enterpris-
es by leveraging multiple avenues such as efficient test data request provisioning systems, data reuse
strategy and test data provisioning tools. He currently leads Cognizant’s Test Data Management Center
of Excellence. Gopi can be reached at Gopinath.Mandala@cognizant.com.
Footnotes
1	
Test data management is the discipline of eliciting/understanding the data requirements for testing and
delivering relevant test data, meeting technical requirements ahead of the test execution timeline.
2	
Our survey was conducted with nearly 250 lines of business that the company serves. Survey partici-
pants included a mix of program managers and test managers.
3	
TD Maxim is a customizable test data solution framework engineered by us in partnership with Infor-
matica.

More Related Content

What's hot (20)

PPTX
Test Data Management: The Underestimated Pain
Chelsea Frischknecht
 
PPT
Ibm Optim Techical Overview 01282009
lucascibm
 
PPTX
DATPROF Test data Management (data privacy & data subsetting) - English
DATPROF
 
PDF
Leveraging Automated Data Validation to Reduce Software Development Timeline...
Cognizant
 
PPTX
Tufts Research: Strategies from Data Management Leaders to Speed Clinical Trials
Veeva Systems
 
PDF
Building a Robust Big Data QA Ecosystem to Mitigate Data Integrity Challenges
Cognizant
 
PPTX
Vertex Reduces EDC Study Build Times by 50%
Veeva Systems
 
PDF
Fraction ERP Overview
Patrick Chester
 
PPT
IBM InfoSphere Optim Solutions - Highlights
Adam Gartenberg
 
PDF
Deliver Trusted Data by Leveraging ETL Testing
Cognizant
 
PDF
Mdm for materials –positive impact of data quality improvement
Verdantis Inc.
 
PDF
Roadmap to next generation digital lab
Stephan Gürtler
 
PDF
Test Data Management: A Healthcare Industry Case Study
TechWell
 
PDF
How to Load Data More Quickly and Accurately into Oracle's Life Sciences Data...
Perficient, Inc.
 
PPTX
PICI’s Best Practices for Building Oncology Studies in an EDC
Veeva Systems
 
PDF
GSK: Preparing the Business for Study Start-up Change
Veeva Systems
 
PDF
Regulatory Considerations for use of Cloud Computing and SaaS Environments
Institute of Validation Technology
 
PDF
Clinical Trial Management System Implementation Guide
Perficient, Inc.
 
PDF
Estuate EDM Checklist
Estuate, Inc.
 
PPTX
Migrating Clinical Data in Various Formats to a Clinical Data Management System
Perficient, Inc.
 
Test Data Management: The Underestimated Pain
Chelsea Frischknecht
 
Ibm Optim Techical Overview 01282009
lucascibm
 
DATPROF Test data Management (data privacy & data subsetting) - English
DATPROF
 
Leveraging Automated Data Validation to Reduce Software Development Timeline...
Cognizant
 
Tufts Research: Strategies from Data Management Leaders to Speed Clinical Trials
Veeva Systems
 
Building a Robust Big Data QA Ecosystem to Mitigate Data Integrity Challenges
Cognizant
 
Vertex Reduces EDC Study Build Times by 50%
Veeva Systems
 
Fraction ERP Overview
Patrick Chester
 
IBM InfoSphere Optim Solutions - Highlights
Adam Gartenberg
 
Deliver Trusted Data by Leveraging ETL Testing
Cognizant
 
Mdm for materials –positive impact of data quality improvement
Verdantis Inc.
 
Roadmap to next generation digital lab
Stephan Gürtler
 
Test Data Management: A Healthcare Industry Case Study
TechWell
 
How to Load Data More Quickly and Accurately into Oracle's Life Sciences Data...
Perficient, Inc.
 
PICI’s Best Practices for Building Oncology Studies in an EDC
Veeva Systems
 
GSK: Preparing the Business for Study Start-up Change
Veeva Systems
 
Regulatory Considerations for use of Cloud Computing and SaaS Environments
Institute of Validation Technology
 
Clinical Trial Management System Implementation Guide
Perficient, Inc.
 
Estuate EDM Checklist
Estuate, Inc.
 
Migrating Clinical Data in Various Formats to a Clinical Data Management System
Perficient, Inc.
 

Similar to Multidimensional Challenges and the Impact of Test Data Management (20)

PDF
Ibm test data_management_v0.4
Rosario Cunha
 
PDF
A Complete Guide to Test Data Management (TDM).pdf
kalichargn70th171
 
PDF
What is Test Data Management? Why Should You Focus on It?
Enov8
 
PDF
4 Test Data Management Techniques That Empower Software Testing
Cigniti Technologies Ltd
 
PDF
Saksham Sarode - Building Effective test Data Management in Distributed Envir...
TEST Huddle
 
PDF
Mind Map Test Data Management Overview
dublinx
 
PPTX
'Test Data Management and Project Quality Go Hand In Hand' by Kristian Fische...
TEST Huddle
 
PPTX
Techniques for effective test data management in test automation.pptx
Knoldus Inc.
 
PDF
Test Data Management: Benefits, Challenges & Techniques
Enov8
 
PDF
Test Data Management Explained: Why It’s the Backbone of Quality Testing
Shubham Joshi
 
PDF
Test Data Management - Keytorc Approach
Keytorc Software Testing Services
 
PDF
Preparing for GDPR
GenRocket
 
PPTX
Test data automation: delivering quality data at speed
Curiosity Software Ireland
 
PDF
MetaSuite and_hp_quality_center_enterprise
Minerva SoftCare GmbH
 
PPTX
Curiosity Software and RCG Global Services Present - Solving Test Data: the g...
Curiosity Software Ireland
 
PDF
A Detailed Guide To Test Data Management.pdf
Enov8
 
PPTX
Data Con LA 2022 - Why Data Quality vigilance requires an End-to-End, Automat...
Data Con LA
 
PDF
The Vital Role of Test Data Management in Software Development.pdf
RohitBhandari66
 
PDF
Test Data Management 101—Featuring a Tour of CA Test Data Manager (Formerly G...
CA Technologies
 
Ibm test data_management_v0.4
Rosario Cunha
 
A Complete Guide to Test Data Management (TDM).pdf
kalichargn70th171
 
What is Test Data Management? Why Should You Focus on It?
Enov8
 
4 Test Data Management Techniques That Empower Software Testing
Cigniti Technologies Ltd
 
Saksham Sarode - Building Effective test Data Management in Distributed Envir...
TEST Huddle
 
Mind Map Test Data Management Overview
dublinx
 
'Test Data Management and Project Quality Go Hand In Hand' by Kristian Fische...
TEST Huddle
 
Techniques for effective test data management in test automation.pptx
Knoldus Inc.
 
Test Data Management: Benefits, Challenges & Techniques
Enov8
 
Test Data Management Explained: Why It’s the Backbone of Quality Testing
Shubham Joshi
 
Test Data Management - Keytorc Approach
Keytorc Software Testing Services
 
Preparing for GDPR
GenRocket
 
Test data automation: delivering quality data at speed
Curiosity Software Ireland
 
MetaSuite and_hp_quality_center_enterprise
Minerva SoftCare GmbH
 
Curiosity Software and RCG Global Services Present - Solving Test Data: the g...
Curiosity Software Ireland
 
A Detailed Guide To Test Data Management.pdf
Enov8
 
Data Con LA 2022 - Why Data Quality vigilance requires an End-to-End, Automat...
Data Con LA
 
The Vital Role of Test Data Management in Software Development.pdf
RohitBhandari66
 
Test Data Management 101—Featuring a Tour of CA Test Data Manager (Formerly G...
CA Technologies
 
Ad

More from Cognizant (20)

PDF
Using Adaptive Scrum to Tame Process Reverse Engineering in Data Analytics Pr...
Cognizant
 
PDF
Data Modernization: Breaking the AI Vicious Cycle for Superior Decision-making
Cognizant
 
PDF
It Takes an Ecosystem: How Technology Companies Deliver Exceptional Experiences
Cognizant
 
PDF
Intuition Engineered
Cognizant
 
PDF
The Work Ahead: Transportation and Logistics Delivering on the Digital-Physic...
Cognizant
 
PDF
Enhancing Desirability: Five Considerations for Winning Digital Initiatives
Cognizant
 
PDF
The Work Ahead in Manufacturing: Fulfilling the Agility Mandate
Cognizant
 
PDF
The Work Ahead in Higher Education: Repaving the Road for the Employees of To...
Cognizant
 
PDF
Engineering the Next-Gen Digital Claims Organisation for Australian General I...
Cognizant
 
PDF
Profitability in the Direct-to-Consumer Marketplace: A Playbook for Media and...
Cognizant
 
PDF
Green Rush: The Economic Imperative for Sustainability
Cognizant
 
PDF
Policy Administration Modernization: Four Paths for Insurers
Cognizant
 
PDF
The Work Ahead in Utilities: Powering a Sustainable Future with Digital
Cognizant
 
PDF
AI in Media & Entertainment: Starting the Journey to Value
Cognizant
 
PDF
Operations Workforce Management: A Data-Informed, Digital-First Approach
Cognizant
 
PDF
Five Priorities for Quality Engineering When Taking Banking to the Cloud
Cognizant
 
PDF
Getting Ahead With AI: How APAC Companies Replicate Success by Remaining Focused
Cognizant
 
PDF
Crafting the Utility of the Future
Cognizant
 
PDF
Utilities Can Ramp Up CX with a Customer Data Platform
Cognizant
 
PDF
The Work Ahead in Intelligent Automation: Coping with Complexity in a Post-Pa...
Cognizant
 
Using Adaptive Scrum to Tame Process Reverse Engineering in Data Analytics Pr...
Cognizant
 
Data Modernization: Breaking the AI Vicious Cycle for Superior Decision-making
Cognizant
 
It Takes an Ecosystem: How Technology Companies Deliver Exceptional Experiences
Cognizant
 
Intuition Engineered
Cognizant
 
The Work Ahead: Transportation and Logistics Delivering on the Digital-Physic...
Cognizant
 
Enhancing Desirability: Five Considerations for Winning Digital Initiatives
Cognizant
 
The Work Ahead in Manufacturing: Fulfilling the Agility Mandate
Cognizant
 
The Work Ahead in Higher Education: Repaving the Road for the Employees of To...
Cognizant
 
Engineering the Next-Gen Digital Claims Organisation for Australian General I...
Cognizant
 
Profitability in the Direct-to-Consumer Marketplace: A Playbook for Media and...
Cognizant
 
Green Rush: The Economic Imperative for Sustainability
Cognizant
 
Policy Administration Modernization: Four Paths for Insurers
Cognizant
 
The Work Ahead in Utilities: Powering a Sustainable Future with Digital
Cognizant
 
AI in Media & Entertainment: Starting the Journey to Value
Cognizant
 
Operations Workforce Management: A Data-Informed, Digital-First Approach
Cognizant
 
Five Priorities for Quality Engineering When Taking Banking to the Cloud
Cognizant
 
Getting Ahead With AI: How APAC Companies Replicate Success by Remaining Focused
Cognizant
 
Crafting the Utility of the Future
Cognizant
 
Utilities Can Ramp Up CX with a Customer Data Platform
Cognizant
 
The Work Ahead in Intelligent Automation: Coping with Complexity in a Post-Pa...
Cognizant
 
Ad

Multidimensional Challenges and the Impact of Test Data Management

  • 1. Multidimensional Challenges and the Impact of Test Data Management To deliver benefits such as reducing test windows, improving test quality, optimizing data processing/storage costs and maintaining complex data model/relationships, TDM must be embedded equally within an IT organization’s QA and development teams. Executive Summary Managing costs without compromising quality is the latest mantra of IT organizations. End-to-end test data management (TDM),1 as against mere provisioning of data for test execution, has far- reaching potential and can help rein in operation- al and processing costs. Over the past few years organizations have focused more on test data management, realizing its potential impact on multiple aspects of the quality assurance (QA) process, such as testing window reduction, improvement in quality of testing, optimizing data processing and storage costs and maintaining complex data models and relationships. To deliver these multidimensional impacts, TDM must be embedded within quality assurance and development teams, alike. Moreover, TDM must also intersect with other disciplines such as release management, environment management and data security (see Figure 1). TDM acts as a conduit between development and QA teams. It also needs to interface with the disciplines shown in the diagram to deliver the larger impacts discussed in the prior paragraph. These interfaces with the various disciplines increase the complexity that must be overcome. Why is TDM so pertinent? Our survey of the top 250 lines of business served by us makes clear why:2 • 80% of IT organizations acknowledge that they either have no tools to query/retrieve the right data or require improvements to their in-house tools. • 61% of IT organizations do not have any TDM tools. • 50% of IT organizations do not have a data reusability strategy in place. • 40% of IT organizations face data integrity issues in their test environments. • 38% of IT organizations have sensitive production data that can be accessed by in-house testing or QA teams. This white paper explores various dimensions of potential benefits organizations can derive from effective TDM and what is needed for QA teams to leverage these elements to achieve critical multi- dimensional benefits. • Cognizant 20-20 Insights cognizant 20-20 insights | july 2014
  • 2. 2 Unleash the Potential of Test Data Management By maximizing the potential of TDM, IT organi- zations can address the above challenges and deliver several valuable benefits: • Increased quality of test conditions and the leveraging of data integration tools. Based on our survey, nearly two-thirds of IT organizations do not have any TDM tools. Many organizations face a common challenge of ensuring that the test environment replicates the exact patterns of production. Some data integration and data profiling and extraction tools are built with features that can help identify data patterns in the production stage. Such tools can be used to capture and mimic similar patterns in the test environments. This can help elevate test data quality to the next level. The tools result in increased business confidence regarding test coverage and, eventually, in higher quality applications. • Optimized storage and processing costs. Typically, the size of a test environment increases as additional projects are brought online. This results in increasing storage costs. Bulky test environments also add indirectly to processing expenses such as archival and batch processing costs, because larger volumes of data are processed. • An efficient TDM framework needs to consider not only making the data available to testers but also must minimize the volumes of data in a test environment. A test data strategy needs to consider a long-term pipeline of activities and arrive at an optimal test bed that does not compromise data avail- ability. In addition, the test data strategy must be supplemented by an optimal archival, refresh or restore plan. • A shorter batch process window. This is a natural extension of the resulting impact of the previous challenge. Not only does it cost more to process unnecessary data, but also the batch process itself takes longer than needed. Today’s TDM approach tries to minimize the batch window. • Improved compliance to data protection policies and regulatory requirements. Our survey also revealed that over one-third of IT organizations let in-house testing teams access sensitive production data for their testing needs. IT organizations require test data architects to devise a solution to comply with their data protection policy, while simul- taneously retaining production-like quality test data. Persistent and dynamic data masking solutions are highly sought-after, especially in the context of the dynamic environment of reg- ulations. This is more pronounced in domains such as healthcare, banking and financial services and insurance where data privacy is a highly sensitive concern. Also, a small percentage of IT organiza- tions quoted data privacy as a roadblock for offshoring. Data masking solutions can enable cognizant 20-20 insights TDM’s Connections with Other Organizational Functions Figure 1 Tools Strategy Release Management Environment Management Program Management & Governance Test Data Management Knowledge Management Data Security Office Tools StTT rategyrr Release Management Environment Management Programrr Management & Governance Test DataTT Management Knowledge Management Data Security Office Test Data Management Program Management & Governance Data Security Office Release Management Knowledge Management Tool Strategy Environment Management
  • 3. 3cognizant 20-20 insights offshore test execution in compliance with the policy and relevant regulatory requirements. The key here is to ensure data integrity of the masked data, without exposing sensitive data. There are also instances in which a synthetic data generation solution has been leveraged to overcome data privacy concerns, paving the way for offshoring of TDM. • Better technical support to black-box testing teams for knowledge management. Nearly half of IT organizations surveyed face data integrity issues in their test environments. Test data teams can act as a bridge between the white-box develop- ment teams and black-box functional testing teams. While the focus of functional testing teams is to treat the application under test as a black box, the setup of the prerequisite test data requires a deep understand- ing of the underlying data structure and flow. Thanks to test data management tools, the knowledge about numerous business rules and the interactions among interfaces can be captured in a more elaborate and robust manner. This can lead to enhanced knowledge quality, better retention and easier maintenance. • Availability of more in-process efficien- cies: data fencing and formal data request systems. Nearly half of IT organizations polled spend at least 10% of their testing efforts on test data preparation. Improvements can also be made in terms of data fencing. That is, the ways in which testing teams earmark their data and avoid data loss and overrides. This process becomes complex when the testing team’s size is large (more than 200) and when it is involved in multiple and parallel projects that can impact the same module within an application. Efficiencies can also be extracted based on how testing teams create and handle their test data requests. More IT organizations now realize the impact of a speedy test data request and provi- sioning system. Moreover, they are also willing to invest in one. Some IT organizations are even deploying commercial request management systems with integrated work flows. • Improved test coverage. It is commonly assumed that production data, though present in large volumes, only covers about 30% of possible functionalities or scenarios. To maximize test coverage, testers need to condition the data to ensure testing of all possible scenarios. Here, a synthetic data generation solution can play a crucial role in creating the much-needed variety of data with desired distribution patterns. It can also be extended to improve coverage of perfor- mance testing by generating volumes that are sometimes many times bigger than the current production size. Nearly two-thirds of IT organizations surveyed have synthetic data generation needs. Also, as part of migration/upgrade scenarios, when large volumes of data are moved from source to target systems, it is only a small portion of target data that is typically tested manually with the help of scripts and queries. The low coverage of testing leads to the risk of untested data at the target. Organizations can leverage validation-related products, either those in-house or the ones available commer- cially, to ensure 100% coverage. • Optimized data provisioning time: TDM accelerators. More than two-thirds of the IT organizations surveyed acknowledge that they have no tools to either query or retrieve the right data. To capture the business rules and understand the information data structures more efficiently, accelerators can be set up that readily contain this information for packages such as SAP or Siebel. Accelerators are also available for regulatory compliances such as Payment Card Industry’s Data Security Standards (PCI DSS) and Protected Health Information (PHI). These accelerators not only cut down the configuration time and effort, but also improve the consistency of how these rules are applied. • Maximized utilization of knowledge and skills. A small percentage of IT organiza- tions surveyed have centralized their test data team services. As such, the benefits of a managed services model can be applied to TDM. TDM requires skills in niche areas such as extract-transform-load (ETL) and online analytical processing (OLAP)/online transac- tion processing (OLTP) which may not be nec- essarily available with functional testing teams. Having such high-skilled resources functioning as TDM specialists can help achieve greater efficiency. Thanks to test data management tools, the knowledge about numerous business rules and the interactions among interfaces can be captured in a more elaborate and robust manner. This can lead to enhanced knowledge quality, better retention and easier maintenance.
  • 4. cognizant 20-20 insights 4 Managed Services TDM in Action Figure 2 Business Outcomes: Case in Point Figure 2 highlights engagements where we helped IT organizations embrace, deploy and benefit from TDM delivered via a managed services approach. Looking Ahead A comprehensive and end-to-end TDM strategy that addresses the multidimensional challenges IT organizations face can deliver multifaceted benefits to enterprises across industries. It can help reduce operational and processing costs, while simultaneously compressing the time needed to achieve the cost savings. It can also eliminate the risk of exposing sensitive data to nonproduction users while also improving test coverage. TDM, in our experience, furthers better knowledge retention and management with QA organiza- tions, and higher data reuse. Applying a managed services model to TDM provides key benefits to IT organizations, including wider access to niche skills and optimal resource utilization. TDM has a far wider potential than mere provisioning of data for test execution. The advent of big vendors such as Informatica and IBM from the data integration space has only advanced TDM’s ability to deliver large-scale benefits, while shrinking processing windows. Challenges Improvements and Changes Results A Multibillion Banking & Financial Services Organization in the UK Manually intensive test data provisioning. Lack of a formal data request tracking mechanism. Lack of consolidated knowledge on data conditions. Centralized team with internal ownership of knowledge management. API-based automation (data extraction and data generation). Centralized data request tracking: SLAs based on the turnaround time and data suitability. Annual savings of $2.5 million in test data provisioning. 40% increase in productivity. A U.S.-Based Retail Conglomerate Complex application design. Sensitive production data used in test environment. Full load of data each time, making it longer than required. Deployed the TD MaximTM framework3 and configured complex rules removing manual efforts. TD Maxim used for applying filters to get just enough data. 75% reduction in TDM efforts. Gained data privacy in test environments. Reduction of data volume by 80%. A Leading U.S.-Based Healthcare Organization Sensitive protected health information (PHI) was unprotected in the test environment. Inconsistent and effort- intensive methods for generating test data. Deployed the TD Maxim framework with necessary masking algorithms and policies. The framework was also utilized for synthetic data generation. Achieved legal compliance as per PHI regulations. Consistent and streamlined data generation process. A Global Biopharmaceuticals Giant A large-scale program facing huge schedule slippage due to unavailability of test data. Simple stubbing ruled out because of the need for high-quality test data. Deployed the TD Maxim framework for synthetic data generation, which ensures data integrity rules across all interfaces. Averted risk of derailment of the project.
  • 5. About Cognizant Cognizant (NASDAQ: CTSH) is a leading provider of information technology, consulting, and business process out- sourcing services, dedicated to helping the world’s leading companies build stronger businesses. Headquartered in Teaneck, New Jersey (U.S.), Cognizant combines a passion for client satisfaction, technology innovation, deep industry and business process expertise, and a global, collaborative workforce that embodies the future of work. With over 75 development and delivery centers worldwide and approximately 178,600 employees as of March 31, 2014, Cognizant is a member of the NASDAQ-100, the S&P 500, the Forbes Global 2000, and the Fortune 500 and is ranked among the top performing and fastest growing companies in the world. Visit us online at www.cognizant.com or follow us on Twitter: Cognizant. World Headquarters 500 Frank W. Burr Blvd. Teaneck, NJ 07666 USA Phone: +1 201 801 0233 Fax: +1 201 801 0243 Toll Free: +1 888 937 3277 Email: [email protected] European Headquarters 1 Kingdom Street Paddington Central London W2 6BD Phone: +44 (0) 20 7297 7600 Fax: +44 (0) 20 7121 0102 Email: [email protected] India Operations Headquarters #5/535, Old Mahabalipuram Road Okkiyam Pettai, Thoraipakkam Chennai, 600 096 India Phone: +91 (0) 44 4209 6000 Fax: +91 (0) 44 4209 6060 Email: [email protected] ­­© Copyright 2014, Cognizant. All rights reserved. No part of this document may be reproduced, stored in a retrieval system, transmitted in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the express written permission from Cognizant. The information contained herein is subject to change without notice. All other trademarks mentioned herein are the property of their respective owners. About the Author Gopinath Mandala leads Cognizant’s QE&A Test Data Management Center of Excellence in India. He has 17-plus years of IT experience spanning application design and development, test environment management, test process and automation consulting, program management and enterprise CMMI assessments. Gopi has mentored and transformed test data management processes for large enterpris- es by leveraging multiple avenues such as efficient test data request provisioning systems, data reuse strategy and test data provisioning tools. He currently leads Cognizant’s Test Data Management Center of Excellence. Gopi can be reached at [email protected]. Footnotes 1 Test data management is the discipline of eliciting/understanding the data requirements for testing and delivering relevant test data, meeting technical requirements ahead of the test execution timeline. 2 Our survey was conducted with nearly 250 lines of business that the company serves. Survey partici- pants included a mix of program managers and test managers. 3 TD Maxim is a customizable test data solution framework engineered by us in partnership with Infor- matica.