SlideShare une entreprise Scribd logo
1  sur  8
Télécharger pour lire hors ligne
• Cognizant 20-20 Insights

Information Lifecycle and the
Actionability Test
By scoring information generated internally and externally, and
assessing its importance and maturity, organizations can better
understand how and where to store key data for quick access and
applicability to strategic planning.
Executive Summary
For the past several years, enterprises have
been drowning in a deluge of data, and the sheer
volume is growing exponentially. That translates
to downstream reporting environments, where,
for instance, data warehouses have become
behemoth data dumps containing multiple
terabytes and even petabytes of information.
While it is fortunate that new technologies have
made it possible to store this volume of information at a significantly reduced cost, professionals
who are dealing with this data deluge still suffer
from information overload.
Information overload is caused by too much data
created by too many sources. In fact, the sheer
number of sources is increasing geometrically, as
is the volume and complexity of information (i.e.,
formats, metadata, frequency, etc.) available for
analysis. (See Figure 1, next page, for illustrative
examples.)
For years, businesses have focused on moving
nonoperational data into archives through an
orderly automated process or archiving the
contents of retired systems. But rarely have they
used information lifecycle techniques to ensure

cognizant 20-20 insights | september 2013

that what is easily accessible in their operational
systems is actionable. This white paper defines a
process to ensure exactly that.
According to IDC,1 the volume of information
available will grow 50-fold over the next several
years (see Figure 2, next page). But data volume
is only part of the problem. The sheer number
of tables, overlapping sources, references to
differing formats of information and other issues
are causing acute problems for organizations
that utilize traditional techniques for analysis and
reporting. The unbridled growth of warehouses
as a means of servicing all constituencies has
reached its limit of usefulness, and a better
approach to managing the growth of data sources
is required, such as allowing warehouse environments to grow organically to service the needs
of all. This unmanaged organic growth is causing
difficulty to those striving to glean insight from
or discover anything useful in these behemoth
warehouses.
While big data makes it possible to store and
retrieve insight from large data pools at an accelerated rate, it does not help people focus on
actionable information nuggets. The economist
The Reality of Information Overload

42%

Organizations that say the
paper records they are
generating continue to
increase.

11

Seconds to generate one
petabyte of information, or
the equivalent of 13 years
of high definition video.

15.4

Source: Intel Corporation

Source: AIIM survey
of 512 AIIM members
conducted in January
and February, 2013.

Millions of dollars in
health insurance
premiums overpaid for
fiscal year 2012/13,
half of which were
to cover workers who
were no longer
employed.
Source: Wisconsin
Legislative Audit Bureau

Figure 1

broad buckets. They do not generally cull out
organizational trivia from warehouses, under the
premise that this information may be needed at
some future date.

Herbert A. Simon had it right when he wrote, “In
an information-rich world, the wealth of information means a dearth of something else: a scarcity
of whatever it is that information consumes.
What information consumes is rather obvious: it
consumes the attention of its recipients. Hence
a wealth of information creates a poverty of
attention and a need to allocate that attention
efficiently among the overabundance of information sources that might consume it.”2 It is exactly
such focused attention that organizations need to
properly derive insights from the explosion of data.
And, most importantly, this is where an appropriate information lifecycle program comes in handy.

According to Omar Tawakol, CEO of BlueKai,
the data warehouse’s historical premise was to
provide a pool of storage where trustworthy
data could be parked for analysis.3 However, the
explosive growth of data has made this pooling
of data superfluous; the enormity of the end data
state made the warehouse virtually unusable.
An example of an actionable data warehouse
is the one used by the U.S. Internal Revenue
Service, which manages a historical base of 230
million annual tax returns and 50 billion tax-

Most information lifecycle programs tackle basic
needs, such as segregating data into several

The Growing Data Universe
Exabytes
40,000

30,000

20,000

10,000

2013 YE Digital Universe Estimate
e
m
Year

2013

2014

2015

2016

Source: IDC Digital Universe Study, 2012
Figure 2

cognizant 20-20 insights

2

5,000
related telephone requests in a single warehouse.
Queries made against this warehouse help to:

•	 Estimate the U.S. tax gap.
•	 Predict identity theft, fraud

•	 Information

that is actionable and routinely
used for decision-making, such as sales information, customer activities, working capital
availability, disputed invoices stuck in receivables, consumer sentiment, market capitalization, fixed and variable costs of operations,
competitive insight and innovation programs
that help differentiate an organization from its
competitors.

and other tax

issues.

•	 Model financial risks.
•	 Measure taxpayer burdens.
•	 Simulate new tax laws and their effects on tax
behavior.

•	 Information

that is actionable for a specific
business need that may be routinely required,
such as changes in long-term debt portfolios,
changes in credit ratings, changes in market
valuation, caustic social media coverage and
changes in the regulatory climate.

•	 Optimize the IRS staffing model.

4

Information Lifecycle Maturity
Information lifecycle management (see Figure 3)
can be viewed in terms of levels of maturity:

•	 Information

that is actionable for a short
timeframe and may be historical trivia shortly
afterwards, such as tracking the construction
of a new plant, monitoring the launch of a new
product, addressing raw materials reject rates
from a supplier, monitoring spikes in healthcare
costs and monitoring online groups for entries
made by a disgruntled employee.

•	 Immature,

where the information lifecycle
management (ILM) process begins.

•	 Aware, where in most cases basic ILM maturity
emerges by partitioning data onto static (not
logged for updates) partitions.

•	 Basic managed, where a basic ILM program is
established.

•	 Advanced

•	 Information that is operationally relevant but

not necessarily required as a primary source
for decision-making, such as the dispensation
of contact center calls, the safety reports of
elevators and the detailed time and attendance
logs of employees.

managed, where actionable information is managed.

•	 Governed, where the ILM program is governed.
Understanding Data Consumption
Organizations should assign a status to the
candidate information to be managed by an
information lifecycle program. These information
priority statuses are as follows:

•	 Information seen as historical trivia, such as

long-lost organizational contacts, building
maintenance reports for divested facilities, old
detailed call detail records, ancient detailed
journal entry logs that no longer align with the

Information Lifecycle Maturity
Information Maturity
Lifecycle Stage
Immature

Aware

Basic Managed

Advanced Managed
Governed

Description of Stage

Typical Stage Activities

Database tables grow linearly
with only a fraction of their
contents usable.

The information lifecycle management journey
normally begins here.

A need for lifecycle is present.

The utilization of database partitioning is
normally a first activity in satisfying ILM needs.

A repeatable Information
lifecycle management program is
present.

Tools are introduced to orchestrate the
movement of information through the ILM
processes.

The information lifecycle management program is directed to deal with the deluge
of information, much of which is never utilized.
A proactive governance body orchestrates the information lifecycle processes.

Figure 3

cognizant 20-20 insights

3
chart of accounts and competitive insight on
defunct companies.

Understanding Information Gleaned
Candidate information for an organization’s information lifecycle program can be obtained from a
variety of sources:

quo — which would require a swift determination of whether action is warranted and, if it is,
execution of the appropriate coordinated action
to protect, extract, originate or capture value for
the organization (see Figure 4).
Data can originate from:

•	 Indirect sources, such as Web sites maintained

•	 A log, or formatted activity logs made available

•	 Direct sources, such as applications, logs and

•	 A

from a call center, Web site, RFID, mobile device
or other facility that records its activity into a
log.

by partnering organizations, distributors and
retailer reports, advertisers, regulators and
watchdogs.

other information from Web sites, logs and
other information from mobile facilities and
logs and other information from call centers
that your organization originates.

•	 Externally sourced information, such as
data collected from RFID tags, mobile devices,
social media, modifications made on tracked
Web sites, traditional and nontraditional news
media, public relations bulletins, acquired intelligence, affiliate referrals, etc.

message, received from an internal or
external source, such as a SWIFT message,
an EDI transaction or other message based
sources.

•	 A stream, or an asynchronous source of infor-

mation that streams information to a specific
listening target, such as Twitter, a market ticker,
newswires and other sources.

•	 A

document, or any form of object that is
available as an information source, such as an
XML file, a public relations briefing, a weather
map, a schematic or other non-tabular information source.

Each of these sources may contain superfluous
information, or may signal a change to the status

Information Source Formats
Possible Source Formats

Information

Retailer Reports
Advertisers
Watchdogs and Regulators
Internal Sources

Direct

Web Sites
Mobile Facilities
Call Centers
Mobile Facilitated
RFID

Externally
Obtained

Social Media
News and PR Media
Acquired Intelligence
Affiliate Referrals
Other

Figure 4

cognizant 20-20 insights

4

Tabular
External

Indirect

Tabular
Internal

Distributor Reports

Document

Partner Web

Stream

Message

Log

Delivery Models
•	 A

tabular source of information, normally
housed in databases and originated from
sources internal to an organization (internal
ERP systems, spreadsheets, CRM systems,
etc.) or external sources (such as vendors,
customers and industry associations).

The information lifecycle management program
must be able to decipher which of these data
sources are usable for long-term analysis and
which ones are relevant only for situational use
and temporal event detection. For example, the
session statistics from a Web log may be usable
for long-term use; however, the Web log detail
may be usable only until it is transformed. At this
point, this data may be very well a candidate for
archiving.

The business information model should serve as a
primary tool for determining placement of information classes in the information management
lifecycle (see Figure 5).
Scoring information consumed by an information
lifecycle program prioritizes information deemed
mandatory for operational use, its importance
in managing operational risks, its importance in
enhancing sustainable value (capturing, extending
or creating sustainable value), defining organizational strategies or diffusing disruptive events.

Data Attributes and the
Information Lifecycle
The information lifecycle program should consider
the following characteristics of data:

A Model for Deciphering
Information Actionability

•	 The ability to recreate information should it be

Information captured for analysis can be used for
managing operational risks, diffusing disruptive
events, enhancing sustainable value and defining
enterprise strategies. The information lifecycle
must determine which of the data captured
about customers, media, competitors, markets,
financing and geographies will be useful in identifying patterns or deriving insight for the organization.

•	 The veracity of sources transformed into what

incorrectly or inadvertently destroyed.
is being archived.

•	 Actionability of the candidate information.
•	 The information’s priority status.
The scoring will facilitate whether the information is:

•	 Actively maintained (and thus should reside in
active online partitions).

Managing
Operational
Risks

n
tio
ra
Service

Customers

Be

Value Chain

ne

fit

Financing

s

Enhancing
Sustainable
Value
Geographies
Ge

Product

le

Focused
Information

op

ts
Markets

Process

Capabilities

Barriers
Diffusing
Disruptive
Events

M
e
cs
tri

Col
lab
o

Business Information Model

Defining
Enterprise
Strategies

Figure 5

cognizant 20-20 insights

5

Pe
•	 Required

Assigning Scores for Lifecycle
Placement Data

•	 Required, but has a low information lifecycle

The primary reason for focusing on the actionability of information — based on its format and
ease of recreation, its age and consumption
pattern — is to derive a mechanism to assign a
score that can be used to place information into
the buckets of storage managed by an information lifecycle program.

regularly but historical in nature
(static online partitions).
pool scoring (near-line storage).

•	 Required infrequently but must be retained for
regulatory or other needs (archived).

•	 Is no longer required for regulatory needs and
is not accessed (to be destroyed).

A viable approach for utilizing data attributes in
an information lifecycle program is automating
the lifecycle scoring algorithm (see Figure 5)
within the information lifecycle orchestration
application.

It is important to note that there will be differing
retention requirements for information utilized
by an organization, which may serve as information lifecycle overrides. For example, vested

Scoring Data Attributes
Managed Information Lifecycle Pools Index
Factors in Scoring the
Placement of Information

Active
Online
(100-86)

Static
Online
(85-76)

Operationally Critical

Near line
(75-61)

Archived
(60-19)

Destructed
(18-11)

Operational Criticality
scored 10-15

(1=not critical, 15=highly critical

Information Priority Status

Information Priority Status
scored 10-15

(1=organizational trivia,
15=highly actionable
Analytic Criticality
(1=not required, 10=critical)

Analytic Criticality scored
7-10

Able to Recreate
(1=easily, 5=not able)
Veracity of Sources
(1=Low, 10=High)
Storage Platform
(1=other,2=Mainframe,
3=portal, 4=database, 5=big
data)
Regulatory Availability
(1=not required, 15=critical)

Regulatory
Availability
scored 3-15

Age of Data (1=>3 years,
3=1 -3 years, 5=<1 year)
Information Format
(5=Tabular, 4=Document,
3=Message, 2=Log, 1=Stream)
Used in Active Program
(1=Not used, 5=actively used)

Used in active
program
scored 2-5

Frequency of Access
(1=rarely if ever, 10 = regularly)

Frequency of
Access scored
3-10

Figure 5

cognizant 20-20 insights

6
pension recipient information must be saved for
employees who have left an organization because
they will be receiving pensions at a future date.
Thus, the information about them has operational
(funding needs for pension plans) and regulatory
(pension regulations) retention requirements that
will override some information lifecycle placement
scores — i.e., you cannot destroy data on potential
pensioners who are less than 60 years old, vested
and have not yet received pensions.

ILM Implementation and Testing
Approaches

Using a matrix similar to the one in Figure 5, a
numeric scoring will be derived, which should be
used to determine the appropriate placement of
data within the information lifecycle. To control
the scoring of information for use in analytics
and regulatory retention requirements, a simple
scoring algorithm can be deployed to enter data
into the organization’s information management
lifecycle management software.

ILM Governance

Just as important as moving data through the
lifecycle into its archival target is the destruction
of data no longer required nor usable by the organization. There are countless examples of organizations that have had old records subpoenaed
and used for tax and HR disputes and patent
infringement disagreements, among other
reasons. When information is no longer required,
it should be disposed of.

The processes that govern an information lifecycle program should be automated in such a
way that are repeatable and auditable (see Figure
6). These processes should ensure that information, particularly that which is moved to near-line
storage or archived, can be retrieved and used
and that information flagged for destruction
cannot be utilized after deletion.

The role of governance is simple in the information lifecycle program. Governance facilitates
the scoring of the placement of information into
the appropriate information lifecycle pools and
informs the proper execution of the information
lifecycle program.
Tasks specific to the administration of the information lifecycle program are as follows:

•	 Scoring

the managed information lifecycle
pools matrix.

•	 Measuring and communicating the program’s
cost and the associated cost avoidance.

•	 Measuring the risk associated with the scoring
of specific regulated and high-risk information and rescoring should an elevated risk be
derived.

ILM Construction
Implementation Approach
ILM Governance

Information
lifecycle
configuration.

Business review of
information necessity
and relevance.

Develop queries
to validate the ILM
pools of data (near-line,
archived, etc.)

Validate that ILM
managed data
is usable.

Establish test cases
to validate changes.
Spot profiling review to
validate proper information
placement.

Review action ability
scoring of ILM managed
information.

Identify changes to
the ILM configuration.
Review rules governing
the placement of
information.

Execute the archival recovery
in test mode and destroyed
data recovery attempt.
Ensure that archived
data is recoverable and
destroyed data is not.
Acceptance
test

Figure 6

cognizant 20-20 insights

7

Implementation

1. Code movement to production.
2. Trigger application retirement,
if applicable.
•	 Measuring

•	 Review

Organizations that have effective information
lifecycle governance programs and high degrees
of information lifecycle maturity routinely revisit
their information lifecycle program and adjust
the information lifecycle pools matrix as part of
the annual planning cycle. This enables organizations to:

•	 Regularly

and communicating information
lifecycle maturity model placement of the
organization.

•	 Ensure

that a partitioning program administered through an information lifecycle product
(Solix, Optim, Informatica ILM, Autonomy,
Opentext, Storetrends, etc.) exists and is
regularly monitored for effectiveness.

key technologies in the information
lifecycle program and monitor storage, floor
space, manpower and power consumption
statistics; and regularly update to newer technologies where they make economic sense and
are not operational disruptions.
monitor the effectiveness of thirdparty information lifecycle products and align
to make the capabilities of the information
lifecycle program effective and make optimal
use of the information lifecycle capabilities of
vendors participating in the program.

Footnotes
1	

IDC Digital Universe Study, December 2012.

2	

H.A. Simon, “Designing Organizations for an Information-Rich World,” in Martin Greenberger, ed.,
Computers, Communication, and the Public Interest, pp. 40–41, Johns Hopkins Press, Baltimore, 1971.

3	

Omar Tawakol, “The Death of the Data Warehouse and the Age of Activation,” Huffington Post, July 12,
2012.

4	

Internal Revenue Service Compliance Data Warehouse case study, General Accounting Office, 2011.

About the Author
Mark Albala is the Service Line Lead of Cognizant’s Data On-Demand Service Line, an organization that
provides services and products that support the effectiveness and efficiency of databases and other
information platforms used for managing information as part of Cognizant’s overall capabilities in
enterprise information management. A graduate of Syracuse University, Mark has held senior consulting,
thought leadership, advanced technical and business development roles for organizations focused
on the disciplines of business intelligence, governance and data warehousing. He can be reached at
Mark.Albala@cognizant.com.

About Cognizant
Cognizant (NASDAQ: CTSH) is a leading provider of information technology, consulting, and business process outsourcing services, dedicated to helping the world’s leading companies build stronger businesses. Headquartered in
Teaneck, New Jersey (U.S.), Cognizant combines a passion for client satisfaction, technology innovation, deep industry
and business process expertise, and a global, collaborative workforce that embodies the future of work. With over 50
delivery centers worldwide and approximately 164,300 employees as of June 30, 2013, Cognizant is a member of the
NASDAQ-100, the S&P 500, the Forbes Global 2000, and the Fortune 500 and is ranked among the top performing
and fastest growing companies in the world. Visit us online at www.cognizant.com or follow us on Twitter: Cognizant.

World Headquarters

European Headquarters

India Operations Headquarters

500 Frank W. Burr Blvd.
Teaneck, NJ 07666 USA
Phone: +1 201 801 0233
Fax: +1 201 801 0243
Toll Free: +1 888 937 3277
Email: inquiry@cognizant.com

1 Kingdom Street
Paddington Central
London W2 6BD
Phone: +44 (0) 20 7297 7600
Fax: +44 (0) 20 7121 0102
Email: infouk@cognizant.com

#5/535, Old Mahabalipuram Road
Okkiyam Pettai, Thoraipakkam
Chennai, 600 096 India
Phone: +91 (0) 44 4209 6000
Fax: +91 (0) 44 4209 6060
Email: inquiryindia@cognizant.com

©
­­ Copyright 2013, Cognizant. All rights reserved. No part of this document may be reproduced, stored in a retrieval system, transmitted in any form or by any
means, electronic, mechanical, photocopying, recording, or otherwise, without the express written permission from Cognizant. The information contained herein is
subject to change without notice. All other trademarks mentioned herein are the property of their respective owners.

Contenu connexe

Tendances

LS_WhitePaper_NextGenAnalyticsMay2016
LS_WhitePaper_NextGenAnalyticsMay2016LS_WhitePaper_NextGenAnalyticsMay2016
LS_WhitePaper_NextGenAnalyticsMay2016
Anjan Roy, PMP
 
Big data - a review (2013 4)
Big data - a review (2013 4)Big data - a review (2013 4)
Big data - a review (2013 4)
Sonu Gupta
 
IBM-Infoworld Big Data deep dive
IBM-Infoworld Big Data deep diveIBM-Infoworld Big Data deep dive
IBM-Infoworld Big Data deep dive
Kun Le
 
Big Data Analytics: Recent Achievements and New Challenges
Big Data Analytics: Recent Achievements and New ChallengesBig Data Analytics: Recent Achievements and New Challenges
Big Data Analytics: Recent Achievements and New Challenges
Editor IJCATR
 

Tendances (20)

LS_WhitePaper_NextGenAnalyticsMay2016
LS_WhitePaper_NextGenAnalyticsMay2016LS_WhitePaper_NextGenAnalyticsMay2016
LS_WhitePaper_NextGenAnalyticsMay2016
 
Modern trends in information systems
Modern trends in information systemsModern trends in information systems
Modern trends in information systems
 
The Post-Relational Reality Sets In: 2011 Survey on Unstructured Data
The Post-Relational Reality Sets In: 2011 Survey on Unstructured DataThe Post-Relational Reality Sets In: 2011 Survey on Unstructured Data
The Post-Relational Reality Sets In: 2011 Survey on Unstructured Data
 
IRJET - Big Data Analysis its Challenges
IRJET - Big Data Analysis its ChallengesIRJET - Big Data Analysis its Challenges
IRJET - Big Data Analysis its Challenges
 
Big data - a review (2013 4)
Big data - a review (2013 4)Big data - a review (2013 4)
Big data - a review (2013 4)
 
Ch03
Ch03Ch03
Ch03
 
IRJET- Analysis of Big Data Technology and its Challenges
IRJET- Analysis of Big Data Technology and its ChallengesIRJET- Analysis of Big Data Technology and its Challenges
IRJET- Analysis of Big Data Technology and its Challenges
 
IBM-Infoworld Big Data deep dive
IBM-Infoworld Big Data deep diveIBM-Infoworld Big Data deep dive
IBM-Infoworld Big Data deep dive
 
Introduction to Big Data
Introduction to Big DataIntroduction to Big Data
Introduction to Big Data
 
Stream Meets Batch for Smarter Analytics- Impetus White Paper
Stream Meets Batch for Smarter Analytics- Impetus White PaperStream Meets Batch for Smarter Analytics- Impetus White Paper
Stream Meets Batch for Smarter Analytics- Impetus White Paper
 
Big Data is Here for Financial Services White Paper
Big Data is Here for Financial Services White PaperBig Data is Here for Financial Services White Paper
Big Data is Here for Financial Services White Paper
 
Idc info brief-choosing_dbms_to_address_challenges_of_the_third-platform
Idc info brief-choosing_dbms_to_address_challenges_of_the_third-platformIdc info brief-choosing_dbms_to_address_challenges_of_the_third-platform
Idc info brief-choosing_dbms_to_address_challenges_of_the_third-platform
 
An introduction to data mining
An introduction to data miningAn introduction to data mining
An introduction to data mining
 
Big Data: Opportunities, Strategy and Challenges
Big Data: Opportunities, Strategy and ChallengesBig Data: Opportunities, Strategy and Challenges
Big Data: Opportunities, Strategy and Challenges
 
Big data analytics and its impact on internet users
Big data analytics and its impact on internet usersBig data analytics and its impact on internet users
Big data analytics and its impact on internet users
 
The Comparison of Big Data Strategies in Corporate Environment
The Comparison of Big Data Strategies in Corporate EnvironmentThe Comparison of Big Data Strategies in Corporate Environment
The Comparison of Big Data Strategies in Corporate Environment
 
Big data
Big dataBig data
Big data
 
Unstructured Data into EHR Systems: Challenges and Solutions
Unstructured Data into EHR Systems: Challenges and SolutionsUnstructured Data into EHR Systems: Challenges and Solutions
Unstructured Data into EHR Systems: Challenges and Solutions
 
Big Data Analytics: Recent Achievements and New Challenges
Big Data Analytics: Recent Achievements and New ChallengesBig Data Analytics: Recent Achievements and New Challenges
Big Data Analytics: Recent Achievements and New Challenges
 
Global Healthcare Information System Market will hit USD 54.0 Billion By 2020
Global Healthcare Information System Market will hit USD 54.0 Billion By 2020Global Healthcare Information System Market will hit USD 54.0 Billion By 2020
Global Healthcare Information System Market will hit USD 54.0 Billion By 2020
 

En vedette

En vedette (6)

Apps for the Connected World: Supercharge Customer Data with Code Halos
Apps for the Connected World: Supercharge Customer Data with Code HalosApps for the Connected World: Supercharge Customer Data with Code Halos
Apps for the Connected World: Supercharge Customer Data with Code Halos
 
Turning Customer Knowledge into Business Growth
Turning Customer Knowledge into Business GrowthTurning Customer Knowledge into Business Growth
Turning Customer Knowledge into Business Growth
 
Realizing the Potential of Social Media: A Strategy and Approach for Insurers
Realizing the Potential of Social Media: A Strategy and Approach for InsurersRealizing the Potential of Social Media: A Strategy and Approach for Insurers
Realizing the Potential of Social Media: A Strategy and Approach for Insurers
 
A Statistical Framework for Cluster Health Assessment and Its Application in ...
A Statistical Framework for Cluster Health Assessment and Its Application in ...A Statistical Framework for Cluster Health Assessment and Its Application in ...
A Statistical Framework for Cluster Health Assessment and Its Application in ...
 
A Blueprint for DAM, MAM Cost Avoidance
A Blueprint for DAM, MAM Cost AvoidanceA Blueprint for DAM, MAM Cost Avoidance
A Blueprint for DAM, MAM Cost Avoidance
 
Drones: The Insurance Industry's Next Game-Changer?
Drones: The Insurance Industry's Next Game-Changer?Drones: The Insurance Industry's Next Game-Changer?
Drones: The Insurance Industry's Next Game-Changer?
 

Similaire à Information Lifecycle and the Actionability Test

Big_data_analytics_for_life_insurers_published
Big_data_analytics_for_life_insurers_publishedBig_data_analytics_for_life_insurers_published
Big_data_analytics_for_life_insurers_published
Shradha Verma
 
Project 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docx
Project 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docxProject 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docx
Project 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docx
stilliegeorgiana
 
Power from big data - Are Europe's utilities ready for the age of data?
Power from big data - Are Europe's utilities ready for the age of data?Power from big data - Are Europe's utilities ready for the age of data?
Power from big data - Are Europe's utilities ready for the age of data?
Steve Bray
 

Similaire à Information Lifecycle and the Actionability Test (20)

Fast Data and Architecting the Digital Enterprise Fast Data drivers, componen...
Fast Data and Architecting the Digital Enterprise Fast Data drivers, componen...Fast Data and Architecting the Digital Enterprise Fast Data drivers, componen...
Fast Data and Architecting the Digital Enterprise Fast Data drivers, componen...
 
Running at the Speed of Digital: Hyper-Digital Information Management
Running at the Speed of Digital: Hyper-Digital Information ManagementRunning at the Speed of Digital: Hyper-Digital Information Management
Running at the Speed of Digital: Hyper-Digital Information Management
 
The Future Of Big Data In Business – 4 Emerging Trends In 2022.pptx
The Future Of Big Data In Business – 4 Emerging Trends In 2022.pptxThe Future Of Big Data In Business – 4 Emerging Trends In 2022.pptx
The Future Of Big Data In Business – 4 Emerging Trends In 2022.pptx
 
Big data analytics for life insurers
Big data analytics for life insurersBig data analytics for life insurers
Big data analytics for life insurers
 
Big_data_analytics_for_life_insurers_published
Big_data_analytics_for_life_insurers_publishedBig_data_analytics_for_life_insurers_published
Big_data_analytics_for_life_insurers_published
 
Project 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docx
Project 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docxProject 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docx
Project 3 – Hollywood and IT· Find 10 incidents of Hollywood p.docx
 
The Economic Value of Data: A New Revenue Stream for Global Custodians
The Economic Value of Data: A New Revenue Stream for Global CustodiansThe Economic Value of Data: A New Revenue Stream for Global Custodians
The Economic Value of Data: A New Revenue Stream for Global Custodians
 
Power from big data - Are Europe's utilities ready for the age of data?
Power from big data - Are Europe's utilities ready for the age of data?Power from big data - Are Europe's utilities ready for the age of data?
Power from big data - Are Europe's utilities ready for the age of data?
 
149.pdf
149.pdf149.pdf
149.pdf
 
Aligning Information Insights with the Speed of Business
Aligning Information Insights with the Speed of BusinessAligning Information Insights with the Speed of Business
Aligning Information Insights with the Speed of Business
 
Mejorar la toma de decisiones con Big Data
Mejorar la toma de decisiones con Big DataMejorar la toma de decisiones con Big Data
Mejorar la toma de decisiones con Big Data
 
Data warehouse,data mining & Big Data
Data warehouse,data mining & Big DataData warehouse,data mining & Big Data
Data warehouse,data mining & Big Data
 
Lecture 5 the information system a general model of ais:update version
Lecture 5  the information system   a general model of ais:update versionLecture 5  the information system   a general model of ais:update version
Lecture 5 the information system a general model of ais:update version
 
Lecture 5 the information system - a general model of ais
Lecture 5  the information system - a general model of aisLecture 5  the information system - a general model of ais
Lecture 5 the information system - a general model of ais
 
Big Data, Analytics and Data Science
Big Data, Analytics and Data ScienceBig Data, Analytics and Data Science
Big Data, Analytics and Data Science
 
Big Data Analytics
Big Data AnalyticsBig Data Analytics
Big Data Analytics
 
Big data analytics
Big data analyticsBig data analytics
Big data analytics
 
IOT & Procuement
IOT & ProcuementIOT & Procuement
IOT & Procuement
 
Business intelligence and big data
Business intelligence and big dataBusiness intelligence and big data
Business intelligence and big data
 
Data deck - CV - AXA - CVC
Data deck - CV - AXA - CVCData deck - CV - AXA - CVC
Data deck - CV - AXA - CVC
 

Plus de Cognizant

Plus de Cognizant (20)

Using Adaptive Scrum to Tame Process Reverse Engineering in Data Analytics Pr...
Using Adaptive Scrum to Tame Process Reverse Engineering in Data Analytics Pr...Using Adaptive Scrum to Tame Process Reverse Engineering in Data Analytics Pr...
Using Adaptive Scrum to Tame Process Reverse Engineering in Data Analytics Pr...
 
Data Modernization: Breaking the AI Vicious Cycle for Superior Decision-making
Data Modernization: Breaking the AI Vicious Cycle for Superior Decision-makingData Modernization: Breaking the AI Vicious Cycle for Superior Decision-making
Data Modernization: Breaking the AI Vicious Cycle for Superior Decision-making
 
It Takes an Ecosystem: How Technology Companies Deliver Exceptional Experiences
It Takes an Ecosystem: How Technology Companies Deliver Exceptional ExperiencesIt Takes an Ecosystem: How Technology Companies Deliver Exceptional Experiences
It Takes an Ecosystem: How Technology Companies Deliver Exceptional Experiences
 
Intuition Engineered
Intuition EngineeredIntuition Engineered
Intuition Engineered
 
The Work Ahead: Transportation and Logistics Delivering on the Digital-Physic...
The Work Ahead: Transportation and Logistics Delivering on the Digital-Physic...The Work Ahead: Transportation and Logistics Delivering on the Digital-Physic...
The Work Ahead: Transportation and Logistics Delivering on the Digital-Physic...
 
Enhancing Desirability: Five Considerations for Winning Digital Initiatives
Enhancing Desirability: Five Considerations for Winning Digital InitiativesEnhancing Desirability: Five Considerations for Winning Digital Initiatives
Enhancing Desirability: Five Considerations for Winning Digital Initiatives
 
The Work Ahead in Manufacturing: Fulfilling the Agility Mandate
The Work Ahead in Manufacturing: Fulfilling the Agility MandateThe Work Ahead in Manufacturing: Fulfilling the Agility Mandate
The Work Ahead in Manufacturing: Fulfilling the Agility Mandate
 
The Work Ahead in Higher Education: Repaving the Road for the Employees of To...
The Work Ahead in Higher Education: Repaving the Road for the Employees of To...The Work Ahead in Higher Education: Repaving the Road for the Employees of To...
The Work Ahead in Higher Education: Repaving the Road for the Employees of To...
 
Engineering the Next-Gen Digital Claims Organisation for Australian General I...
Engineering the Next-Gen Digital Claims Organisation for Australian General I...Engineering the Next-Gen Digital Claims Organisation for Australian General I...
Engineering the Next-Gen Digital Claims Organisation for Australian General I...
 
Profitability in the Direct-to-Consumer Marketplace: A Playbook for Media and...
Profitability in the Direct-to-Consumer Marketplace: A Playbook for Media and...Profitability in the Direct-to-Consumer Marketplace: A Playbook for Media and...
Profitability in the Direct-to-Consumer Marketplace: A Playbook for Media and...
 
Green Rush: The Economic Imperative for Sustainability
Green Rush: The Economic Imperative for SustainabilityGreen Rush: The Economic Imperative for Sustainability
Green Rush: The Economic Imperative for Sustainability
 
Policy Administration Modernization: Four Paths for Insurers
Policy Administration Modernization: Four Paths for InsurersPolicy Administration Modernization: Four Paths for Insurers
Policy Administration Modernization: Four Paths for Insurers
 
The Work Ahead in Utilities: Powering a Sustainable Future with Digital
The Work Ahead in Utilities: Powering a Sustainable Future with DigitalThe Work Ahead in Utilities: Powering a Sustainable Future with Digital
The Work Ahead in Utilities: Powering a Sustainable Future with Digital
 
AI in Media & Entertainment: Starting the Journey to Value
AI in Media & Entertainment: Starting the Journey to ValueAI in Media & Entertainment: Starting the Journey to Value
AI in Media & Entertainment: Starting the Journey to Value
 
Operations Workforce Management: A Data-Informed, Digital-First Approach
Operations Workforce Management: A Data-Informed, Digital-First ApproachOperations Workforce Management: A Data-Informed, Digital-First Approach
Operations Workforce Management: A Data-Informed, Digital-First Approach
 
Five Priorities for Quality Engineering When Taking Banking to the Cloud
Five Priorities for Quality Engineering When Taking Banking to the CloudFive Priorities for Quality Engineering When Taking Banking to the Cloud
Five Priorities for Quality Engineering When Taking Banking to the Cloud
 
Getting Ahead With AI: How APAC Companies Replicate Success by Remaining Focused
Getting Ahead With AI: How APAC Companies Replicate Success by Remaining FocusedGetting Ahead With AI: How APAC Companies Replicate Success by Remaining Focused
Getting Ahead With AI: How APAC Companies Replicate Success by Remaining Focused
 
Crafting the Utility of the Future
Crafting the Utility of the FutureCrafting the Utility of the Future
Crafting the Utility of the Future
 
Utilities Can Ramp Up CX with a Customer Data Platform
Utilities Can Ramp Up CX with a Customer Data PlatformUtilities Can Ramp Up CX with a Customer Data Platform
Utilities Can Ramp Up CX with a Customer Data Platform
 
The Work Ahead in Intelligent Automation: Coping with Complexity in a Post-Pa...
The Work Ahead in Intelligent Automation: Coping with Complexity in a Post-Pa...The Work Ahead in Intelligent Automation: Coping with Complexity in a Post-Pa...
The Work Ahead in Intelligent Automation: Coping with Complexity in a Post-Pa...
 

Dernier

Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Victor Rentea
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Safe Software
 

Dernier (20)

Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
CNIC Information System with Pakdata Cf In Pakistan
CNIC Information System with Pakdata Cf In PakistanCNIC Information System with Pakdata Cf In Pakistan
CNIC Information System with Pakdata Cf In Pakistan
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectors
 
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 AmsterdamDEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
DEV meet-up UiPath Document Understanding May 7 2024 Amsterdam
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with Milvus
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Ransomware_Q4_2023. The report. [EN].pdf
Ransomware_Q4_2023. The report. [EN].pdfRansomware_Q4_2023. The report. [EN].pdf
Ransomware_Q4_2023. The report. [EN].pdf
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
DBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor PresentationDBX First Quarter 2024 Investor Presentation
DBX First Quarter 2024 Investor Presentation
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
 

Information Lifecycle and the Actionability Test

  • 1. • Cognizant 20-20 Insights Information Lifecycle and the Actionability Test By scoring information generated internally and externally, and assessing its importance and maturity, organizations can better understand how and where to store key data for quick access and applicability to strategic planning. Executive Summary For the past several years, enterprises have been drowning in a deluge of data, and the sheer volume is growing exponentially. That translates to downstream reporting environments, where, for instance, data warehouses have become behemoth data dumps containing multiple terabytes and even petabytes of information. While it is fortunate that new technologies have made it possible to store this volume of information at a significantly reduced cost, professionals who are dealing with this data deluge still suffer from information overload. Information overload is caused by too much data created by too many sources. In fact, the sheer number of sources is increasing geometrically, as is the volume and complexity of information (i.e., formats, metadata, frequency, etc.) available for analysis. (See Figure 1, next page, for illustrative examples.) For years, businesses have focused on moving nonoperational data into archives through an orderly automated process or archiving the contents of retired systems. But rarely have they used information lifecycle techniques to ensure cognizant 20-20 insights | september 2013 that what is easily accessible in their operational systems is actionable. This white paper defines a process to ensure exactly that. According to IDC,1 the volume of information available will grow 50-fold over the next several years (see Figure 2, next page). But data volume is only part of the problem. The sheer number of tables, overlapping sources, references to differing formats of information and other issues are causing acute problems for organizations that utilize traditional techniques for analysis and reporting. The unbridled growth of warehouses as a means of servicing all constituencies has reached its limit of usefulness, and a better approach to managing the growth of data sources is required, such as allowing warehouse environments to grow organically to service the needs of all. This unmanaged organic growth is causing difficulty to those striving to glean insight from or discover anything useful in these behemoth warehouses. While big data makes it possible to store and retrieve insight from large data pools at an accelerated rate, it does not help people focus on actionable information nuggets. The economist
  • 2. The Reality of Information Overload 42% Organizations that say the paper records they are generating continue to increase. 11 Seconds to generate one petabyte of information, or the equivalent of 13 years of high definition video. 15.4 Source: Intel Corporation Source: AIIM survey of 512 AIIM members conducted in January and February, 2013. Millions of dollars in health insurance premiums overpaid for fiscal year 2012/13, half of which were to cover workers who were no longer employed. Source: Wisconsin Legislative Audit Bureau Figure 1 broad buckets. They do not generally cull out organizational trivia from warehouses, under the premise that this information may be needed at some future date. Herbert A. Simon had it right when he wrote, “In an information-rich world, the wealth of information means a dearth of something else: a scarcity of whatever it is that information consumes. What information consumes is rather obvious: it consumes the attention of its recipients. Hence a wealth of information creates a poverty of attention and a need to allocate that attention efficiently among the overabundance of information sources that might consume it.”2 It is exactly such focused attention that organizations need to properly derive insights from the explosion of data. And, most importantly, this is where an appropriate information lifecycle program comes in handy. According to Omar Tawakol, CEO of BlueKai, the data warehouse’s historical premise was to provide a pool of storage where trustworthy data could be parked for analysis.3 However, the explosive growth of data has made this pooling of data superfluous; the enormity of the end data state made the warehouse virtually unusable. An example of an actionable data warehouse is the one used by the U.S. Internal Revenue Service, which manages a historical base of 230 million annual tax returns and 50 billion tax- Most information lifecycle programs tackle basic needs, such as segregating data into several The Growing Data Universe Exabytes 40,000 30,000 20,000 10,000 2013 YE Digital Universe Estimate e m Year 2013 2014 2015 2016 Source: IDC Digital Universe Study, 2012 Figure 2 cognizant 20-20 insights 2 5,000
  • 3. related telephone requests in a single warehouse. Queries made against this warehouse help to: • Estimate the U.S. tax gap. • Predict identity theft, fraud • Information that is actionable and routinely used for decision-making, such as sales information, customer activities, working capital availability, disputed invoices stuck in receivables, consumer sentiment, market capitalization, fixed and variable costs of operations, competitive insight and innovation programs that help differentiate an organization from its competitors. and other tax issues. • Model financial risks. • Measure taxpayer burdens. • Simulate new tax laws and their effects on tax behavior. • Information that is actionable for a specific business need that may be routinely required, such as changes in long-term debt portfolios, changes in credit ratings, changes in market valuation, caustic social media coverage and changes in the regulatory climate. • Optimize the IRS staffing model. 4 Information Lifecycle Maturity Information lifecycle management (see Figure 3) can be viewed in terms of levels of maturity: • Information that is actionable for a short timeframe and may be historical trivia shortly afterwards, such as tracking the construction of a new plant, monitoring the launch of a new product, addressing raw materials reject rates from a supplier, monitoring spikes in healthcare costs and monitoring online groups for entries made by a disgruntled employee. • Immature, where the information lifecycle management (ILM) process begins. • Aware, where in most cases basic ILM maturity emerges by partitioning data onto static (not logged for updates) partitions. • Basic managed, where a basic ILM program is established. • Advanced • Information that is operationally relevant but not necessarily required as a primary source for decision-making, such as the dispensation of contact center calls, the safety reports of elevators and the detailed time and attendance logs of employees. managed, where actionable information is managed. • Governed, where the ILM program is governed. Understanding Data Consumption Organizations should assign a status to the candidate information to be managed by an information lifecycle program. These information priority statuses are as follows: • Information seen as historical trivia, such as long-lost organizational contacts, building maintenance reports for divested facilities, old detailed call detail records, ancient detailed journal entry logs that no longer align with the Information Lifecycle Maturity Information Maturity Lifecycle Stage Immature Aware Basic Managed Advanced Managed Governed Description of Stage Typical Stage Activities Database tables grow linearly with only a fraction of their contents usable. The information lifecycle management journey normally begins here. A need for lifecycle is present. The utilization of database partitioning is normally a first activity in satisfying ILM needs. A repeatable Information lifecycle management program is present. Tools are introduced to orchestrate the movement of information through the ILM processes. The information lifecycle management program is directed to deal with the deluge of information, much of which is never utilized. A proactive governance body orchestrates the information lifecycle processes. Figure 3 cognizant 20-20 insights 3
  • 4. chart of accounts and competitive insight on defunct companies. Understanding Information Gleaned Candidate information for an organization’s information lifecycle program can be obtained from a variety of sources: quo — which would require a swift determination of whether action is warranted and, if it is, execution of the appropriate coordinated action to protect, extract, originate or capture value for the organization (see Figure 4). Data can originate from: • Indirect sources, such as Web sites maintained • A log, or formatted activity logs made available • Direct sources, such as applications, logs and • A from a call center, Web site, RFID, mobile device or other facility that records its activity into a log. by partnering organizations, distributors and retailer reports, advertisers, regulators and watchdogs. other information from Web sites, logs and other information from mobile facilities and logs and other information from call centers that your organization originates. • Externally sourced information, such as data collected from RFID tags, mobile devices, social media, modifications made on tracked Web sites, traditional and nontraditional news media, public relations bulletins, acquired intelligence, affiliate referrals, etc. message, received from an internal or external source, such as a SWIFT message, an EDI transaction or other message based sources. • A stream, or an asynchronous source of infor- mation that streams information to a specific listening target, such as Twitter, a market ticker, newswires and other sources. • A document, or any form of object that is available as an information source, such as an XML file, a public relations briefing, a weather map, a schematic or other non-tabular information source. Each of these sources may contain superfluous information, or may signal a change to the status Information Source Formats Possible Source Formats Information Retailer Reports Advertisers Watchdogs and Regulators Internal Sources Direct Web Sites Mobile Facilities Call Centers Mobile Facilitated RFID Externally Obtained Social Media News and PR Media Acquired Intelligence Affiliate Referrals Other Figure 4 cognizant 20-20 insights 4 Tabular External Indirect Tabular Internal Distributor Reports Document Partner Web Stream Message Log Delivery Models
  • 5. • A tabular source of information, normally housed in databases and originated from sources internal to an organization (internal ERP systems, spreadsheets, CRM systems, etc.) or external sources (such as vendors, customers and industry associations). The information lifecycle management program must be able to decipher which of these data sources are usable for long-term analysis and which ones are relevant only for situational use and temporal event detection. For example, the session statistics from a Web log may be usable for long-term use; however, the Web log detail may be usable only until it is transformed. At this point, this data may be very well a candidate for archiving. The business information model should serve as a primary tool for determining placement of information classes in the information management lifecycle (see Figure 5). Scoring information consumed by an information lifecycle program prioritizes information deemed mandatory for operational use, its importance in managing operational risks, its importance in enhancing sustainable value (capturing, extending or creating sustainable value), defining organizational strategies or diffusing disruptive events. Data Attributes and the Information Lifecycle The information lifecycle program should consider the following characteristics of data: A Model for Deciphering Information Actionability • The ability to recreate information should it be Information captured for analysis can be used for managing operational risks, diffusing disruptive events, enhancing sustainable value and defining enterprise strategies. The information lifecycle must determine which of the data captured about customers, media, competitors, markets, financing and geographies will be useful in identifying patterns or deriving insight for the organization. • The veracity of sources transformed into what incorrectly or inadvertently destroyed. is being archived. • Actionability of the candidate information. • The information’s priority status. The scoring will facilitate whether the information is: • Actively maintained (and thus should reside in active online partitions). Managing Operational Risks n tio ra Service Customers Be Value Chain ne fit Financing s Enhancing Sustainable Value Geographies Ge Product le Focused Information op ts Markets Process Capabilities Barriers Diffusing Disruptive Events M e cs tri Col lab o Business Information Model Defining Enterprise Strategies Figure 5 cognizant 20-20 insights 5 Pe
  • 6. • Required Assigning Scores for Lifecycle Placement Data • Required, but has a low information lifecycle The primary reason for focusing on the actionability of information — based on its format and ease of recreation, its age and consumption pattern — is to derive a mechanism to assign a score that can be used to place information into the buckets of storage managed by an information lifecycle program. regularly but historical in nature (static online partitions). pool scoring (near-line storage). • Required infrequently but must be retained for regulatory or other needs (archived). • Is no longer required for regulatory needs and is not accessed (to be destroyed). A viable approach for utilizing data attributes in an information lifecycle program is automating the lifecycle scoring algorithm (see Figure 5) within the information lifecycle orchestration application. It is important to note that there will be differing retention requirements for information utilized by an organization, which may serve as information lifecycle overrides. For example, vested Scoring Data Attributes Managed Information Lifecycle Pools Index Factors in Scoring the Placement of Information Active Online (100-86) Static Online (85-76) Operationally Critical Near line (75-61) Archived (60-19) Destructed (18-11) Operational Criticality scored 10-15 (1=not critical, 15=highly critical Information Priority Status Information Priority Status scored 10-15 (1=organizational trivia, 15=highly actionable Analytic Criticality (1=not required, 10=critical) Analytic Criticality scored 7-10 Able to Recreate (1=easily, 5=not able) Veracity of Sources (1=Low, 10=High) Storage Platform (1=other,2=Mainframe, 3=portal, 4=database, 5=big data) Regulatory Availability (1=not required, 15=critical) Regulatory Availability scored 3-15 Age of Data (1=>3 years, 3=1 -3 years, 5=<1 year) Information Format (5=Tabular, 4=Document, 3=Message, 2=Log, 1=Stream) Used in Active Program (1=Not used, 5=actively used) Used in active program scored 2-5 Frequency of Access (1=rarely if ever, 10 = regularly) Frequency of Access scored 3-10 Figure 5 cognizant 20-20 insights 6
  • 7. pension recipient information must be saved for employees who have left an organization because they will be receiving pensions at a future date. Thus, the information about them has operational (funding needs for pension plans) and regulatory (pension regulations) retention requirements that will override some information lifecycle placement scores — i.e., you cannot destroy data on potential pensioners who are less than 60 years old, vested and have not yet received pensions. ILM Implementation and Testing Approaches Using a matrix similar to the one in Figure 5, a numeric scoring will be derived, which should be used to determine the appropriate placement of data within the information lifecycle. To control the scoring of information for use in analytics and regulatory retention requirements, a simple scoring algorithm can be deployed to enter data into the organization’s information management lifecycle management software. ILM Governance Just as important as moving data through the lifecycle into its archival target is the destruction of data no longer required nor usable by the organization. There are countless examples of organizations that have had old records subpoenaed and used for tax and HR disputes and patent infringement disagreements, among other reasons. When information is no longer required, it should be disposed of. The processes that govern an information lifecycle program should be automated in such a way that are repeatable and auditable (see Figure 6). These processes should ensure that information, particularly that which is moved to near-line storage or archived, can be retrieved and used and that information flagged for destruction cannot be utilized after deletion. The role of governance is simple in the information lifecycle program. Governance facilitates the scoring of the placement of information into the appropriate information lifecycle pools and informs the proper execution of the information lifecycle program. Tasks specific to the administration of the information lifecycle program are as follows: • Scoring the managed information lifecycle pools matrix. • Measuring and communicating the program’s cost and the associated cost avoidance. • Measuring the risk associated with the scoring of specific regulated and high-risk information and rescoring should an elevated risk be derived. ILM Construction Implementation Approach ILM Governance Information lifecycle configuration. Business review of information necessity and relevance. Develop queries to validate the ILM pools of data (near-line, archived, etc.) Validate that ILM managed data is usable. Establish test cases to validate changes. Spot profiling review to validate proper information placement. Review action ability scoring of ILM managed information. Identify changes to the ILM configuration. Review rules governing the placement of information. Execute the archival recovery in test mode and destroyed data recovery attempt. Ensure that archived data is recoverable and destroyed data is not. Acceptance test Figure 6 cognizant 20-20 insights 7 Implementation 1. Code movement to production. 2. Trigger application retirement, if applicable.
  • 8. • Measuring • Review Organizations that have effective information lifecycle governance programs and high degrees of information lifecycle maturity routinely revisit their information lifecycle program and adjust the information lifecycle pools matrix as part of the annual planning cycle. This enables organizations to: • Regularly and communicating information lifecycle maturity model placement of the organization. • Ensure that a partitioning program administered through an information lifecycle product (Solix, Optim, Informatica ILM, Autonomy, Opentext, Storetrends, etc.) exists and is regularly monitored for effectiveness. key technologies in the information lifecycle program and monitor storage, floor space, manpower and power consumption statistics; and regularly update to newer technologies where they make economic sense and are not operational disruptions. monitor the effectiveness of thirdparty information lifecycle products and align to make the capabilities of the information lifecycle program effective and make optimal use of the information lifecycle capabilities of vendors participating in the program. Footnotes 1 IDC Digital Universe Study, December 2012. 2 H.A. Simon, “Designing Organizations for an Information-Rich World,” in Martin Greenberger, ed., Computers, Communication, and the Public Interest, pp. 40–41, Johns Hopkins Press, Baltimore, 1971. 3 Omar Tawakol, “The Death of the Data Warehouse and the Age of Activation,” Huffington Post, July 12, 2012. 4 Internal Revenue Service Compliance Data Warehouse case study, General Accounting Office, 2011. About the Author Mark Albala is the Service Line Lead of Cognizant’s Data On-Demand Service Line, an organization that provides services and products that support the effectiveness and efficiency of databases and other information platforms used for managing information as part of Cognizant’s overall capabilities in enterprise information management. A graduate of Syracuse University, Mark has held senior consulting, thought leadership, advanced technical and business development roles for organizations focused on the disciplines of business intelligence, governance and data warehousing. He can be reached at Mark.Albala@cognizant.com. About Cognizant Cognizant (NASDAQ: CTSH) is a leading provider of information technology, consulting, and business process outsourcing services, dedicated to helping the world’s leading companies build stronger businesses. Headquartered in Teaneck, New Jersey (U.S.), Cognizant combines a passion for client satisfaction, technology innovation, deep industry and business process expertise, and a global, collaborative workforce that embodies the future of work. With over 50 delivery centers worldwide and approximately 164,300 employees as of June 30, 2013, Cognizant is a member of the NASDAQ-100, the S&P 500, the Forbes Global 2000, and the Fortune 500 and is ranked among the top performing and fastest growing companies in the world. Visit us online at www.cognizant.com or follow us on Twitter: Cognizant. World Headquarters European Headquarters India Operations Headquarters 500 Frank W. Burr Blvd. Teaneck, NJ 07666 USA Phone: +1 201 801 0233 Fax: +1 201 801 0243 Toll Free: +1 888 937 3277 Email: inquiry@cognizant.com 1 Kingdom Street Paddington Central London W2 6BD Phone: +44 (0) 20 7297 7600 Fax: +44 (0) 20 7121 0102 Email: infouk@cognizant.com #5/535, Old Mahabalipuram Road Okkiyam Pettai, Thoraipakkam Chennai, 600 096 India Phone: +91 (0) 44 4209 6000 Fax: +91 (0) 44 4209 6060 Email: inquiryindia@cognizant.com © ­­ Copyright 2013, Cognizant. All rights reserved. No part of this document may be reproduced, stored in a retrieval system, transmitted in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the express written permission from Cognizant. The information contained herein is subject to change without notice. All other trademarks mentioned herein are the property of their respective owners.