Commendable Comments (Part 6)

Last September, and on the exact day of the sixth mensiversary (yes, that’s a real word, look it up) of my blog, I started this series as an ongoing celebration of the truly commendable comments that I regularly receive from my heroes—my readers.

 

Commendable Comments

On The Circle of Quality, Kelly Lautt commented:

“One of the offerings I provide as a consultant is around data readiness specifically for BI.  Sometimes, you have to sneak an initial data quality project into a company tightly connected to a project or initiative with a clear, already accepted (and budgeted) ROI.  Once the client sees the value of data quality vis a vis the BI requirements, it is easier to then discuss overall data quality (from multiple perspectives).

And, I have to add, I do feel that massive, cumbersome enterprise DQ programs sometimes lose the plot by blindly ‘improving’ data without any value in sight.  I think there has to be a balance between ignoring generalized DQ versus going overboard when there will be a diminishing return at some point.

Always drive effort and investment in any area (including DQ) from expected business value!”

On The Poor Data Quality Jar, Daragh O Brien commented:

“We actually tried to implement something like this with regard to billing data quality issues that created compliance problems.  Our aim was to have the cost of fixing the problem borne by the business area which created the issue, with the ‘swear jar’ being the budget pool for remediation projects.

We ran into a few practical problems:

1) Many problems ultimately had multiple areas with responsibility (line-of-business workers bypassing processes, IT historically ‘right-sizing’ scope on projects, business processes and business requirements not necessarily being defined properly resulting in inevitable errors)

2) Politics often prevented us from pushing the evidence we did have too hard as to the weighting of contributions towards any issue.

3) More often than not it was not possible to get hard metrics on which to base a weighting of contribution, and people tended to object to being blamed for a problem that was obviously complex with multiple inputs.

That said, the attempt to do it did help us to:

1) Justify our ‘claims’ that these issues were often complex with multiple stakeholders involved.

2) Get stakeholders to think about the processes end-to-end, including the multiple IT systems that were involved in even the simplest process.

3) Ensure we had human resources assigned to projects because we had metrics to apply to a business case.

4) Start building a focus on prevention of defect rather than just error detection and fix.

We never got around to using electric shocks on anyone.  But I’d be lying if I said it wasn’t a temptation.”

On The Poor Data Quality Jar, Julian Schwarzenbach commented:

“As data accuracy issues in some cases will be identified by front line staff, how likely are they going to be to report them?  Whilst the electric chair would be a tempting solution for certain data quality transgressions, would it mean that more data quality problems are reported?

This presents a similar issue to that in large companies when they look at their accident reporting statistics and reports of near misses/near hits:

* Does a high number of reported accidents and near hits mean that the company is unsafe, or does it mean that there are high levels of reporting coupled with a supportive, learning culture?

* Does a low number of reported accidents and near hits mean that the company is safe, or does it mean that staff are too scared of repercussions to report anything?

If staff risk a large fine/electric shock for owning up to transgressions, they will not do it and will work hard to hide the evidence, if they can.

In organizational/industrial situations, there are often multiple contributing factors to accidents and data quality problems.  To minimize the level of future problems, all contributory causes need to be identified and resolved.  To achieve this, staff should not be victimized/blamed in any way and should be encouraged to report issues without fear.”

On The Scarlet DQ, Henrik Liliendahl Sørensen commented:

“When I think about the root causes of many of the data quality issues I have witnessed, the original data entry was actually made in good faith by people trying to make data fit for the immediate purpose of use.  Honest, loyal, and hardworking employees striving to get the work done.

Who are the bad guys then?  Either it is no one or everyone or probably both.

When I have witnessed data quality problems solved it is most often done by a superhero taking the lead in finding solutions.  That superhero has been different kinds of people.  Sometimes it is a CEO, sometimes a CFO, sometimes a CRM-manager, sometimes it is anyone else.”

On The Scarlet DQ, Jacqueline Roberts commented:

“I work with engineering data and I find that the users of the data are not the creators of data, so by the time that data quality is questioned the engineering project has been completed, the engineering teams have been disbanded and moved on to other projects for other facilities. 

I am sure that if the engineers had to put the spare part components on purchasing contracts for plant maintenance, the engineers would start to understand some of the data quality issues such as incomplete part numbers or descriptions, missing information, etc.”

On The Scarlet DQ, Thorsten Radde commented:

“Is the question of ‘who is to blame’ really that important?

For me, it is more important to ask ‘what needs to be done to improve the situation.’

I don’t think that assigning blame helps much in improving the situation.  It is very rare that people cooperate to ‘cover up their mistakes.’  I found it more helpful to point out why the current situation is ‘wrong’ and then brainstorm with people on what can be done about it - which additional conventions are required, what can be checked automatically, if new functionality is needed, etc.

Of course, to be able to do that, youve got to have the right people on board that trust each other - and the blame game doesn’t help at all.  Maybe you need a ‘blame doll’ that everyone can beat in order to vent their frustrations and then move on to more constructive behavior?”

On Can Enterprise-Class Solutions Ever Deliver ROI?, James Standen commented:

“Fantastic question.  I think the short answer of course as always is ‘it depends’.

However, what’s important is exactly WHAT does it depend on.  And I think while the vendors of these solutions would like you to believe that it depends on the features and functionality of their various applications, that what it all depends on far more is the way they are installed, and to what degree the business actually uses them.

(Insert buzz words here like: ‘business process alignment’, ‘project ownership’, ‘Business/IT collaboration’)

But if you spend Gazillions on a new ERP, then customize it like crazy to ensure that none of your business processes have to change and none of your siloed departments have to talk to each other (which will cost another gazillion in development and consulting by the way), which will then ensure that ongoing maintenance and configuration is more expensive as well, and will eliminate any ability to use pre-built business intelligence solutions etc., etc.  Your ROI is going to be a big, negative number.

Unfortunately, this is often how it’s done.  So my first comment in this debate is - If enterprise systems enable real change and optimization in business processes, then they CAN have ROI.  But it’s hard. And doesn't happen often enough.”

On Microwavable Data Quality, Dylan Jones commented:

“Totally agree with you that data cleansing has been by far the most polarizing topic featured on our site since the launch.  Like you, I agree that data governance is a marathon not a sprint but I do object to a lot of the data cleansing bashing that goes on.

I think that sometimes we should give people who purchase cleansing software far more credit than many of the detractors would be willing to offer.  In the vast majority of cases data cleansing does provide a positive ROI and whilst some could argue it creates a cost base within the organization it is still a step in the direction of data quality maturity.

I think this particular debate is going to run and run however so thanks for fanning the flames.”

On The Challenging Gift of Social Media, Crysta Anderson commented:

“This is the biggest mindshift for a lot of people.  When we started Social Media, many wanted to build our program based only on the second circle - existing customers.  We had to fight hard to prove that the third circle not only existed (we had a hunch it did), but that it was worth our time to pursue.  Sure, we can't point to a direct sales ROI, but the value of building a ‘tribe’ that raises the conversation about data quality, MDM, data governance and other topics has been incredible and continues to grow.”

Thank You

Thank you all for your comments.  Your feedback is greatly appreciated—and truly is the best part of my blogging experience.

Since there have been so many commendable comments, please don’t be offended if one of your comments wasn’t featured. 

Please keep on commenting and stay tuned for future entries in the series.

 

Related Posts

Commendable Comments (Part 5)

Commendable Comments (Part 4)

Commendable Comments (Part 3)

Commendable Comments (Part 2)

Commendable Comments (Part 1)

 

Follow OCDQ

For more blog posts and commendable comments, subscribe to OCDQ via my RSS feed, my E-mail updates, or Google Reader.

You can also follow OCDQ on Twitter, fan the Facebook page for OCDQ, and connect with me on LinkedIn.


Data, data everywhere, but where is data quality?

“Two young fish are swimming along when they happen to meet an older fish swimming the other way, who nods at them and says: ‘Morning, boys.  How’s the water?’  And the two young fish swim on for a bit, then eventually one of them looks over at the other and goes: ‘What the hell is water?’”

The acclaimed novelist David Foster Wallace told that story during a speech he delivered at Kenyon College in 2005.  Although he certainly wasn’t speaking on the topic of data management, I believe that his story can easily be adapted as a data metaphor:

“Two young kids are walking along, tweeting and uploading new pictures to Facebook on their iPhones, when they happen to meet an older man walking the other way checking his work e-mail on his BlackBerry, who nods at them and says: ‘Morning, boys.  How’s the data?’  And the two young kids walk on for a bit, then eventually one of them looks over at the other and goes: ‘What the hell is data?’”

My point is that what once was a seemingly esoteric word (“data”) used mostly by computer geeks such as myself, has now so thoroughly pervaded mainstream culture that we hardly seem to notice we are quite literally swimming in data on a daily basis.

 

Why Data Matters

In his recent blog post, Rich Murnane was hit with the realization that data isn’t for data geeks anymore.  The post included an excellent IBM video (and commercial) about “Why Data Matters” that states every day we are creating fifteen petabytes of data, which is eight times as much data as there is in all of the libraries in the United States combined. 

Data matters because everything—and not just the rows in our relational databases and spreadsheets, but also our status updates from Facebook and Twitter, our blog posts, and even most of our daily conversations—is data. 

The growing challenge is can we extract meaningful insights from these vast and veritable oceans of unrelenting data volumes, and use those insights to make better decisions in near real-time in order to positively impact the various aspects of our lives.

 

Paradoxical Business Situation

Even in the business world, where data management used to be viewed solely as a concern for those computer geeks down in IT, more and more people all throughout more and more organizations are coming to view data as a strategic corporate asset.

In his recent Network World article Data Everywhere, But Not Enough Smart Management, Thomas Wailgum described the “data, data everywhere” phenomenon as “an awe-inspiring and unprecedented push and pull of data and information needs.”  Wailgum described the push as a growing surge of terabytes of data flooding enterprise systems and applications, and the pull as the growing demand from users for sweeping, individualized access to analytics and business information.

However, just because data is flowing everywhere doesn’t automatically mean that data quality is sure to follow.

Wailgum cites research from a recent Forbes survey where executives reported that the “bad data problem” is currently estimated to be costing their organizations between five and twenty million dollars annually, which leads him to ask the question:

“If everyone agrees on the strategic importance of data and information management, and everyone knows what the negative consequences are, then why are there still so many problems?” 

Wailgum calls this the “paradoxical business situation” and cited survey results indicating “fragmented data ownership” is the single biggest roadblock to successful enterprise information management.  Nearly 80% of IT managers said data quality was their responsibility, whereas nearly 75% of business (finance, sales, and marketing) managers said it was their responsibility.

“While IT managers largely concede that information is the users’ not theirs, they take the position that data and information management systems are under IT’s purview,” concludes the survey.  “This differing perspective puts IT and business executives in conflicting camps, particularly when it comes to data quality.”

This debate over data ownership reminded me of the great discussion sparked by a recent Henrik Liliendahl Sørensen blog post questioning whether “data owner” was a bad word.  Many commenters agreed that “data stewardship” was more relevant and that although data quality is a shared responsibility for the entire enterprise, corporate culture is far more challenging than what can amount to a largely semantic argument over the proper use of terminology such as “data ownership” or “ data stewardship.”

 

Why Data Quality Matters

As I posited in The Circle of Quality, an organization’s success is measured by the quality of its results, which are dependent on the quality of its business decisions, which rely on the quality of its information, which is based on the quality of its data. 

Therefore, data quality matters because high quality data serves as a solid foundation for business success.

Organizations are not only facing the challenging realities that data is everywhere and its burgeoning volumes continue to rise, but also that data is no longer limited to the traditional structured forms stored in relational databases.  Unstructured data from social media, the Internet, and mobile devices are contributing an abundant new source to the enterprise’s information ocean.

In The Rime of the Ancient Mariner, Samuel Taylor Coleridge wrote:

“Day after day, day after day,
We stuck, nor breath nor motion;
As idle as a painted ship
Upon a painted ocean.

Water, water, everywhere,
And all the boards did shrink;
Water, water, everywhere,
Nor any drop to drink.”

When data is abundant, but data quality remains scarce, then the thirst to acquire knowledge and insight remains unquenched, and data hangs like a heavy albatross around the enterprise’s neck.

 

Related Posts

The Circle of Quality

Beyond a “Single Version of the Truth”

Poor Data Quality is a Virus

DQ-Tip: “Don't pass bad data on to the next person...”

The Only Thing Necessary for Poor Data Quality

Hyperactive Data Quality (Second Edition)

The General Theory of Data Quality

Data Governance and Data Quality

The Data-Information Continuum

 

Follow OCDQ

If you enjoyed this blog post, then please subscribe to OCDQ via my RSS feed, my E-mail updates, or Google Reader.

You can also follow OCDQ on Twitter, fan the Facebook page for OCDQ, and connect with me on LinkedIn.


The Fellowship of #FollowFriday

During the dawn of the Second Age of Digital-Earth, in the land of Twitter there was formed a group of like-minded tweeps who were well known for their wisdom about Data Quality, Data Governance, Master Data Management, and Business Intelligence.

They battled against the dark forces of poor data quality, undisciplined organizations, multiple conflicting versions of the truth, flawed business decisions, vast boiling oceans of unmanaged data assets, uncontrolled costs, and unmitigated compliance risks.

Collectively, these valiant heroes were known as: The Fellowship of FollowFriday.

Okay, so clearly I am a total dork—geek, nerd, and dweeb are also completely acceptable alternatives.

J. R. R. Tolkien's The Lord of the Rings three-volume book and Peter Jackson’s adapted movie trilogy were awe inspiring epics, and also the theme of this blog post about FollowFriday, the weekly tradition of recommending great folks to follow on Twitter.

Please note that simply for the purposes of organizing the following lists, I have made the United States the kingdom of Gondor, Canada the kingdom of Rohan, and all of Europe collectively The Shire.  No offense intended to my tweeps from other lands.

I hope that everyone has a great FollowFriday and an even greater weekend.  See you all around the Twittersphere.

 

Tweeps of Gondor

 

Tweeps of Rohan

 

Tweeps of The Shire

 

Related Posts

Social Karma (Part 7)

The Wisdom of the Social Media Crowd

The Twitter Clockwork is NOT Orange

Video: Twitter #FollowFriday – January 15, 2010

Video: Twitter Search Tutorial

Live-Tweeting: Data Governance

Brevity is the Soul of Social Media

If you tweet away, I will follow

Tweet 2001: A Social Media Odyssey

Microwavable Data Quality

Data quality is definitely not a one-time project, but instead requires a sustained program of enterprise-wide best practices that are best implemented within a data governance framework that “bakes in” defect prevention, data quality monitoring, and near real-time standardization and matching services—all ensuring high quality data is available to support daily business decisions.

However, implementing a data governance program is an evolutionary process requiring time and patience.

Baking and cooking also require time and patience.  Microwavable meals can be an occasional welcome convenience, and if you are anything like me (my condolences) and you can’t bake or cook, then microwavable meals can be an absolute necessity.

Data cleansing can also be an occasional (not necessarily welcome) convenience, or a relative necessity (i.e., a “necessary evil”).

Last year on Data Quality Pro, Dylan Jones hosted a great debate on the necessity of data cleansing, which is well worth reading, especially since the over 25 (and continuing) comments it received proves it is a polarizing topic for the data quality profession.

I reheated this debate (using the Data Quality Microwave, of course) earlier this year with my A Tale of Two Q’s blog post, which also received many commendable comments (but far less than Dylan’s blog post—not that I am counting or anything).

Similarly, a heated debate can be had over the health implications of the microwave.  Eating too many microwavable meals is certainly not healthy, but I have many friends and family who would argue quite strongly for either side of this “food fight.”

Both of these great debates can be as deeply polarizing as Pepsi vs. Coke and Soccer vs. Football.  Just for the official record, I am firmly for both Pepsi and Football—and by Football, I mean NFL Football—and firmly against both Coke and Soccer. 

Just as I advocate that everyone (myself included) should learn how to cook, but still accept the eternal reality of the microwave, I definitely advocate the implementation of a data governance program, but I also accept the eternal reality of data cleansing.   

However, my lawyers have advised me to report that beta testing for an actual Data Quality Microwave has not been promising.

 

Related Posts

A Tale of Two Q’s

Hyperactive Data Quality (Second Edition)

The General Theory of Data Quality

 

Follow OCDQ

If you enjoyed this blog post, then please subscribe to OCDQ via my RSS feed, my E-mail updates, or Google Reader.

You can also follow OCDQ on Twitter, fan the Facebook page for OCDQ, and connect with me on LinkedIn.


The Scarlet DQ

The Scarlet DQ

The Scarlet DQ is the superhero name of Jill Wanless (aka sheezaredhead).

Just kidding—I would never reveal a superhero’s secret identity.

Although I was never a big fan of the book, the title of this blog post is inspired by The Scarlet Letter by Nathaniel Hawthorne, where the novel’s protagonist Hester Prynne is forced to wear The Scarlet Letter A as a badge of shame for committing the act of adultery, which lead to the birth of her daughter Pearl.

The book came to mind while I was reading the commendable comments received last week on The Poor Data Quality Jar, where a recurring theme was the valid criticism of the “public humiliation” aspect of having employees put money into the jar when they contribute to the organization's poor data quality.

Using such an approach to help organizations illustrate the costs of poor data quality is equivalent to making the offenders wear The Scarlet DQ as a badge of shame, which will only make it far more likely that data quality issues will not be reported at all.

But even without my “swear jar” inspired idea, I think that the fear of public humiliation is what prevents poor data quality from being acknowledged by many organizations, which often leads to a major data quality related crisis that “no one saw coming.” 

For example, if you are in need of some quiet time alone for taking a good power nap in a conference room, then try scheduling a meeting to discuss known data quality issues and their root causes.  If your organization is like most, then you could probably book one of those really nice conference rooms with the big comfy reclining chairs—because nobody will attend your meeting.

Data quality can be somewhat of a taboo topic.  Many organization assume that their data quality must be “good enough” otherwise “how could we possibly still be in business?”  Nobody likes to talk about data quality for one simple reason:

If data quality issues exist (and they do), then no one wants to be blamed for causing or failing to fix them.

It’s as if everyone is afraid that they will be forced to wear The Scarlet DQ.

 

This is one of the many human dynamics that can render even the best technology and proven methodology completely useless. 

 

What Say You?

Please post a comment and share your recommendations about how to foster an environment in which poor data quality can be reported freely without fear of blame or reprisal.  All viewpoints are welcome.  Nathaniel Hawthorne references are not required.

 

Related Posts

The Poor Data Quality Jar

The Third Law of Data Quality

The Dumb and Dumber Guide to Data Quality

Not So Strange Case of Dr. Technology and Mr. Business

You're So Vain, You Probably Think Data Quality Is About You

 

Follow OCDQ

If you enjoyed this blog post, then please subscribe to OCDQ via my RSS feed, my E-mail updates, or Google Reader.

You can also follow OCDQ on Twitter, fan the Facebook page for OCDQ, and connect with me on LinkedIn.


Subterranean Computing

Cloud computing continues to receive significant industry buzz and endorsements from many industry luminaries:

  • Tim O'Reilly of O'Reilly Media calls cloud computing “the platform for all computing.”
  • Connor MacLeod of the Clan MacLeod says “there can be only one—and that one is cloud computing.”  
  • Marc Benioff of SalesForce.com refers to companies in the “anti-cloud crowd” as “innovationless.”
  • Lando Calrissian of Cloud City calls anyone not using cloud computing a “slimy, double-crossing, no-good swindler.”

Therefore, I was happy to hear a cogent alternative viewpoint from a member of the “anti-cloud crowd” when I recently interviewed Sidd Finch, the Founder and President of the New York based startup company Kremvax, which recently secured another $4.1 billion in venture capital to pursue an intriguing alternative to cloud computing called Subterranean Computing.

 

The Truth about Cloud Computing

Mr. Finch began the interview by discussing some of the common criticisms of cloud computing, which include issues such as data privacy, data protection, and data security.  However, I was most intrigued by the new research Mr. Finch cited from Professor Nat Tate of the College of Nephology at the University of Southern North Dakota at Hoople.

According to Professor Tate, here is the truth about cloud computing:

  • Cloud computing's viability depends greatly on the type of cloud, not public or private, but rather cirrus, stratus, or cumulus.
  • Cirrus clouds are not good for data privacy concerns because they tend to be wispy and therefore completely transparent.
  • Stratus clouds are not good for data protection concerns because “data drizzling” occurs frequently and without warning. 
  • Cumulus clouds are not good for data security concerns because “fair weather clouds” disperse at the first sign of trouble. 

 

The Underlying Premise of Subterranean Computing

Later in the interview, Mr. Finch described the underlying premise of subterranean computing:

“Instead of beaming your data up into the cloud, bury your data down underground.”  

According to Mr. Finch, here are the basic value propositions of subterranean computing:

  • Subterranean computing's viability is limited only to your imagination (but real money is required, and preferably cash).
  • Data privacy is not a concern because your data gets buried in its own completely (not virtually) private hole in the ground.
  • Data protection is not a concern because once it is buried, your data will never be used again for any purpose whatsoever.
  • Data security is not a concern, but for an additional fee, we bury your data where nobody will ever find it (we know a guy).

 

Brown is the new Green

Environmentally sustainable computing (i.e., “Green IT”) is another buzzworthy industry trend these days.  Reduce your carbon footprint, utilize electricity more efficiently, evaluate alternative power sources, and leverage recyclable materials. 

All great ideas.  But according to Mr. Finch, subterranean computing takes it to the next level by running entirely on geothermal power, a sustainable and renewable energy source, as well as converting your databases into Composting Data Stores (CDS).

In subterranean computing, your data is buried deep underground, where CDS can draw the very minimal amount of power it requires directly from the heat emanating from the Earth's core.  The CDS biodegradable data format (BDF) also minimizes your data storage requirements by automatically composting old data, which creates the raw material used to store your new data.

In the words of Kremvax customer and award-eligible environmentalist Isaac Bickerstaff: “brown is the new green.” 

Bickerstaff is the Lord Mayor of the English village of Spiggot, which has “gone subterranean” with its computing infrastructure.

 

Conclusion

So which new industry trend will your organization be implementing this year: cloud computing or subterranean computing? 

Well, before you make your final decision, please be advised that Industry Analyst Lirpa Sloof has recently reported rumors are circulating that Larry Ellison of Oracle is planning on announcing the first Cloud-Subterranean hybrid computing platform at the Oracle OpenWorld 2010 conference, which is also rumored to be changing its venue from San Francisco to Spiggot.

But whenever you're evaluating new technology, remember the wise words from Subterranean Homesick Blues by Bob Dylan:

“You don’t need a weatherman to know which way the wind blows.”

The Poor Data Quality Jar

The Poor Data Quality Jar

Today I am pondering whether or not the venerable tradition of The Swear Jar could be adapted to help organizations illustrate the costs of poor data quality.

As an example for those unfamiliar with the concept, my family used a swear jar when I was growing up.  Anytime a family member swore (i.e., used profanity), they put an amount of money into the jar based on the severity of the swear.

Of course in my family, what exactly constituted “profanity” as well as what the severity of a particular swear should be would often cause considerable debate, which somewhat ironically lead to the increased use of unquestionable profanity.

Therefore, The Swear Jar was far from a perfect system (at least for my family). 

But I am still imaging every organization instituting The Poor Data Quality Jar.

When an employee contributes to the organization's poor data quality, they put an amount of money into the jar based on the severity of the data quality issue, and perhaps with an increasing scale to be more punitive to repeat offenders.

Do you think The Poor Data Quality Jar can help your organization?  If so, how much would you charge for different types of data quality issues?  How would you determine the severity (i.e., financial impact) of each data quality issue?

 

Photo via Flickr (Creative Commons License) by: Karen Roe


Enterprise Data World 2010

Enterprise Data World 2010

Enterprise Data World 2010 was held March 14-18 in San Francisco, California at the Hilton San Francisco Union Square.

Congratulations and thanks to Tony Shaw, Maya Stosskopf, the entire Wilshire Conferences staff, as well as Cathy Nolan and everyone with DAMA International, for their outstanding efforts on delivering yet another wonderful conference experience.

I wish I could have attended every session on the agenda, but this blog post provides some quotes from a few of my favorites.

 

Applying Agile Software Engineering Principles to Data Governance

Conference session by Marty Moseley, CTO of Initiate Systems, an IBM company.

Quotes from the session:

  • “Data governance is 80% people and only 20% technology”
  • “Data governance is an ongoing, evolutionary practice”
  • “There are some organizational problems that are directly caused by poor data quality”
  • “Build iterative 'good enough' solutions – not 'solve world hunger' efforts”
  • “Traditional approaches to data governance try to 'boil the ocean' and solve every data problem”
  • “Agile approaches to data governance laser focus on iteratively solving one problem at a time”
  • “Quality is everything, don't sacrifice accuracy for performance, you can definitely have both”

Seven iterative steps of Agile Data Governance:

  1. “Form the Data Governance Board – Small guidance team of executives who can think cross-organizationally”
  2. “Define the Problem and the Team – Root cause analysis, build the business case, appoint necessary resources”
  3. “Nail Down Size and Scope – Prioritize the scope in order to implement the current iteration in less than 9 months”
  4. “Validate Your Assumptions – Challenge all estimates, perform data profiling, list data quality issues to resolve”
  5. “Establishing Data Policies – Measurable statements of 'what must be achieved' for which kinds of data”
  6. “Implement the data quality solution for the current iteration”
  7. “Evaluate the overall progress and plan for the next iteration”

 

Monitor the Quality of your Master Data

Conference session by Thomas Ravn, MDM Practice Director at Platon.

Quotes from the session:

  • “Ensure master data is taken into account each and every time a business process or IT system is changed”
  • “Web forms requiring master data attributes can NOT be based on a single country's specific standards”
  • “There is no point in monitoring data quality if no one within the business feels responsible for it”
  • “The greater the business impact of a data quality dimension, the more difficult it is to measure”
  • “Data quality key performance indicators (KPI) should be tied directly to business processes”
  • “Implement a data input validation rule rather than allow bad data to be entered”
  • “Sometimes the business logic is too ambiguous to be enforced by a single data input validation rule”
  • “Data is not always clean or dirty in itself – it depends on the viewpoint or defined standard”
  • “Data quality is in the eye of the beholder”

 

Measuring the Business Impact of Data Governance

Conference session by Tony Fisher, CEO of DataFlux, and Dr. Walid el Abed, CEO of Global Data Excellence.

Quotes from the session:

  • “The goal of data governance is to position the business to improve”
  • “Revenue optimization, cost control, and risk mitigation are the business drivers of data management”
  • “You don't manage data to manage data, you manage data to improve your business”
  • “Business rules are rules that data should comply with in order to have the process execute properly”
  • “For every business rule, define the main impact (cost of failure) and the business value (result of success)”
  • “Power Shift – Before: Having information is power – Now: Sharing information is power”
  • “You must translate technical details into business language, such as cost, revenue, risk”
  • “Combine near-term fast to value with long-term alignment with business strategy”
  • “Data excellence must be a business value added driven program”
  • “Communication is key to data excellence, make it visible and understood by all levels of the organization”

 

The Effect of the Financial Meltdown on Data Management

Conference session by April Reeve, Consultant at EMC Consulting.

Quotes from the session:

  • “The recent financial crisis has greatly increased the interest in both data governance and data transparency”
  • “Data Governance is a symbiotic relationship of Business Governance and Technology Governance”
  • “Risk management is a data problem in the forefront of corporate concern – now viewing data as a corporate asset”
  • “Data transparency increases the criticality of data quality – especially regarding the accuracy of financial reporting”

 

What the Business Wants

Closing Keynote Address by Graeme Simsion, Principal at Simsion & Associates.

Quotes from the keynote:

  • “You can get a lot done if you don't care who gets the credit”
  • “People will work incredibly hard to implement their own ideas”
  • “What if we trust the business to know what's best for the business?”
  • “Let's tell the business what we (as data professionals) do – and then ask the business what they want”

 

Social Karma

My Badge for Enterprise Data World 2010

I presented this session about the art of effectively using social media in business.

An effective social media strategy is essential for organizations as well as individual professionals.  Using social media effectively can definitely help promote you, your expertise, your company, and its products and services. However, too many businesses and professionals have a selfish social media strategy.  You should not use social media to exclusively promote only yourself or your business. 

You need to view social media as Social Karma.

For free related content with no registration required, click on this link: Social Karma

 

Live-Tweeting at Enterprise Data World 2010

Twitter at Enterprise Data World 2010

The term “live-tweeting” describes using Twitter to provide near real-time reporting from an event.  When a conference schedule has multiple simultaneous sessions, Twitter is great for sharing insights from the sessions you are in with other conference attendees at other sessions, as well as with the on-line community not attending the conference.

Enterprise Data World 2010 had a great group of tweeps (i.e., people using Twitter) and I want to thank all of them, and especially the following Super-Tweeps in particular:   

Karen Lopez – @datachick

April Reeve – @Datagrrl

Corinna Martinez – @Futureratti

Eva Smith – @datadeva

Alec Sharp – @alecsharp

Ted Louie – @tedlouie

Rob Drysdale – @projmgr

Loretta Mahon Smith – @silverdata 

 

Additional Resources

Official Website for DAMA International

LinkedIn Group for DAMA International

Twitter Account for DAMA International

Facebook Group for DAMA International

Official Website for Enterprise Data World 2010

LinkedIn Group for Enterprise Data World

Twitter Account for Enterprise Data World

Facebook Group for Enterprise Data World 

Enterprise Data World 2011 will take place in Chicago, Illinois at the Chicago Sheraton and Towers on April 3-7, 2011.

 

Related Posts

Enterprise Data World 2009

TDWI World Conference Chicago 2009

DataFlux IDEAS 2009

Recently Read: March 22, 2010

Recently Read is an OCDQ regular segment.  Each entry provides links to blog posts, articles, books, and other material I found interesting enough to share.  Please note “recently read” is literal – therefore what I share wasn't necessarily recently published.

 

Data Quality

For simplicity, “Data Quality” also includes Data Governance, Master Data Management, and Business Intelligence.

  • The Data Quality Herald Magazine – Dylan Jones, the founder and community manager of Data Quality Pro, recently released the first edition of a unique new magazine focused squarely on the needs of the data quality community.
  • Defining Master Data for Your Organization – Loraine Lawson recaps a recent David Loshin MDM vendor panel discussion, and looks at both the simple answer and the complex, but more useful, answer to the question “what is master data?”
  • What is Data Quality anyway? – Henrik Liliendahl Sørensen asks two excellent questions in this blog post (which also received great comments): “is data quality an independent discipline?” and “is data quality an independent technology?”
  • Business logic – Peter Thomas provides a hilarious adapted comic strip.
  • Police Untelligence – from IQTrainwrecks.com, which is provided by the IAIDQ, read the story about the home of an elderly Brooklyn couple that has been raided by the New York City Police Department 50 times over the last 4 years.
  • Metadata and 3-D Glasses – David Loshin explains the data governance, data stewardship, and metadata/harmonization albatross hanging around the neck of the common question “what is the definition of ‘customer’?”
  • No Enterprise wide Data Model – Ken O’Connor continues his excellent series about common enterprise wide data governance issues with this entry about the impact of not having an enterprise wide data model.
  • Putting data on the web – Rich Murnane shares an excellent recent TED video by Tim Berners-Lee showing some of the benefits of shared data on the web.
  • Building your Data Governance Board – Marty Moseley continues his overview of agile data governance by discussing how you select a data governance board, and how you establish data governance priorities.
  • The Change Paradox – Carol Newcomb examines the “change is good, but change is bad” paradox often encountered in consulting when recommended new technology or new methodology conflicts with your client's corporate culture.  
  • Data Quality Non-Believers – Phil Simon takes on the data quality non-believers making “dataless decisions” by relying on gut instincts to explain such things as customer churn, employee turnover, and intelligent spending of corporate funds. 
  • Data Cleansing to Achieve Information Quality – Jackie Roberts raises some interesting questions regarding the efforts needed to cleanse data though multiple stages of analytics and processes to achieve appropriate information quality.
  • Data Quality Principles within the PMO – Phil Wright provides a list of six excellent principles that must be met in order to help embed a culture of data quality, data assurance, and data governance within each new project.
  • Is computer analysis accurate? – Julian Schwarzenbach considers the accuracy of computer analysis in decision making, especially automated decision making that attempts to mimic human logic, intuition, and insight.

 

Related Posts

Recently Read: March 6, 2010

Recently Read: January 23, 2010

Recently Read: December 21, 2009

Recently Read: December 7, 2009

Recently Read: November 28, 2009

 

Recently Read Resources

Data Quality via My Google Reader

Blogs about Data Quality, Data Governance, Master Data Management, and Business Intelligence

Books about Data Quality, Data Governance, Master Data Management, and Business Intelligence

Data Quality Mad Libs (Part 2)

Data Quality Mad Libs is an ongoing OCDQ series.

For the uninitiated, Mad Libs are sentences with several of their key words or phrases left intentionally blank.

Next to each blank is indicated what type of word should be entered, but you get to choose the actual words.

The completed sentence can be as thought-provoking, comical, or nonsensical as you want to make it.

 

Data Quality Mad Lib

“If you want to

_______________ (verb or phrase)

your

_______________ (noun or phrase)

initiative, then

_______________ (verb)

your

_______________ (noun or phrase)

that

_______________ (phrase)

is highly recommended.”

 

My Version

“If you want to doom your data quality initiative, then advise your technical stakeholders that ignoring the business context is highly recommended.”

 

Share Your Version

Post a comment below and share your completed version of this Data Quality Mad Lib.

 

Related Posts

Data Quality Mad Libs (Part 1)

Recently Read: March 6, 2010

Recently Read is an OCDQ regular segment.  Each entry provides links to blog posts, articles, books, and other material I found interesting enough to share.  Please note “recently read” is literal – therefore what I share wasn't necessarily recently published.

 

Data Quality

For simplicity, “Data Quality” also includes Data Governance, Master Data Management, and Business Intelligence.

  • Let the Data Geeks Play – Rob Paller is hosting a contest on his blog challenging all data geeks to submit an original song (or parody of an existing one) related to MDM, Data Governance, or Data Quality.  Deadline for submissions is March 20.
  • The First Step on your Data Quality Roadmap – Phil Wright describes how to learn lessons from what has happened before, and use this historical analysis as a basis for planning a successful strategy for your data quality initiative.
  • Bad word?: Data Owner – Henrik Liliendahl Sørensen examines how the common data quality terms “data owner” and “data ownership” are used and whether they are truly useful.  Excellent commentary was also received on this blog post.
  • Data as a smoke screen – Charles Blyth discusses how to get to the point where your consumers trust the data that you are providing to them.  This post includes a great graphic and received considerable commentary.
  • MDM Streamlines the Supply Chain – Evan Levy ruminates on the change management challenge for MDM—where change truly is constant—and how the supply chain can become incredibly flexible and streamlined as a result of MDM.
  • MDM as a Vendor Fight to Own Enterprise Data – Loraine Lawson (with help from actor Peter Boyle) looks at another angle of the recent MDM vendor consolidation, based on the recent remark “MDM is the new ERP” made by Jill Dyché. 
  • Data Quality Open Issues and Questions? – Jackie Roberts of DATAForge issues the blogosphere challenge of discussing real-world best practices for MDM, data governance, and data quality.  This blog post received some great comments.
  • Noise and Signal – David Loshin examines the implications of the rising volumes of unstructured data (especially from social media sources) and the related need for data (and metadata) quality to help filter out the signal from the noise.  
  • A gold DQ team! – Daniel Gent, inspired by the recent Winter Olympics and his country's success in ice hockey, discusses the skills and characteristics necessary for assembling a golden data quality team. 
  • Unpredictable Inaccuracy – Henrik Liliendahl Sørensen incites another thought-provoking discussion in the comments section of his blog with this post about the impact on data quality initiatives caused by the challenging reality of time.
  • Does your data quality help customers succeed? – Dylan Jones searches for the holy grail of data quality—providing your customers with great information quality that enables them to achieve their goals as quickly and simply as possible.
  • Charm School: It’s Not Just for IT Anymore – Jill Dyché reminds the business that it’s their business, too—and illustrates the need for a sustained hand-off cycle between IT and the business—and the days of the IT-business mind-meld are over.
  • Data Quality Lip Service – Phil Simon examines why leaders at many organizations merely pay lip service to data quality, and makes some recommendations for getting data quality its due.  Simon Says: “Read this blog post!”
  • What is the name of that block? – Rich Murnane provides a fascinating discussion about looking at things differently by sharing a TED video with Derek Sivers, who explains the different way locations are identified in Japan.
  • Aphorism of the week – Peter Thomas recently (and thankfully) returned to active blogging.  This blog post is a great signature piece representative of his excellent writing style, which proves that long blog posts can be worth reading.
  • How tasty is your data quality cheese? – Julian Schwarzenbach explains data quality using a cheese analogy, where cheese represents a corporate data set, mold represents poor data quality, which causes indigestion—and poor business decisions.
  • Wild stuff: Nines complement date format – Thorsten Radde provides a great example of the unique data quality challenges presented by legacy applications by explaining the date format known as Nine’s complement

 

Social Media

For simplicity, “Social Media” also includes Blogging, Writing, Social Networking, and Online Marketing.

  • Ten Things Social Media Can't Do – B.L. Ochman provides a healthy reminder for properly setting realistic expectations about social media, and provides a great list of ten things you should not expect from social media.
  • A Manifesto for Social Business – Graham Hill discusses how the nature of business is inexorably changing into a new kind of Social Business that is driven by social relationships, and lists fifteen themes (the Manifesto) of this change.
  • Framing Your Social Media Efforts – Chris Brogan explains there are three fundamental areas of practice for social media: (1) Listening, (2) Connecting, and (3) Publishing.
  • Minding the Gap – Tara Hunt examines the gap between the underlying values of business and the underlying human values that drive community.  This blog post also includes an excellent SlideShare presentation that I highly recommend.
  • The Albert Einstein Guide to Social Media – Amber Naslund channels the wisdom of Albert Einstein by using some of his most insightful quotes to frame a practical guide to a better understanding of social media.

 

Book Quotes

An eclectic list of quotes from some recently read (and/or simply my favorite) books.

  • From Linchpin: Are You Indispensable? by Seth Godin – “You don't become indispensable merely because you are different.  But the only way to be indispensable is to be different.  That's because if you're the same, so are plenty of other people.  The only way to get what you're worth is to stand out, to exert emotional labor, to be seen as indispensable, and to produce interactions that organizations and people care deeply about.”

 

Related Posts

Recently Read: January 23, 2010

Recently Read: December 21, 2009

Recently Read: December 7, 2009

Recently Read: November 28, 2009

 

Recently Read Resources

Data Quality via My Google Reader

Social Media via My Google Reader

Books about Data Quality, Data Governance, Master Data Management, and Business Intelligence

Blogs about Data Quality, Data Governance, Master Data Management, and Business Intelligence

Books about Social Media, Blogging, Social Networking, and Online Marketing

Blogs and Websites about Social Media, Social Networking, and Online Marketing

Adventures in Data Profiling

Data profiling is a critical step in a variety of information management projects, including data quality initiatives, MDM implementations, data migration and consolidation, building a data warehouse, and many others.

Understanding your data is essential to using it effectively and improving its quality – and to achieve these goals, there is simply no substitute for data analysis.

 

Webinar

In this vendor-neutral eLearningCurve webinar, I discuss the common functionality provided by data profiling tools, which can help automate some of the work needed to begin your preliminary data analysis.

You can download (no registration required) the webinar (.wmv file) using this link: Adventures in Data Profiling Webinar

 

Presentation

You can download the presentation (no registration required) used in the webinar as an Adobe Acrobat Document (.pdf file) using this link: Adventures in Data Profiling Presentation

 

Complete Blog Series

You can read (no registration required) the complete OCDQ blog series Adventures in Data Profiling by following these links:

The Circle of Quality

Explaining why data quality is so vitally important to an organization's success that it needs to be viewed as a corporate asset is unfortunately not an easy task to accomplish. 

A common mistake made during such attempts is failing to frame data quality issues in a business context, which leads the organization's business stakeholders to understandably mistake data quality for a purely technical issue apparently lacking any tangible impact on their daily business decisions.

An organization's success is measured by the quality of the results it produces.  The results are dependent on the quality of its business decisions.  Those decisions rely on the quality of its information.  That information is based on the quality of its data. 

Therefore, data must be viewed as a corporate asset because high quality data serves as a solid foundation for business success.

As the above diagram illustrates, quality is a fundamental requirement and success criterion all throughout the interconnected Data–>Information–>Decision–>Result business context continuum, which I refer to as The Circle of Quality.

 

The Circle of Quality

Peter Benson of the ECCMA explains that data is intrinsically simple and can be divided into one of two categories:

  1. Master Data – data that identifies and describes things
  2. Transaction Data – data that describes events

In other words, master data is an abstract description of the real-world entities with which the organization conducts business (e.g., customers and vendors).  Transaction data is an abstract description of the real-world interactions that the organization has with those entities (e.g., sales and purchases).

Although a common definition for data quality is fitness for the purpose of use, the common challenge is that all data has multiple uses—and each specific use has its own specific fitness requirements. 

Viewing each specific use as the information that is derived from data, I define information as data in use or data in action.

Although data's quality can be objectively measured separate from its many uses (i.e., data can be fit to serve as at least the basis for each and every purpose), information's quality can only be subjectively measured according to its specific use.

Therefore, information is being customized to meet the subjective needs of a particular business unit and/or a particular tactical or strategic initiative.  In other words, the information is being used as the basis for making a critical business decision.

The quality of the decision is measured by the business result that it produces.  Of course, the reality is that the result is often not immediate and also contingent upon a complex interplay of multiple business decisions.

The result can also produce more data, which could come in the form of new transaction data associated with either existing master data (e.g., sales to existing customers) or new master data (e.g., purchases from new vendors). 

Either way, with the arrival of this new data, yet another spin around The Circle of Quality begins all over again . . .

 

Conclusion

The Circle of Quality illustrates the interconnected business context continuum formed by data, information, decisions, and results.  Additionally, it demonstrates the need for a sustained enterprise-wide program of data governance and data quality, which is necessary for managing data as a corporate asset.

The Circle of Quality also helps illustrate the true challenge of root cause analysis, where poor quality could be occurring in one or more places within the business context continuum. 

And of course, even total quality management is no guarantee of success since it is certainly possible to have high quality data, derive high quality information from it, and then make high quality business decisions based upon it—but still get poor results.

However, it's also easy to imagine the highly questionable results produced when data quality is not considered vital to an organization's success.  Therefore, not managing data as a corporate asset is nothing less than extremely risky business.

 

Related Posts

Beyond a “Single Version of the Truth”

Poor Data Quality is a Virus

DQ-Tip: “Don't pass bad data on to the next person...”

The Only Thing Necessary for Poor Data Quality

Hyperactive Data Quality (Second Edition)

The General Theory of Data Quality

Data Governance and Data Quality

The Data-Information Continuum