"Data, data everywhere, and not a drop to drink." All businesses are fully aware of how much data they're swimming through on a daily basis. And because its buzzy and trendy, most of these businesses are looking to do more with their data, striving to implement cool sounding technologies like machine learning and predictive analytics.
How many, exactly? 41% of executives in a recent 451 Research survey of advanced analytics are looking to begin implementing applications such as Machine Learning or Predictive Modeling in the next 12 months, and an additional 14% plan to do so in the next 24.
And why shouldn't they? These sophisticated programs are highly efficient and represent the future of many different verticals supported by the technology industry.
Yet as enterprises and their leadership see these initiatives on the horizon, a startling number are overlooking a crucial factor that could make or break the success of these investments: the quality of their own data. With some enterprises curating up to 200 disparate data sources, ensuring data quality is no easy task. But getting it right can literally make the difference between a very public crash 'n' burn, or being the standard that everyone tries to emulate.
Here are three reasons why the average enterprise isn't properly prepared for an advanced analytics strategy.
Reason 1: Medieval Methods for Managing Data Quality
According to the survey, 37% of enterprises employ a manual data cleansing process. Given current data volumes, manually cleaning something isn't so 1990s, it's actually more like 1500s. Many of these enterprises are starting to look towards algorithmic automation – but how can they successfully automate advanced processes when their back-end data quality checks remain manual?
44.5% of respondents are in a reactive mode, meaning they only deal with their data quality when it becomes a problem … that they notice (and by the way, their customers noticed way before they did).
The majority of respondents (65%) acknowledge up to 50% of business value can be lost to poor data quality – think that number is going to decrease when the number of initiatives that rely on clean data increases?
Reason 2: Businesses Don't Know The Exact Quality of Their Data
Because of these current Data Quality Management "strategies", IT departments and C-suite executives have a lack of faith in the actual quality of their data.
Over half (57%) of respondents in this survey were "somewhat confident", "unaware", or "less than confident" in the state of their data. Not exactly a resounding endorsement.
This feeling is compounded by the dependency on manual effort to drive remediation in many enterprises' data quality process. Manual entry was the leading cause of poor data quality, also coming in at 57%.
To be fair, you can't blame employees for making mistakes in data entry or processing, but you can blame their management for not providing them with the right tools to handle the volume of data they face every day.
Reason 3: The Stream of Data Today is About to Become a Tsunami
If proper preparations aren't undertaken right now with the relatively manageable amount of data that currently exists, it will be not just be harder, it will be impossible to get a handle on it at the rate that data sources and volumes will continue to expand over the next 3-5 years.
95% of survey respondents acknowledge they expect data to increase (the other 5% presumably in businesses that won't be around in five years).
70% expect data volumes to grow by 70%, while nearly all of the remaining 30% expect it to grow by more than 75%. Chances are, all of them are underestimating what's headed in their direction.
The problems faced by the enterprise today are significant, but can be managed if IT executives deal with the data quality issue now. Tools and technologies are available to ensure viable data quality, which becomes the foundation for growth and value-add, but the choice to act now or quickly get buried is in our collective face, and requires immediate action.
Dan Ortega is VP of Marketing at Blazent.
The increased complexity of new computing architectures coupled with new application development methodologies – especially in the face of time-to-market and security threat pressures – should make secure UX the first strategic decision for CEOs and CFOs on the path to digital transformation ...
IT professionals tend to go above and beyond the scope of their core responsibilities as the changing business landscape demands more of their attention, both inside and outside of the office, according to the Little-Known Facts survey conducted by SolarWinds in honor of IT Professionals Day ...
Digital video consumption is viral and, according to a new study released by IBM and International Broadcasting Convention (IBC), more than half of the 21,000 consumers surveyed are using mobiles every day to watch streaming videos, and that number is expected to grow 45 percent in the next three years ...
No technology that touches more than one IT stakeholder, no matter how good and how transformative, can deliver its potential without attention to leadership, process considerations and dialog. In this blog, I'd like to share effective strategies for AIA adoption ...
Enterprise IT environments are becoming more heterogeneous and complex, with fragmentation permeating cloud infrastructure, tooling and culture, according to a survey recently conducted by IOD Cloud Technologies Research in partnership with Cloudify ...
One area that enables enterprises to reduce complexity and streamline operations is their virtual desktop infrastructure (VDI). Virtualization is a linchpin of digital transformation and effectively optimizing an enterprise's VDI is essential to moving forward with digital technologies. Delivering the best possible VDI performance means taking a fresh look at what "desktop" means today. The endpoint, or desktop, now can be a physical thin client, a software-defined thin client, a traditional laptop, a phone or tablet. To reduce operational waste and achieve better performance across the desktop environment, consider these five actions ...
In incident management, we often overlook the simple things in favor of trying to do too much, too soon. Why not make sure we've done the fundamentals properly? ...
The Input/Output Operations per Second (I/O) capabilities of modern computer systems are truly a modern wonder. Yet no matter how powerful the processors, no matter how many cores, how perfectly formed the bus architecture, or how many flash modules are added, somehow it never seems to be enough ...
By taking advantage of performance monitoring, IT and business decision makers can gain better visibility into their cloud and application performance. Dedicated performance monitoring has become essential for providing visibility into all areas of application performance and keeping the business running optimally ...