Data is one of a company's most valuable assets — but Talend's recent Data Health Survey found only 40% of executives always trust the data they work with.
Today, we have codes and inspections for physical infrastructure, satisfaction surveys for employees, and up-time monitors and stability tests for websites. But are we doing everything we can to understand the degree to which our data is exposed to risk?
There's more to security than protecting yourself from hackers. On one end of the spectrum, you have those big exposures to governmental regulations and security breaches that can shake an entire organization. But even small things — like a little bit of bad data entering the system — can cause a trickle down effect that impacts every department.
We could all be doing a better job of assessing (and mitigating) risk to our data. The key is to start small: just make sure that you have the right data in the right place.
Then you want to make sure that the right people have access to the data and the wrong people don't have access to the data.
Once you have that covered, and you've defined processes for keeping your data clean and standardized, then you can start focusing on making that a daily practice. All it takes is the right combination of people, processes and technology.
What Do We Mean by "Risk?"
When most people think about the risks associated with data, they immediately recall the headline-grabbing data breaches that seem to flood our news feeds with alarming regularity.
But it doesn't take an epic leak affecting millions of users to have serious consequences for most companies. Even a handful of exposed records could have serious legal, financial and reputational repercussions. Fines for GDPR violations alone can run in the millions of dollars, to say nothing of the incalculable cost of losing consumer trust in an increasingly connected and competitive marketplace.
How do these breaches happen?
It can be something as simple as the right data in the wrong place. So much of our conversation about security centers around personally identifiable information (PII). If PII data isn't identified or isn't in the right field — for example, payment information erroneously mapped to an unprotected field and viewed by unauthorized individuals — you could be at risk of exposing some very sensitive information.
But external risks aren't the only dangers we should be worried about. A few years ago, IBM famously calculated that bad data costs US businesses over$3 trillion per year. This is death by a thousand cuts, parceled out in seconds, minutes and hours lost to manual data correction, re-running suspect reports and pursuing strategies and programs that were originally scoped based on data that was later revealed to be faulty.
Of course, the volumes of data we must deal with has grown by over 400% since IBM released that study — and it's only growing.
So how much could we be losing today?
And how much do we stand to lose over the coming years?
Taking all these dangers into account, one thing is clear: no company can afford to expose its data to risk.
Go to: Is Your Data Safe? How to Assess Your Data Risk - Part 2
The Latest
On average, only 48% of digital initiatives enterprise-wide meet or exceed their business outcome targets according to Gartner's annual global survey of CIOs and technology executives ...
Artificial intelligence (AI) is rapidly reshaping industries around the world. From optimizing business processes to unlocking new levels of innovation, AI is a critical driver of success for modern enterprises. As a result, business leaders — from DevOps engineers to CTOs — are under pressure to incorporate AI into their workflows to stay competitive. But the question isn't whether AI should be adopted — it's how ...
The mobile app industry continues to grow in size, complexity, and competition. Also not slowing down? Consumer expectations are rising exponentially along with the use of mobile apps. To meet these expectations, mobile teams need to take a comprehensive, holistic approach to their app experience ...
Users have become digital hoarders, saving everything they handle, including outdated reports, duplicate files and irrelevant documents that make it difficult to find critical information, slowing down systems and productivity. In digital terms, they have simply shoved the mess off their desks and into the virtual storage bins ...
Today we could be witnessing the dawn of a new age in software development, transformed by Artificial Intelligence (AI). But is AI a gateway or a precipice? Is AI in software development transformative, just the latest helpful tool, or a bunch of hype? To help with this assessment, DEVOPSdigest invited experts across the industry to comment on how AI can support the SDLC. In this epic multi-part series to be posted over the next several weeks, DEVOPSdigest will explore the advantages and disadvantages; the current state of maturity and adoption; and how AI will impact the processes, the developers, and the future of software development ...
Half of all employees are using Shadow AI (i.e. non-company issued AI tools), according to a new report by Software AG ...
On their digital transformation journey, companies are migrating more workloads to the cloud, which can incur higher costs during the process due to the higher volume of cloud resources needed ... Here are four critical components of a cloud governance framework that can help keep cloud costs under control ...
Operational resilience is an organization's ability to predict, respond to, and prevent unplanned work to drive reliable customer experiences and protect revenue. This doesn't just apply to downtime; it also covers service degradation due to latency or other factors. But make no mistake — when things go sideways, the bottom line and the customer are impacted ...
Organizations continue to struggle to generate business value with AI. Despite increased investments in AI, only 34% of AI professionals feel fully equipped with the tools necessary to meet their organization's AI goals, according to The Unmet AI Needs Surveywas conducted by DataRobot ...
High-business-impact outages are costly, and a fast MTTx (mean-time-to-detect (MTTD) and mean-time-to-resolve (MTTR)) is crucial, with 62% of businesses reporting a loss of at least $1 million per hour of downtime ...