If you work in an IT organization, you've likely heard the term "observability" lately. If you're a DevOps pro, you probably know exactly what vendors are talking about when they use the term. If you're a NetOps pro, you might be scratching your head.
DevOps Knows Observability
The DevOps community is very familiar with the observability concept. It refers to the ability to understand the internal state of a system by measuring its external outputs. In DevOps, the system is the application, and the outputs are metrics, logs, and traces. DevOps pros know how to navigate messaging from application performance management and cloud monitoring vendors to find solutions that can deliver the observability they need.
More recently, network monitoring vendors have started talking about network observability. Here is where things get fuzzy. In my opinion, DevOps observability and network observability are not interchangeable. Why would they be?
DevOps teams want to understand the state of applications and the infrastructure on which they reside. NetOps teams need to understand a much larger universe of networks, from the cloud to the user edge.
Both DevOps observability and network observability refer to the need to understand the internal state of a system, but that need to understand is only a problem statement. The solution to that problem is where the differences occur.
Does Anyone Have Network Observability?
First, most NetOps teams care about application performance. They want to collect data from the application environment if they can, such as hypervisors and containers. But they don't stop there. They need to monitor data center networks, wide-area networks (WANs), and campus and branch networks. More recently, they've had to worry about home office networks.
Each network they monitor has become more complex. The data center network has been virtualized and partially extended into the public cloud. The WAN has hybridized, with a mix of managed WAN connectivity, public internet, and 4G/5G. Office networks are a mix of ethernet and Wi-Fi, connected via home internet.
A network observability system must monitor and analyze an extremely diverse and ever-growing data set to understand end-to-end network state. A NetOps team might use five, ten, or even fifty tools to monitor a network by collecting packets, flows, device logs, device metrics, test data, DNS logs, routing table changes, configuration changes, synthetic traffic, and more.
It's a lot to keep track of, and it's hard to find a single tool that can handle it all. In fact, my new research on the concept of network observability found that 83% of IT organizations are interested in streaming data from their network observability tool(s) to a central data lake. Why? Nearly half of them believe a data lake will help them correlate network data across their tools.
Earlier, I wrote that network observability is a bit "fuzzy." I'd argue that it's fuzzy because the problem of network observability is much bigger and more complex than DevOps observability. It may prove impossible for any single tool to solve this problem. That's perfectly okay. But IT organizations must keep this in mind and take a comprehensive approach to network operations tools as they steer toward the promise of network observability.
To learn more about network observability, check out EMA's November 9 webinar, which will highlight market research findings on the topic.
The Latest
Artificial intelligence (AI) is rapidly reshaping industries around the world. From optimizing business processes to unlocking new levels of innovation, AI is a critical driver of success for modern enterprises. As a result, business leaders — from DevOps engineers to CTOs — are under pressure to incorporate AI into their workflows to stay competitive. But the question isn't whether AI should be adopted — it's how ...
The mobile app industry continues to grow in size, complexity, and competition. Also not slowing down? Consumer expectations are rising exponentially along with the use of mobile apps. To meet these expectations, mobile teams need to take a comprehensive, holistic approach to their app experience ...
Users have become digital hoarders, saving everything they handle, including outdated reports, duplicate files and irrelevant documents that make it difficult to find critical information, slowing down systems and productivity. In digital terms, they have simply shoved the mess off their desks and into the virtual storage bins ...
Today we could be witnessing the dawn of a new age in software development, transformed by Artificial Intelligence (AI). But is AI a gateway or a precipice? Is AI in software development transformative, just the latest helpful tool, or a bunch of hype? To help with this assessment, DEVOPSdigest invited experts across the industry to comment on how AI can support the SDLC. In this epic multi-part series to be posted over the next several weeks, DEVOPSdigest will explore the advantages and disadvantages; the current state of maturity and adoption; and how AI will impact the processes, the developers, and the future of software development ...
Half of all employees are using Shadow AI (i.e. non-company issued AI tools), according to a new report by Software AG ...
On their digital transformation journey, companies are migrating more workloads to the cloud, which can incur higher costs during the process due to the higher volume of cloud resources needed ... Here are four critical components of a cloud governance framework that can help keep cloud costs under control ...
Operational resilience is an organization's ability to predict, respond to, and prevent unplanned work to drive reliable customer experiences and protect revenue. This doesn't just apply to downtime; it also covers service degradation due to latency or other factors. But make no mistake — when things go sideways, the bottom line and the customer are impacted ...
Organizations continue to struggle to generate business value with AI. Despite increased investments in AI, only 34% of AI professionals feel fully equipped with the tools necessary to meet their organization's AI goals, according to The Unmet AI Needs Surveywas conducted by DataRobot ...
High-business-impact outages are costly, and a fast MTTx (mean-time-to-detect (MTTD) and mean-time-to-resolve (MTTR)) is crucial, with 62% of businesses reporting a loss of at least $1 million per hour of downtime ...
Organizations recognize the benefits of generative AI (GenAI) yet need help to implement the infrastructure necessary to deploy it, according to The Future of AI in IT Operations: Benefits and Challenges, a new report commissioned by ScienceLogic ...