Ciena uses machine learning to heal the scars, horror of network management
“The scars” and “that horrible world” are some of the terms for network management, according to one who’s been in the trenches.
Kailem Anderson was with Cisco Systems for 12 years prior to joining fiber-optics giant Ciena last year. As vice president of portfolio and engineering for the Blue Planet, a software division of Ciena, he is trying to help avoid such pain for those who must keep networks running.
“I managed customer networks, and I spent a lot of time hiring analysts to watch the network, to watch alarms, and to build big strings of rules,” for networking monitoring, says Anderson. His breezy Aussie accent gives a certain lightness to what sounds like a rather miserable affair.
At $26 million in revenue in 2018, Blue Planet was a tiny fraction of Ciena’s roughly $200 million in software revenue in 2018 and $3 billion in total revenue. But it increased by a healthy 66%, and it can bring higher profit margin than Ciena’s optical networking equipment sale. It also offers the company a recurring revenue stream that is highly appreciated by Wall Street. Those economic aspects, plus the fact that it can be strategic in designing customers’ networks, make it an important part of where Ciena is headed as a company.
Also: Is Google’s Snorkel DryBell the future of enterprise data management?
Figuring out what’s gone wrong in a network involves detective work at several levels of what’s known as the “stack” of protocols, the Open Systems Interconnect, or “OSI.” Some information comes from the bottom of the stack, if you will, the “layer one,” which consists of the physical medium of transmission. That could be, for example, coaxial cabling or fiber-optic links.
At the next layer above that, layer two, raw bits are packaged into bundles, such as Ethernet frames, and there’s all kinds of information to be gleaned about the state of those frames of data as they move through the fibers and cables of the network. The next layer up is layer three, where data is packaged as Internet-addressable packets, again, with lots of their owing information to be gleaned, such as routing and switching information about where the packets are going.
From there, one can go on up to higher levels, layers four through seven, the domain of applications, and get information about who an individual application is placing its data into those internet packets and whether it is having any trouble doing so.
Take the example where there is an transponder failure on one of two optical links. That leads to a route change in the multi-protocol label system, or MPLS. The network equipment reports congestion along the IP route as a link shoulders the burden of more traffic, and an end user experiences heavy delays using the network. All these are part of the same problem, Anderson explains, but getting from the user experience to the transponder failure can be a mystery.
Traditionally, a systems administrator sees the various items in a disparate fashion, with signals at each of the OSI layers coming from different telemetry systems, such as SNMP monitors, the systems log, a third thing that tracks “flows,” and then information coming from an individual piece of equipment, such as information about a recent configuration change — none of which are coordinated.
What looks like bad user performance from one angle looks like an MPLS routing issue or an IP bandwidth issue at another level, leading to a serious piece of detective work to find the culprit, the transponder failure.
Also: Google Brain, Microsoft plumb the mysteries of networks with AI
A ticket gets created, and it ping-pongs between teams, with no one team having visibility into the other side, says Anderson. “Eventually they solve it, they have engineers inspect the matter, but it’s very inefficient.”
Sys admins must try and construct systems of rules as to what every possible combination of factors could mean. “They spend 1,000s of hours building these rules,” says Anderson. “It’s a zero sum game to spend that time to identify all the different scenarios.”
Instead, Blue Planet tools can train the network software using a combination of labeled examples, known as supervised learning and reinforcement learning, where the computer explores states of affairs and possible next steps.
With that combination, the software can be trained to identify patterns “up and down the stack” that are difficult to piece together with a rules-based system.
“We want to have the system learn to identify those scenarios, to basically help us get to the root cause much more quickly, and to use that information to close the loop,” he says, and then have a supervisor come into the picture only once that outline has been determined.
Also: Intel-backed startup Nyansa chases the total problem in the AI of network monitoring
The tools necessary to do this are mostly starting from off-the-shelf machine learning models, says Anderson. “Most of this, yes, we can get from the cloud guys,” he says, referring to the various enterprise-grade machine learning offerings in cloud computing facilities. “We use them all,” though the tools can also be run solely on-prem. “It’s six and one half dozen of the other at the moment, but I think analytics is ultimately a good thing to move into the cloud.”
Open-source tools such as SparkML play a big role in organizing all the telemetry data.
The technology of machine learning, says Anderson, has matured substantially in recent years to make the investment in labeling network events pay off.
“Five years ago I was playing with this and with the amount of effort that needed to go into labeling, the risk versus value I was getting was questionable,” he says. “With the hardening of the algorithm, and the maturity of AI, that effort-to-reward ratio has compressed significantly. You only have to do a reasonable amount of tagging now and the outputs are significant.”
Anderson maintains there is another dimension in the shift to machine learning, which is that a more comprehensive sense of the network emerges that may lead to different ways or structuring and maintaining networks.
Traditionally, many sys admins will simply turn off sources of information, says Anderson, which is understandable, because of the information overload, but it means that network administrators are throwing away important clues.
“That’s the complexity in operating with a million different data sources,” he observes. “The traditional way to mange an operations team is to filter the information, almost turn off the information that is too much.
“At Cisco, if I was running a service provider network, I would get in the vicinity of a million events a day, and I might have an operations team of 40 to 50 people who have to handle all that.”
As a consequence, admins end up only looking for “what they deem fair scenarios,” and “are turning off performance-based scenarios,” information about the relative quality of the network.
But, says Anderson, “you don’t want to turn off the information, you want to funnel it, and use it to identify what conditions are driving consistent scenarios,
“Eventually, solutions could be different if they’re trained,” he offers. Data may lead to structuring things differently. “Usually, you have a planned network condition, but then an actual network condition; through learning, you might find the actual is more optimal than planned, and then execute a policy” based on that new insight.
There are new frontiers to achieve, such as delivering analysis of the data in a “graph database” format, says Anderson. “We are in the operations and network world, and so you want to visualize all this in a network graph concept.” Some customers “want to see it just programmatically propagate to northbound systems that are going to leverage that information, to be able to visualize with a graph database and have APIs to send that northbound information to the BSS layer.”
The one catch at the moment in all this is that systems administrators are not yet ready to close the loop, so to speak, and let machine learning completely take over and automate both detection and resolution of network issues.
“This isn’t a tech limit, it’s a cultural aspect,” he says. Machine learning systems are probabilistic, not deterministic. Hence, while they can detect many failure issues, there is a reluctance to automate what could be a false positive scenario. “You only need to screw up .0001% of the time and that’s a big issue.”
“I still think we are a little bit away in terms of closing the loop, I think it’s trust in the technology. It will happen incrementally, where you can close the loop on something non-catastrophic, that doesn’t create a failure scenario, where there is low risk, and then other areas over time
Two Ways To Use Windows 11 On Your Mac Computer
Evaluating the need to incorporate Windows 11 on a Mac involves multiple considerations, particularly within professional, academic, and entertainment contexts. A primary factor to consider is the significance of Windows 11 for your work, studies, or leisure activities. Many jobs and academic programs rely on specific Windows-exclusive applications. At the same time, certain popular games like Microsoft Flight Simulator are only available on Windows. Having Windows 11 on your Mac could be essential for fulfilling your professional, educational, or gaming needs.
It’s also essential to ensure your Mac is compatible with either Windows 365 Cloud or Parallels Desktop. Check that your device meets the necessary hardware and connectivity requirements, including a reliable internet connection for Windows 365 Cloud. Another aspect to examine is how Windows 11 may affect your Mac’s performance. Review your Mac’s current specifications, including RAM and storage, to ensure it can handle both operating systems efficiently without compromising speed or stability.
Finally, consider the financial implications of running Windows 11 on your Mac. Windows 365 Cloud and Parallels Desktop come with costs, as the first option charges a monthly subscription, and the latter requires a one-time purchase of at least $99.99 or an annual subscription. Assess your budget and the advantages of using Windows 11 on your Mac to make a sound decision.
The Fastest Car In BMWs Line-Up Might Surprise You
When it comes to the high-performance lineups of brands like BMW, it's easy to assume that the top-end is going to be the fastest. Not always.
5 Of The Most Underrated Ducati Motorcycles Ever Made
The Ducati 916, produced from 1994 to 1998, sports a masterful body style, and the 999, introduced to the market in 2003, departs sharply from the previous footprint. The 999 is more aerodynamic than its predecessor, but the design was created from scratch and didn’t resemble the much-loved 916. Aesthetics alone can speak to much of the rejection this new model experienced.
In terms of performance, the 999 models opted for a high-speed V-Twin engine, reaching a top speed of 168 miles per hour and weighing just 438 pounds. The 999S variant utilized a new engine build, dubbed the Testasetretta, with a displacement measuring 998cc and producing 136 horsepower alongside a six-speed transmission. The torque rating for the 999 (in a 2005 model, specifically) was 106 Nm and boosted to 111.8 Nm with the racing kit.
These bikes feature a narrower engine than previous models, meaning the bike is slim and minimally profiled by design. Tom Cruise even owns one (a 999R), which has given the model some notoriety. The 999 performance redesign improved on an iconic Ducati, but the 916 continued overshadowing this new addition to the lineup.
Today's Wordle Answer #648 – March 29, 2023 Solution And Hints
Are you stuck on the Wordle for March 29, 2023? Worry not, and don't stress, because we have clues and...
2024 Kia EV9 Three-Row Electric SUV Reveals Range, Power And Tech
The EV9 will be available with two different battery sizes and three different configurations of electric motors. The Standard RWD...
Two Ways To Use Windows 11 On Your Mac Computer
Evaluating the need to incorporate Windows 11 on a Mac involves multiple considerations, particularly within professional, academic, and entertainment contexts....
Ransomware crooks are exploiting IBM file exchange bug with a 9.8 severity
Threat actors are exploiting a critical vulnerability in an IBM file-exchange application in hacks that install ransomware on servers, security...
EV Maker Lucid Plans Widespread Layoffs In Effort To Cut Costs
Beyond the letter sent to staff, Lucid shared some details about the upcoming layoffs in a regulatory filing, the most...
Social12 months ago
Web.com website builder review
Social3 years ago
CrashPlan for Small Business Review
Gadgets4 years ago
A fictional Facebook Portal videochat with Mark Zuckerberg – TechCrunch
Cars4 years ago
What’s the best cloud storage for you?
Social4 years ago
iPhone XS priciest yet in South Korea
Mobile4 years ago
Memory raises $5M to bring AI to time tracking – TechCrunch
Security4 years ago
Google latest cloud to be Australian government certified
Social4 years ago
Apple’s new iPad Pro aims to keep enterprise momentum