The sky’s the limit with the two sides of AI and networking

Generative AI is a hyped topic in virtually all industries. But within the realms of infrastructure management, there are two aspects to consider: AI for networking, and networking for AI

The big change for artificial intelligence (AI) and networking was noted at the HPE Discover 2024 conference by Jensen Huang, founder and CEO of NVIDIA – a company with a pivotal role in the AI tech ecosystem.

During his keynote, Huang observed that the era of generative AI (GenAI) was here and that enterprises had to engage with “the single most consequential technology in history”. He told the audience that what is now happening in the industry is the greatest fundamental computing platform transformation in 60 years, encompassing general purpose computing to accelerated computing, from processing on CPUs plus GPUs.

“Every company is going to be an intelligence manufacturer. Every company is built fundamentally on domain-specific intelligence. For the very first time, we can now digitise that intelligence and turn it into our AI – the corporate AI,” he said.

“AI is a lifecycle that lives forever. What we are looking to do in all of our companies is to turn our corporate intelligence into digital intelligence. Once we do that, we connect our data and our AI flywheel so that we collect more data, harvest more insight, and create better intelligence. This allows us to provide better services or to be more productive, run faster, be more efficient, and to do things at a larger scale.”

While the general potential ramifications of the partnership between parent company Hewlett Packard Enterprise (HPE) and NVIDIA are unknown right now, David Hughes, chief product officer of HPE Aruba Networking (HPE’s security and networking subsidiary), said that there are more pressing issues about the usage of AI in enterprise networks – in particular, around harnessing the benefits that GenAI can offer in the world of CPUs plus GPUs. Hughes believes that the deployment of AI in its industry has two sides – one is AI for networking, and the other is networking for AI.

He said that there are subtle but fundamental differences between these two sides: “Networking for AI is about building out first and foremost the kind of switching infrastructure that’s needed to interconnect these GPU clusters. And then a little bit beyond that, thinking about the impact of collecting telemetry on a network and the changes in the way that people might want to build out their network. So, that’s all networking with AI.

“The other area, AI for networking, is one where we spend time from an engineering and data science point of view. It’s really about [questioning] how we use AI technology, to turn IT admins into super admins so that they can handle their escalating workloads independent of GenAI, which is kind of a load on top of everything else, such as escalating cyber threats and concerns about privacy. The business is asking IT to do new things, deploy new apps all the time, but they’re [asking this of] the same number of people.”

Hughes believes it is important to demonstrate to these hard-pressed IT admins how to take best advantage of automation and AI to take more off their plate so that they can scale. Hughes revealed that his company has a team of a few thousand data scientists working on how to leverage AI more broadly, including classification AI and GenAI, into the company’s products, particularly the Aruba Central cloud-based management system.

After making the distinctions between the categories of AI for networking and networking for AI, Hughes said the main job will be to take this technology to those who will be using it. The challenge will be to articulate what AI and networking means in the job for those running the networks at HPE Aruba customers such as at Espai Barça, said to be is the “largest and most innovative” sports and entertainment space in a European city; the Tottenham Hotspur football stadium; the AT&T Arena of the Dallas Cowboys; and the Mercedes-AMG Petronas Formula One team.

Instead of users having to figure out how to use AI to make lives better, we’re going to do that for them
David Hughes, HPE Aruba Networking

“The key [for users] is how it transforms their jobs,” said Hughes. “For us, that’s about explaining the change rather than just saying, ‘Here’s some more tech this year.’ Our main job is taking this technology to make [operations] more efficient. So, instead of users having to figure out how to use AI to make lives better, we’re going to do that for them.

“There are obviously some domain-specific things that they need to take care of, but in terms of building a network that largely runs itself, we should be doing that. So, that’s really where we are investing, taking a very inspirational high level [of technology] down to the absolute nuts and bolts.”

For HPE Aruba customers such as BMW and General Electrics, which are moving into the realms of AI-based digital twins to support their advanced engineering environments, getting down to the nuts and bolts is an almost literal requirement.

The granularity of the HPE Aruba AI offer extends to recommendations for not only individual customers, but also individual sites and even individual access points, with the latter offering direction on the best firmware to run. Hughes explained why this is important to customers in a typical wireless network deployment.

“If you ask someone, they say the thing that is most successful is based on their personal experience,” said Hughes. “When we have AI, it is looking at all different factors about that particular access point [AP] – the size of the venue, what is the people density, the types of things they’re doing with it, the types of end systems they’ve got. Maybe it’s a place where everyone’s got iPhones or some other type of phone.

“AI realises that you’ve got those kinds of end points in the mix, and then perhaps these particular releases of our AP don’t work so well as those other ones [because of a bug, for example], it will take all of that into account to recommend ... [the hardware] for a site, the framework, the APs, and so on.”

Key to this is that the intelligence and output delivered is based on the data collected and processed through the Aruba Central system. This is currently managing data from roughly four million devices – meaning access points and switches – representing about a billion plus actual telemetry endpoints such as phones and laptops collected into one big data lake, which is used to train the language model for the AI offer, said Hughes.

“It’s way better than any admin can do based on their individual experience just for that firmware example, and it’s multiplied about 100 times with all other kinds of recommendations. For many of these recommendations, we suggest to the admin, ‘Do this’, but there’s a checkbox saying, ‘Yes, I want to do it.’ And if you see something similar to this in future, do it automatically.

“And so that turns into closed loop AI automation. And this has been a major push for us. I believe that we’re really only at the beginning, because there’s a lot more we can be doing. We should be getting to the point where things really are running themselves and the triaging is completely automatic. We’re making really good progress, but the sky’s the limit.”

The AI wheelhouse

But is this generally the opinion of end users who realise this potential in the context of how their jobs are changing and reacting appropriately to how the world is changing around them, or is it a case that AI-solutions are being pushed as the next big thing by the technology industry? Hughes accepts that it is a bit of both right now.

“I think there’s both cases. There are people looking for how to automate and how to reduce the loads on teams, and they’re very willing to adopt new technology and make changes to do them. But there are a lot of customers too that are [happy] networking with traditional network management, where you’re managing individual devices and controlling individual devices.

“Whereas really, you’re just defining security policy once for your whole network, for example. You’re not configuring thousands of access control lists [ACLs] with thousands of lines across all your firewalls. You’re defining at once and then letting automation and AI take care of the rest. But for some people there’s some discomfort, and partly that’s why we have the checkbox. Some people are nervous about letting AI tweak their network. They want that to be mediated.

“So, that’s the default mode. They’re worried AI might do something stupid … especially with GenAI. It’s quite amazing in many senses, but [GenAI’s] not very well-suited to some problems. The one area where it excels is enhanced understanding of language.”

Yet even with this personal view, Hughes stresses his feeling that AI is going to have a broad impact. He said that HPE Aruba is using GenAI in terms of providing natural language interfaces, referencing actual documentation, and providing a summary with links and summaries. In general, he believes that the use of generative AI is making good progress and, by keeping an eye on guide rails and by having it reference actual documentation, that will stop it going off track and into hallucinations.

As with any technology, its success or otherwise is totally down to the use cases that it will support. Hughes believes that with the increased maturity of AI and in particular GenAI and new models of natural language interfaces, there are around 10 or 20 new kinds of insights released to general categories.

“We are looking at all the different things that people would otherwise need to tune to provide those insights. We also look for anomalies and identify what seems unusual, bringing them to the attention of the admin to figure out if it is something they need to care about. That’s another whole area around anomaly detection, because it can’t have people doing that. There’s too much data for them to be sifting through it all.

“Another one is correlation. When something goes wrong, often it has many downstream effects and so IT gets a whole lot of alerts. But what people really want to find out is the root cause because that’s the thing they need to go and address. So, all these things are absolutely in the wheelhouse of different types of AI.”

And these different types of AI are going to be running on differing models and places according to individual needs, despite the general cloudification of the tech industry. On-premise use is very much on agenda, in particular for those in areas such as banking and healthcare where Hughes believes there are good reasons to maintain this mode of operation. For other companies, SMEs as a prime example, the cloud just works. The bottom line is that there won’t be a cookie-cutter solution.

Looking at the forthcoming challenges of successful deployment of AI in networking and the way in which success will be measured, Hughes believes one key criterion will simply be how many devices the company has under management. This is regarded as a key metric because it tells HPE Aruba just how it is doing in terms of bringing its customer into this new world.

The biggest challenge to achieving aims with AI may not be technological but actually, despite many fears to the contrary, a lot more human.

“The biggest challenge for people will be anywhere you need to make a change,” said Hughes. “Making a change means a burst of effort to get over business disruption. So, how do you get people over the hump more easily? There’s also probably many people in the category of, ‘I think it’s time [for AI], but I don’t know when we’re going to do this.’ So, how do we make that easier for them?”

Read more about AI in networking

Read more on Internet infrastructure