How AI Is Poised to Upend Cloud Networking – Data Center Knowledge

Much has been said about how AI will accelerate the growth of cloud platforms and enable a new generation of AI-powered tools for managing cloud environments.

But here's another facet of the cloud that AI is likely to upend: networking. As more and more AI workloads enter the cloud, the ability to deliver better cloud networking solutions will become a key priority.

Related: Why AI Workloads Probably Won't Transform the Data Center Industry

Here's why, and what the future of cloud networking may look like in the age of AI.

The reason why AI will place new demands on cloud networks is simple enough: To work well at scale, AI workloads will require unprecedented levels of performance from cloud networks.

Related: Explosion of Data in the Cloud Era Leading to Observability Complexity

That's because the data that AI workloads need to access will reside in many cases on remote servers located either within the same cloud platform where the workloads live or in a different cloud. (In some cases, the data could also live on-prem while the workloads reside in the cloud, or vice versa.)

Cloud networks will provide the essential link that connects AI workloads to data. The volumes of data will be vast in many cases (even training a simple AI model could require many terabytes' worth of information), and models will need to access the data at low latency rates. Thus, networks will need to be able to support very high bandwidth with very high levels of performance.

To be sure, AI is not the only type of cloud workload that requires great network performance. The ability to deliver low-latency, high-bandwidth networking has long been important for use cases like cloud desktops and video streaming.

Cloud vendors have also long offered solutions to help meet these network performance needs. All of the major clouds provide "direct connect" networking services that can dramatically boost network speed and reliability, especially when moving data between clouds in a multicloud architecture, or between a private data center and the public cloud as part of a hybrid cloud model.

But for AI workloads with truly exceptional network performance needs, direct connect services may not suffice. Workloads may also require optimizations at the hardware level in the form of solutions such as data processing units (DPUs), which can help process network traffic hyper-efficiently. Indeed, vendors like Nvidia, which has unveiled an Ethernet platform tailored for generative AI, are already investing in this area and it says a lot that a company mostly known for selling video cards is also recognizing that unlocking the full potential of AI requires networking hardware innovations, too.

For now, it remains to be seen exactly how cloud vendors, hardware vendors, and AI developers will respond to the special challenges that AI brings to the realm of cloud networking. But in general, it's likely that we'll see changes such as the following:

There's no way around it: If you want to take full advantage of the cloud to help host AI workloads, you need to optimize your cloud networking strategy a move that requires taking advantage of advanced networking services and hardware, while also adjusting cloud cost optimization and network performance management strategies.

For now, the solutions available to help with these goals are still evolving, but this is a space to follow closely for any business seeking to deploy AI workloads in the cloud.

About the author

Read this article:
How AI Is Poised to Upend Cloud Networking - Data Center Knowledge

Related Posts

Comments are closed.