Skip to main content

Considerations of network/system for AI services
draft-irtf-nmrg-ai-deploy-01

The information below is for an old version of the document.
Document Type
This is an older version of an Internet-Draft whose latest revision state is "Active".
Expired & archived
Authors Yong-Geun Hong , Joo-Sang Youn , Seung-Woo Hong , Pedro Martinez-Julia , Qin Wu
Last updated 2025-09-04 (Latest revision 2025-03-03)
RFC stream Internet Research Task Force (IRTF)
Formats
Additional resources Mailing list discussion
Stream IRTF state (None)
Consensus boilerplate Unknown
Document shepherd (None)
IESG IESG state Expired
Telechat date (None)
Responsible AD (None)
Send notices to (None)

This Internet-Draft is no longer active. A copy of the expired Internet-Draft is available in these formats:

Abstract

As the development of AI technology has matured and AI technology has begun to be applied in various fields, AI technology is changing from running only on very high performance servers to commodity servers with small affordable hardware, including microcontrollers, low performance CPUs, and AI chipsets. This document considers how to configure the network and system in terms of AI inference service to provide AI service in a distributed manner. It also describes the points to consider in the environment where a client connects to a cloud server and an edge device and requests an AI service. Some use cases of deploying network-based AI services, such as self-driving vehicles and network digital twins, are described.?

Authors

Yong-Geun Hong
Joo-Sang Youn
Seung-Woo Hong
Pedro Martinez-Julia
Qin Wu

(Note: The e-mail addresses provided for the authors of this Internet-Draft may no longer be valid.)