Considerations of network/system for AI services
draft-irtf-nmrg-ai-deploy-01
| Document | Type |
This is an older version of an Internet-Draft whose latest revision state is "Active".
Expired & archived
|
|
|---|---|---|---|
| Authors | Yong-Geun Hong , Joo-Sang Youn , Seung-Woo Hong , Pedro Martinez-Julia , Qin Wu | ||
| Last updated | 2025-09-04 (Latest revision 2025-03-03) | ||
| RFC stream | Internet Research Task Force (IRTF) | ||
| Formats | |||
| Additional resources | Mailing list discussion | ||
| Stream | IRTF state | (None) | |
| Consensus boilerplate | Unknown | ||
| Document shepherd | (None) | ||
| IESG | IESG state | Expired | |
| Telechat date | (None) | ||
| Responsible AD | (None) | ||
| Send notices to | (None) |
This Internet-Draft is no longer active. A copy of the expired Internet-Draft is available in these formats:
Abstract
As the development of AI technology has matured and AI technology has begun to be applied in various fields, AI technology is changing from running only on very high performance servers to commodity servers with small affordable hardware, including microcontrollers, low performance CPUs, and AI chipsets. This document considers how to configure the network and system in terms of AI inference service to provide AI service in a distributed manner. It also describes the points to consider in the environment where a client connects to a cloud server and an edge device and requests an AI service. Some use cases of deploying network-based AI services, such as self-driving vehicles and network digital twins, are described.?
Authors
Yong-Geun Hong
Joo-Sang Youn
Seung-Woo Hong
Pedro Martinez-Julia
Qin Wu
(Note: The e-mail addresses provided for the authors of this Internet-Draft may no longer be valid.)