Information and Communications Technology and Policy

Information and Communications Technology and Policy

Information and Communications Technology and Policy ›› 2025, Vol. 51 ›› Issue (10): 2-6.doi: 10.12267/j.issn.2096-5931.2025.10.001

Previous Articles     Next Articles

Research on the development of compute-storage collaboration driven by large model inference

ZHOU Lan, CHEN Lei   

  1. Informatization and Industrialization Integration Research Institute, China Academy of Information and Communications Technology, Beijing 100191, China
  • Received:2025-09-10 Online:2025-10-25 Published:2025-11-06

Abstract:

With the continuous enhancement of large model capabilities and the deepening of inference applications, the scale of data processing has expanded drastically, and data processing requirements have become increasingly diversified, this has imposed higher demands on the collaborative between storage and computing power. In response to the new demands on storage systems by larger data volumes, larger model sizes, and longer context windows in current large model inference scenarios, this study first conducts an in-depth analysis of the implementation mechanisms, key technologies, and practical applications of both “computing-in-place-of-storage” and “storage-in-place-of-computing”, Subsequently, by integrating the current technological and industrial foundation as well as application scenario requirements, this paper proposes that based on access latency and bandwidth demands,a hierarchical and systematic collaborative storage model for the future development of computing-storage synergy is important. This paper aims to explore the specific implementation mechanisms and evolutionary pathways of compute-storage collaboration, providing valuable references for promoting the improvement of intelligent computing cluster utilization efficiency and better supporting the development of large model inference.

Key words: large model inference, AI storage, KV Cache, computing-storage collaboration

CLC Number: