학술논문

A Scalable RDMA Network Interface Card with Efficient Cache Management
Document Type
Conference
Source
2023 IEEE International Symposium on Circuits and Systems (ISCAS) Circuits and Systems (ISCAS), 2023 IEEE International Symposium on. :1-5 May, 2023
Subject
Components, Circuits, Devices and Systems
Power, Energy and Industry Applications
Signal Processing and Analysis
Circuits and systems
Scalability
Bandwidth
Behavioral sciences
Network interfaces
Low latency communication
RDMA
Network scalability
RNIC Architecture
Quality of service
Language
ISSN
2158-1525
Abstract
Remote Direct Memory Access (RDMA) has been applied in large-scale clusters due to its high bandwidth and low latency features in recent decades. However, the scalability issue is a known intractable problem that the current RDMA Network Interface Card (RNIC) cannot overcome as the number of connections increases to thousands. The key to the scalability bottleneck is managing the connection information cached on the RNIC. To solve the scalability problem, firstly, we test and analyze the behavior of the commercial RNIC when the network performance declines in the case of large-scale connections. Further, we give the performance model of RNIC and point out that the cache design is the key to scalability issues. Then, we propose a Scalable RDMA NIC (ScalaRNIC) architecture with a non-blocking and priority-programmable cache design. Besides, our cache design supports parametric configuration by the extended API. ScalaRNIC can maintain $\mathrm{a}\approx 100\%$ cache hit ratio for higher priority connections and keep message rates nearly equal to peak performance. In contrast, commercial RNIC's performance drops by 47% when there are a few thousand connections.