Abstract
Network bandwidth demand in datacenters is doubling every 12 to 15 months. In response to this demand, high-bandwidth network interface cards, each capable of transferring 100s of Gigabits of data per second, are making inroads into the servers of next-generation datacenters. Such unprecedented data delivery rates on server endpoints raise new challenges, as inbound network traffic placement decisions within the memory hierarchy have a direct impact on end-to-end performance. Modern server-class Intel processors leverage DDIO technology to steer all inbound network data into the last-level cache (LLC), regardless of the network traffic's nature. This static data placement policy is suboptimal, both from a performance and an energy efficiency standpoint. In this work, we design IDIO, a framework that - unlike DDIO - dynamically decides where to place inbound network traffic within a server's multi-level memory hierarchy. IDIO dynamically monitors system behavior and distinguishes between different traffic classes to determine and periodically re-evaluate the best placement location for each flow: LLC, mid-level (L2) cache or DRAM. Our results show that IDIO increases a server's maximum sustainable load by up to ∼ 33.3% across various network functions.
| Original language | English |
|---|---|
| Article number | 9294125 |
| Pages (from-to) | 30-33 |
| Number of pages | 4 |
| Journal | IEEE Computer Architecture Letters |
| Volume | 20 |
| Issue number | 1 |
| DOIs | |
| State | Published - 1 Jan 2021 |
Bibliographical note
Publisher Copyright:© 2002-2011 IEEE.
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 7 Affordable and Clean Energy
Keywords
- Cache
- data direct I/O
- datacenters
- network
Fingerprint
Dive into the research topics of 'IDIO: Orchestrating inbound network data on server processors'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver