Rearchitecting the {TCP} Stack for {I/O-Offloaded} Content Delivery

T Kim, DM Ng, J Gong, Y Kwon, M Yu… - 20th USENIX Symposium …, 2023 - usenix.org
The recent advancement of high-bandwidth I/O devices enables scalable delivery of online
content. Unfortunately, the traditional programming model for content servers has a tight …

{FVM}:{FPGA-assisted} Virtual Device Emulation for Fast, Scalable, and Flexible Storage Virtualization

D Kwon, J Boo, D Kim, J Kim - 14th USENIX Symposium on Operating …, 2020 - usenix.org
Emerging big-data workloads with massive I/O processing require fast, scalable, and flexible
storage virtualization support. Hardware-assisted virtualization can achieve reasonable …

Data motion acceleration: Chaining cross-domain multi accelerators

ST Wang, H Xu, A Mamandipoor… - … Symposium on High …, 2024 - ieeexplore.ieee.org
There has been an arms race for devising accelerators for deep learning in recent years.
However, real-world applications are not only neural networks but often span across …

BM-Store: A Transparent and High-performance Local Storage Architecture for Bare-metal Clouds Enabling Large-scale Deployment

Y Chen, J Xu, C Wei, Y Wang, X Yuan… - … Symposium on High …, 2023 - ieeexplore.ieee.org
Bare-metal instances are crucial for high-value, mission-critical applications on the cloud.
Tenants exclusively use these dedicated hardware resources. Local virtualized disks are …

TrainBox: an extreme-scale neural network training server architecture by systematically balancing operations

P Park, H Jeong, J Kim - 2020 53rd Annual IEEE/ACM …, 2020 - ieeexplore.ieee.org
Neural network is a major driving force of another golden age of computing; the computer
architects have proposed specialized accelerators (eg, TPU), high-speed interconnects (eg …

FIDR: A scalable storage system for fine-grain inline data reduction with efficient memory handling

M Ajdari, W Lee, P Park, J Kim, J Kim - … of the 52nd Annual IEEE/ACM …, 2019 - dl.acm.org
Storage systems play a critical role in modern servers which run highly data-intensive
applications. To satisfy the high performance and capacity demands of such applications …

PREBA: A Hardware/Software Co-Design for Multi-Instance GPU based AI Inference Servers

G Yeo, J Kim, Y Choi, M Rhu - arXiv preprint arXiv:2411.19114, 2024 - arxiv.org
NVIDIA's Multi-Instance GPU (MIG) is a feature that enables system designers to reconfigure
one large GPU into multiple smaller GPU slices. This work characterizes this emerging GPU …

Smartfvm: A fast, flexible, and scalable hardware-based virtualization for commodity storage devices

D Kwon, W Lee, D Kim, J Boo, J Kim - ACM Transactions on Storage …, 2022 - dl.acm.org
A computational storage device incorporating a computation unit inside or near its storage
unit is a highly promising technology to maximize a storage server's performance. However …

High accuracy positioning for C-V2X

Q Liu, M Song, X Xv, J Qiu - IOP Conference Series: Earth and …, 2021 - iopscience.iop.org
With the rapid development and popularization of 5G and C-V2X, services based on C-V2X
are rapidly expanding. Specially, the positioning accuracy is the most basic requirement in …

Peripheral device controlling device, operation method thereof, and operation method of peripheral device controlling device driver

J Kim, J Ahn, DU Kwon - US Patent 10,664,418, 2020 - Google Patents
A peripheral device controlling device according to an embodiment of the inventive concept
includes a command queue for storing at least one Device to Device (D2D) command for …