Overcoming the Memory Wall with {CXL-Enabled}{SSDs}

SP Yang, M Kim, S Nam, J Park, JY Choi… - 2023 USENIX Annual …, 2023 - usenix.org
This paper investigates the feasibility of using inexpensive flash memory on new
interconnect technologies such as CXL (Compute Express Link) to overcome the memory …

Flash-Cosmos: In-flash bulk bitwise operations using inherent computation capability of nand flash memory

J Park, R Azizi, GF Oliveira… - 2022 55th IEEE/ACM …, 2022 - ieeexplore.ieee.org
Bulk bitwise operations, ie, bitwise operations on large bit vectors, are prevalent in a wide
range of important application domains, including databases, graph processing, genome …

{Hardware/Software}{Co-Programmable} framework for computational {SSDs} to accelerate deep learning service on {Large-Scale} graphs

M Kwon, D Gouk, S Lee, M Jung - 20th USENIX Conference on File and …, 2022 - usenix.org
Graph neural networks (GNNs) process large-scale graphs consisting of a hundred billion
edges. In contrast to traditional deep learning, unique behaviors of the emerging GNNs are …

{λ-IO}: A Unified {IO} Stack for Computational Storage

Z Yang, Y Lu, X Liao, Y Chen, J Li, S He… - 21st USENIX Conference …, 2023 - usenix.org
The emerging computational storage device offers an opportunity for in-storage computing. It
alleviates the overhead of data movement between the host and the device, and thus …

Deepburning-gl: an automated framework for generating graph neural network accelerators

S Liang, C Liu, Y Wang, H Li, X Li - Proceedings of the 39th International …, 2020 - dl.acm.org
Building FPGA-based graph learning accelerators is very time-consuming due to the low-
level RTL programming and the complicated design flow of FPGA development. It also …

{GLIST}: Towards {in-storage} graph learning

C Li, Y Wang, C Liu, S Liang, H Li, X Li - 2021 USENIX Annual Technical …, 2021 - usenix.org
Graph learning is an emerging technique widely used in diverse applications such as
recommender system and medicine design. Real-world graph learning applications typically …

{DeepSketch}: A new machine {Learning-Based} reference search technique for {Post-Deduplication} delta compression

J Park, J Kim, Y Kim, S Lee, O Mutlu - 20th USENIX Conference on File …, 2022 - usenix.org
Data reduction in storage systems is an effective solution to minimize the management cost
of a data center. To maximize data-reduction efficiency, prior works propose post …

Behemoth: a flash-centric training accelerator for extreme-scale {DNNs}

S Kim, Y Jin, G Sohn, J Bae, TJ Ham… - 19th USENIX Conference …, 2021 - usenix.org
The explosive expansion of Deep Neural Networks (DNN) model size expedites the need for
larger memory capacity. This movement is particularly true for models in natural language …

{NVMeVirt}: A Versatile Software-defined Virtual {NVMe} Device

SH Kim, J Shim, E Lee, S Jeong, I Kang… - 21st USENIX Conference …, 2023 - usenix.org
There have been drastic changes in the storage device landscape recently. At the center of
the diverse storage landscape lies the NVMe interface, which allows high-performance and …

Optimstore: In-storage optimization of large scale dnns with on-die processing

J Kim, M Kang, Y Han, YG Kim… - 2023 IEEE International …, 2023 - ieeexplore.ieee.org
Training deep neural network (DNN) models is a resource-intensive, iterative process. For
this reason, nowadays, complex optimizers like Adam are widely adopted as it increases the …