Bluedbm: An appliance for big data analytics
Complex data queries, because of their need for random accesses, have proven to be slow
unless all the data can be accommodated in DRAM. There are many domains, such as …
unless all the data can be accommodated in DRAM. There are many domains, such as …
StRoM: smart remote memory
Big data applications often incur large costs in I/O, data transfer and copying overhead,
especially when operating in cloud environments. Since most such computations are …
especially when operating in cloud environments. Since most such computations are …
Ibex: An intelligent storage engine with support for advanced sql offloading
Modern data appliances face severe bandwidth bottlenecks when moving vast amounts of
data from storage to the query processing nodes. A possible solution to mitigate these …
data from storage to the query processing nodes. A possible solution to mitigate these …
CoNDA: Efficient cache coherence support for near-data accelerators
Specialized on-chip accelerators are widely used to improve the energy efficiency of
computing systems. Recent advances in memory technology have enabled near-data …
computing systems. Recent advances in memory technology have enabled near-data …
Caribou: Intelligent distributed storage
The ever increasing amount of data being handled in data centers causes an intrinsic
inefficiency: moving data around is expensive in terms of bandwidth, latency, and power …
inefficiency: moving data around is expensive in terms of bandwidth, latency, and power …
Accelerating pattern matching queries in hybrid CPU-FPGA architectures
Taking advantage of recently released hybrid multicore architectures, such as the Intel's
Xeon+ FPGA machine, where the FPGA has coherent access to the main memory through …
Xeon+ FPGA machine, where the FPGA has coherent access to the main memory through …
Quicksel: Quick selectivity learning with mixture models
Y Park, S Zhong, B Mozafari - Proceedings of the 2020 ACM SIGMOD …, 2020 - dl.acm.org
Estimating the selectivity of a query is a key step in almost any cost-based query optimizer.
Most of today's databases rely on histograms or samples that are periodically refreshed by …
Most of today's databases rely on histograms or samples that are periodically refreshed by …
Aquoman: An analytic-query offloading machine
Analytic workloads on terabyte data-sets are often run in the cloud, where application and
storage servers are separate and connected via network. In order to saturate the storage …
storage servers are separate and connected via network. In order to saturate the storage …
Fpga-based data partitioning
Implementing parallel operators in multi-core machines often involves a data partitioning
step that divides the data into cache-size blocks and arranges them so to allow concurrent …
step that divides the data into cache-size blocks and arranges them so to allow concurrent …
UDP: a programmable accelerator for extract-transform-load workloads and more
Big data analytic applications give rise to large-scale extract-transform-load (ETL) as a
fundamental step to transform new data into a native representation. ETL workloads pose …
fundamental step to transform new data into a native representation. ETL workloads pose …