Haotian Zhang, PhD candidate
David R. Cheriton School of Computer Science
Dynamic sampling (DS) is applied to create a sampled set of relevance judgments in our participation of TREC Common Core Track 2018. One goal was to test the effectiveness and efficiency of this technique with a set of non-expert, secondary relevance assessors. We consider NIST assessors to be the experts and the primary assessors. Another goal was to make available to other researchers a sampled set of relevance judgments (prels) and thus allow the estimation of retrieval metrics that have the potential to be more robust than the standard NIST provided relevance judgments (qrels). In addition to creating the prels, we also submitted several runs based on our manual judging and the models produced by our HiCAL system.
Zeynep Korkmaz, PhD candidate
David R. Cheriton School of Computer Science
Analysis on graphs have powerful impact on solving many social and scientific problems, and applications often perform expensive traversals on large scale graphs. Caching approaches on top of persistent storage are among the classical solutions to handle high request throughput. However, graph processing applications have poor access locality, and caching algorithms do not improve disk I/O sufficiently. We present GAL, a graph-aware layout for disk-resident graph databases that generates a storage layout for large-scale graphs on disk with the objective of increasing locality of disk blocks and reducing the number of I/O operations for transactional workloads.
Panos K. Chrysanthis, University of Pittsburgh
Abstract: Online analytics, in most advanced scientific, business, and defense applications, rely heavily on the efficient execution of large numbers of Aggregate Continuous Queries (ACQs). ACQs continuously aggregate streaming data and periodically produce results such as max or average over a given window of the latest data. It was shown that in processing ACQs it is beneficial to use incremental evaluation, which involves storing and reusing calculations performed over the unchanged parts of the window, rather than performing the re-evaluation of the entire window after each update.
Jaemyung Kim, PhD candidate
David R. Cheriton School of Computer Science
Transaction durability guarantees the ability to recover committed transactions from failures. However, making every transaction durable impacts transaction processing performance. Some ad-hoc durability mechanisms (e.g., delayed durability) improve performance, but they risk transactions losing their effects due to failures. The current one-size-fits-all transaction durability model does not solve this problem. We propose a new generalized transaction durability model to trade-off performance and durability and argue that transactions should also provide flexible durability like they provide multiple isolation levels. We evaluate the performance of a modified PostgreSQL that supports the new durability model using a micro-benchmark to show the durability/performance trade-offs.
Hemant Saxena, PhD candidate
David R. Cheriton School of Computer Science
We address the problem of discovering dependencies from distributed big data. Existing (non-distributed) algorithms focus on minimizing computation by pruning the search space of possible dependencies. However, distributed algorithms must also optimize data communication costs, especially in current shared-nothing settings. To do this, we define a set of primitives for dependency discovery, which corresponds to data processing steps separated by communication barriers, and we present efficient implementations that optimize both computation and communication costs. Using real data, we show that algorithms built using our primitives are significantly faster and more communication-efficient than straightforward distributed implementations.
A. Erdem Sarıyüce, University at Buffalo
Abstract: Finding dense substructures in a network is a fundamental graph mining operation, with applications in bioinformatics, social networks, and visualization to name a few. Yet most standard formulations of this problem (like clique, quasi-clique, densest at-least-k subgraph) are NP-hard. Furthermore, the goal is rarely to find the “true optimum” but to identify many (if not all) dense substructures, understand their distribution in the graph, and ideally determine relationships among them. In this talk, I will talk about a framework that we designed to find dense regions of the graph with hierarchical relations.
Amine Mhedhbi, PhD candidate
David R. Cheriton School of Computer Science
We study the problem of optimizing subgraph queries (SQs) using the new worst-case optimal (WCO) join plans in Selinger-style cost-based optimizers. WCO plans evaluate SQs by matching one query vertex at a time using multiway intersections. The core problem in optimizing WCO plans is to pick an ordering of the query vertices to match.
Mustafa Korkmaz, PhD candidate
David R. Cheriton School of Computer Science
Mina Farid, PhD candidate
David R. Cheriton School of Computer Science
RDF has become a prevalent format to represent disparate data that is ingested from heterogeneous sources. However, data often contains errors due to extraction, transformation, and integration problems, leading to missing or contradicting information that propagate to downstream applications.
Speaker: Mohammad Sadoghi, UC Davis
Haotian Zhang, PhD candidate
David R. Cheriton School of Computer Science
Speaker: Stratos Idreos, Harvard University
Paolo Atzeni, Database Professor and Head of the Department of Engineering
Università Roma Tre
NoSQL systems have gained their popularity for many reasons, including the flexibility they provide with modeling, which tries to relax the rigidity provided by the relational model and by the other structured models.
Anil Pacaci, PhD candidate
David R. Cheriton School of Computer Science
Torben Bach Pedersen, Professor of Computer Science
Aalborg University, Denmark
Data collected from new sources such as sensors and smart devices is large, fast, and often complex. There is a universal wish to perform multidimensional OLAP-style analytics on such data, i.e., to turn it into “Big Multidimensional Data”. Supporting this is a multi-stage journey, requiring new tools and systems, and forming a new, extended data cycle with models as a key concept.
Panos Ipeirotis, Professor and George A. Kellner Faculty Fellow
Department of Information, Operations, and Management Sciences
New York University
Dallas Fraser, Master’s candidate
David R. Cheriton School of Computer Science
Combining text and mathematics when searching in a corpus with extensive mathematical notation remains an open problem. Recent results for math information retrieval systems on the math and text retrieval task at NTCIR-12, for example, show room for improvement, even though formula retrieval appears to be fairly successful.
Babar Naveed Memon, Master’s candidate
David R. Cheriton School of Computer Science
Remote Direct Memory Access (RDMA) can be used to implement a shared storage abstraction or a shared nothing abstraction for distributed applications. We argue that the shared storage abstraction is an overkill for loosely coupled applications and that the shared nothing abstraction does not leverage all the benefits of RDMA.
Daniel Lemire
Université Télug
Maximizing performance in data engineering is a daunting challenge. We present some of our work on designing faster indexes, with a particular emphasis on compressed indexes. Some of our prior work includes (1) Roaring indexes which are part of multiple big-data systems such as Spark, Hive, Druid, Atlas, Pinot, Kylin, (2) EWAH indexes are part of Git (GitHub) and included in major Linux distributions.
Rachel Pottinger, Department of Computer Science
University of British Columbia
Users are faced with an increasing onslaught of data, whether it's in their choices of movies to watch, assimilating data from multiple sources, or finding information relevant to their lives on open data registries.
Barzan Mozafari, Department of Computer Science and Engineering
University of Michigan
Lei Zou, Institute of Computer Science and Technology
Peking University
In this talk, I focus on accelerating a widely employed computing pattern — set intersection, to boost a group of relevant graph algorithms. Graph’s adjacency-lists can be naturally considered as node sets, thus set intersection is a primitive operation in many graph algorithms. We propose QFilter, a set intersection algorithm using SIMD instructions. QFilter adopts a merge-based framework and compares two blocks of elements iteratively by SIMD instructions.
Jennifer Widom
Frederick Emmons Terman Dean, School of Engineering
Fletcher Jones Professor, Computer Science and Electrical Engineering
Stanford University
Xi He, PhD candidate
Computer Science Department, Duke University
Michael Abebe, PhD candidate
David R. Cheriton School of Computer Science