%%% -*-BibTeX-*-
%%% ====================================================================
%%% BibTeX-file{
%%% author = "Nelson H. F. Beebe",
%%% version = "1.19",
%%% date = "06 August 2020",
%%% time = "09:04:21 MDT",
%%% filename = "topc.bib",
%%% address = "University of Utah
%%% Department of Mathematics, 110 LCB
%%% 155 S 1400 E RM 233
%%% Salt Lake City, UT 84112-0090
%%% USA",
%%% telephone = "+1 801 581 5254",
%%% FAX = "+1 801 581 4148",
%%% URL = "http://www.math.utah.edu/~beebe",
%%% checksum = "36536 6948 38872 363870",
%%% email = "beebe at math.utah.edu, beebe at acm.org,
%%% beebe at computer.org (Internet)",
%%% codetable = "ISO/ASCII",
%%% keywords = "ACM Transactions on Parallel Computing
%%% (TOPC); bibliography; BibTeX",
%%% license = "public domain",
%%% supported = "yes",
%%% docstring = "This is a COMPLETE BibTeX bibliography for
%%% ACM Transactions on Parallel Computing (TOPC)
%%% (CODEN ????, ISSN 2329-4949 (print),
%%% 2329-4957 (electronic)). The journal appears
%%% quarterly, and publication began with volume
%%% 1, number 1, in September 2014.
%%%
%%% At version 1.20, the COMPLETE journal
%%% coverage looked like this:
%%%
%%% 2014 ( 8) 2017 ( 16) 2020 ( 20)
%%% 2015 ( 29) 2018 ( 22)
%%% 2016 ( 25) 2019 ( 29)
%%%
%%% Article: 149
%%%
%%% Total entries: 149
%%%
%%% The journal Web page can be found at:
%%%
%%% http://topc.acm.org/
%%%
%%% The journal table of contents page is at:
%%%
%%% http://dl.acm.org/pub.cfm?id=J1496
%%% http://dl.acm.org/citation.cfm?id=2632163
%%%
%%% Qualified subscribers can retrieve the full
%%% text of recent articles in PDF form.
%%%
%%% The initial draft was extracted from the ACM
%%% Web pages.
%%%
%%% ACM copyrights explicitly permit abstracting
%%% with credit, so article abstracts, keywords,
%%% and subject classifications have been
%%% included in this bibliography wherever
%%% available. Article reviews have been
%%% omitted, until their copyright status has
%%% been clarified.
%%%
%%% bibsource keys in the bibliography entries
%%% below indicate the entry originally came
%%% from the computer science bibliography
%%% archive, even though it has likely since
%%% been corrected and updated.
%%%
%%% URL keys in the bibliography point to
%%% World Wide Web locations of additional
%%% information about the entry.
%%%
%%% BibTeX citation tags are uniformly chosen
%%% as name:year:abbrev, where name is the
%%% family name of the first author or editor,
%%% year is a 4-digit number, and abbrev is a
%%% 3-letter condensation of important title
%%% words. Citation tags were automatically
%%% generated by software developed for the
%%% BibNet Project.
%%%
%%% In this bibliography, entries are sorted in
%%% publication order, using ``bibsort -byvolume.''
%%%
%%% The checksum field above contains a CRC-16
%%% checksum as the first value, followed by the
%%% equivalent of the standard UNIX wc (word
%%% count) utility output of lines, words, and
%%% characters. This is produced by Robert
%%% Solovay's checksum utility.",
%%% }
%%% ====================================================================
@Preamble{"\input bibnames.sty" #
"\def \TM {${}^{\sc TM}$}"
}
%%% ====================================================================
%%% Acknowledgement abbreviations:
@String{ack-nhfb = "Nelson H. F. Beebe,
University of Utah,
Department of Mathematics, 110 LCB,
155 S 1400 E RM 233,
Salt Lake City, UT 84112-0090, USA,
Tel: +1 801 581 5254,
FAX: +1 801 581 4148,
e-mail: \path|beebe@math.utah.edu|,
\path|beebe@acm.org|,
\path|beebe@computer.org| (Internet),
URL: \path|http://www.math.utah.edu/~beebe/|"}
%%% ====================================================================
%%% Journal abbreviations:
@String{j-TOPC = "ACM Transactions on Parallel Computing
(TOPC)"}
%%% ====================================================================
%%% Bibliography entries:
@Article{Gibbons:2014:ATP,
author = "Phillip B. Gibbons",
title = "{ACM Transactions on Parallel Computing}: an
introduction",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "1:1--1:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2661651",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Techniques",
acknowledgement = ack-nhfb,
articleno = "1",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Lilja:2014:I,
author = "David J. Lilja",
title = "Introduction",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "2:1--2:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2609798",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "2",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Rane:2014:EPO,
author = "Ashay Rane and James Browne",
title = "Enhancing Performance Optimization of Multicore\slash
Multichip Nodes with Data Structure Metrics",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "3:1--3:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2588788",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Program performance optimization is usually based
solely on measurements of execution behavior of code
segments using hardware performance counters. However,
memory access patterns are critical performance
limiting factors for today's multicore chips where
performance is highly memory bound. Therefore diagnoses
and selection of optimizations based only on
measurements of the execution behavior of code segments
are incomplete because they do not incorporate
knowledge of memory access patterns and behaviors. This
article presents a low-overhead tool (MACPO) that
captures memory traces and computes metrics for the
memory access behavior of source-level (C, C++,
Fortran) data structures. MACPO explicitly targets the
measurement and metrics important to performance
optimization for multicore chips. The article also
presents a complete process for integrating measurement
and analyses of code execution with measurements and
analyses of memory access patterns and behaviors for
performance optimization, specifically targeting
multicore chips and multichip nodes of clusters. MACPO
uses more realistic cache models for computation of
latency metrics than those used by previous tools.
Evaluation of the effectiveness of adding memory access
behavior characteristics of data structures to
performance optimization was done on subsets of the
ASCI, NAS and Rodinia parallel benchmarks and two
versions of one application program from a domain not
represented in these benchmarks. Adding characteristics
of the behavior of data structures enabled easier
diagnoses of bottlenecks and more accurate selection of
appropriate optimizations than with only code centric
behavior measurements. The performance gains ranged
from a few percent to 38 percent.",
acknowledgement = ack-nhfb,
articleno = "3",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Jimenez:2014:APP,
author = "V{\'\i}ctor Jim{\'e}nez and Francisco J. Cazorla and
Roberto Gioiosa and Alper Buyuktosunoglu and Pradip
Bose and Francis P. O'Connell and Bruce G. Mealey",
title = "Adaptive Prefetching on {POWER7}: Improving
Performance and Power Consumption",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "4:1--4:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2588889",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Hardware data prefetch engines are integral parts of
many general purpose server-class microprocessors in
the field today. Some prefetch engines allow users to
change some of their parameters. But, the prefetcher is
usually enabled in a default configuration during
system bring-up, and dynamic reconfiguration of the
prefetch engine is not an autonomic feature of current
machines. Conceptually, however, it is easy to infer
that commonly used prefetch algorithms---when applied
in a fixed mode---will not help performance in many
cases. In fact, they may actually degrade performance
due to useless bus bandwidth consumption and cache
pollution, which in turn, will also waste power. We
present an adaptive prefetch scheme that dynamically
modifies the prefetch settings in order to adapt to
workloads' requirements. We use a commercial processor,
namely the IBM POWER7 as a vehicle for our study. First
we characterize---in terms of performance and power
consumption---the prefetcher in that processor using
microbenchmarks and SPEC CPU2006. We then present our
adaptive prefetch mechanism showing performance
improvements with respect to the default prefetch
setting up to 2.7X and 1.3X for single-threaded and
multiprogrammed workloads, respectively. Adaptive
prefetching is also able to reduce power consumption in
some cases. Finally, we also evaluate our mechanism
with SPECjbb2005, improving both performance and power
consumption.",
acknowledgement = ack-nhfb,
articleno = "4",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Heil:2014:APH,
author = "Timothy Heil and Anil Krishna and Nicholas Lindberg
and Farnaz Toussi and Steven Vanderwiel",
title = "Architecture and Performance of the Hardware
Accelerators in {IBM}'s {PowerEN} Processor",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "5:1--5:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2588888",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Computation at the edge of a datacenter has unique
characteristics. It deals with streaming data from
multiple sources, going to multiple destinations, often
requiring repeated application of one or more of
several standard algorithmic kernels. These kernels,
related to encryption, compression, XML Parsing and
regular expression searching on the data, demand a high
data processing rate and power efficiency. This
suggests the use of hardware acceleration for key
functions. However, robust general purpose processing
support is necessary to orchestrate the flow of data
between accelerators, as well as perform tasks that are
not suited to acceleration. Further, these accelerators
must be tightly integrated with the general purpose
computation in order to keep invocation overhead and
latency low. The accelerators must be easy for software
to use, and the system must be flexible enough to
support evolving networking standards. In this article,
we describe and evaluate the architecture of IBM's
PowerEN processor, with a focus on PowerEN's
architectural enhancements and its on-chip hardware
accelerators. PowerEN unites the throughput of
application-specific accelerators with the
programmability of general purpose cores on a single
coherent memory architecture. Hardware acceleration
improves throughput by orders of magnitude in some
cases compared to equivalent computation on the general
purpose cores. By offloading work to the accelerators,
general purpose cores are freed to simultaneously work
on computation less suited to acceleration.",
acknowledgement = ack-nhfb,
articleno = "5",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Wu:2014:MAG,
author = "Xing Wu and Frank Mueller and Scott Pakin",
title = "A methodology for automatic generation of executable
communication specifications from parallel {MPI}
applications",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "6:1--6:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2660249",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Portable parallel benchmarks are widely used for
performance evaluation of HPC systems. However, because
these are manually produced, they generally represent a
greatly simplified view of application behavior,
missing the subtle but important-to-performance nuances
that may exist in a complete application. This work
contributes novel methods to automatically generate
highly portable and customizable communication
benchmarks from HPC applications. We utilize
ScalaTrace, a lossless yet scalable
parallel-application tracing framework to collect
selected aspects of the run-time behavior of HPC
applications, including communication operations and
computation time, while abstracting away the details of
the computation proper. We subsequently generate
benchmarks with nearly identical run-time behavior to
the original applications. Results demonstrate that the
generated benchmarks are in fact able to preserve the
run-time behavior (including both the communication
pattern and the execution time) of the original
applications. Such automated benchmark generation is
without precedent and particularly valuable for
proprietary, export-controlled, or classified
application codes.",
acknowledgement = ack-nhfb,
articleno = "6",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ravishankar:2014:APC,
author = "Mahesh Ravishankar and John Eisenlohr and
Louis-No{\"e}l Pouchet and J. Ramanujam and Atanas
Rountev and P. Sadayappan",
title = "Automatic parallelization of a class of irregular
loops for distributed memory systems",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "7:1--7:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2660251",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Many scientific applications spend significant time
within loops that are parallel, except for dependences
from associative reduction operations. However these
loops often contain data-dependent control-flow and
array-access patterns. Traditional optimizations that
rely on purely static analysis fail to generate
parallel code in such cases. This article proposes an
approach for automatic parallelization for distributed
memory environments, using both static and runtime
analysis. We formalize the computations that are
targeted by this approach and develop algorithms to
detect such computations. We also describe algorithms
to generate a parallel inspector that performs a
runtime analysis of control-flow and array-access
patterns, and a parallel executor to take advantage of
this information. The effectiveness of the approach is
demonstrated on several benchmarks that were
automatically transformed using a prototype compiler.
For these, the inspector overheads and performance of
the executor code were measured. The benefit on
real-world applications was also demonstrated through
similar manual transformations of an atmospheric
modeling software.",
acknowledgement = ack-nhfb,
articleno = "7",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Shun:2014:SPC,
author = "Julian Shun and Guy E. Blelloch",
title = "A simple parallel {Cartesian} tree algorithm and its
application to parallel suffix tree construction",
journal = j-TOPC,
volume = "1",
number = "1",
pages = "8:1--8:??",
month = sep,
year = "2014",
CODEN = "????",
DOI = "https://doi.org/10.1145/2661653",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Oct 17 12:28:03 MDT 2014",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present a simple linear work and space, and
polylogarithmic time parallel algorithm for generating
multiway Cartesian trees. We show that bottom-up
traversals of the multiway Cartesian tree on the
interleaved suffix array and longest common prefix
array of a string can be used to answer certain string
queries. By adding downward pointers in the tree (e.g.
using a hash table), we can also generate suffix trees
from suffix arrays on arbitrary alphabets in the same
bounds. In conjunction with parallel suffix array
algorithms, such as the skew algorithm, this gives a
rather simple linear work parallel, $ O(n \epsilon) $
time $ (0 < \epsilon < 1) $, algorithm for generating
suffix trees over an integer alphabet $ \Sigma \{ 1,
\ldots {}, n \} $, where $n$ is the length of the input
string. It also gives a linear work parallel algorithm
requiring $ O(\log_2 n) $ time with high probability
for constant-sized alphabets. More generally, given a
sorted sequence of strings and the longest common
prefix lengths between adjacent elements, the algorithm
will generate a patricia tree (compacted trie) over the
strings. Of independent interest, we describe a
work-efficient parallel algorithm for solving the all
nearest smaller values problem using Cartesian trees,
which is much simpler than the work-efficient parallel
algorithm described in previous work. We also present
experimental results comparing the performance of the
algorithm to existing sequential implementations and a
second parallel algorithm that we implement. We present
comparisons for the Cartesian tree algorithm on its own
and for constructing a suffix tree. The results show
that on a variety of strings our algorithm is
competitive with the sequential version on a single
processor and achieves good speedup on multiple
processors. We present experiments for three
applications that require only the Cartesian tree, and
also for searching using the suffix tree.",
acknowledgement = ack-nhfb,
articleno = "8",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Pingali:2015:ISI,
author = "Keshav Pingali and J. Ramanujam and P. Sadayappan",
title = "Introduction to the Special Issue on {PPoPP'12}",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "9:1--9:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2716343",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "9",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bouteiller:2015:ABF,
author = "Aurelien Bouteiller and Thomas Herault and George
Bosilca and Peng Du and Jack Dongarra",
title = "Algorithm-Based Fault Tolerance for Dense Matrix
Factorizations, Multiple Failures and Accuracy",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "10:1--10:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2686892",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/bibnet/authors/d/dongarra-jack-j.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Dense matrix factorizations, such as LU, Cholesky and
QR, are widely used for scientific applications that
require solving systems of linear equations,
eigenvalues and linear least squares problems. Such
computations are normally carried out on
supercomputers, whose ever-growing scale induces a fast
decline of the Mean Time To Failure (MTTF). This
article proposes a new hybrid approach, based on
Algorithm-Based Fault Tolerance (ABFT), to help matrix
factorizations algorithms survive fail-stop failures.
We consider extreme conditions, such as the absence of
any reliable node and the possibility of losing both
data and checksum from a single failure. We will
present a generic solution for protecting the right
factor, where the updates are applied, of all above
mentioned factorizations. For the left factor, where
the panel has been applied, we propose a scalable
checkpointing algorithm. This algorithm features high
degree of checkpointing parallelism and cooperatively
utilizes the checksum storage leftover from the right
factor protection. The fault-tolerant algorithms
derived from this hybrid solution is applicable to a
wide range of dense matrix factorizations, with minor
modifications. Theoretical analysis shows that the
fault tolerance overhead decreases inversely to the
scaling in the number of computing units and the
problem size. Experimental results of LU and QR
factorization on the Kraken (Cray XT5) supercomputer
validate the theoretical evaluation and confirm
negligible overhead, with- and without-errors.
Applicability to tolerate multiple failures and
accuracy after multiple recovery is also considered.",
acknowledgement = ack-nhfb,
articleno = "10",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ballard:2015:ACS,
author = "Grey Ballard and James Demmel and Nicholas Knight",
title = "Avoiding Communication in Successive Band Reduction",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "11:1--11:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2686877",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The running time of an algorithm depends on both
arithmetic and communication (i.e., data movement)
costs, and the relative costs of communication are
growing over time. In this work, we present sequential
and distributed-memory parallel algorithms for
tridiagonalizing full symmetric and symmetric band
matrices that asymptotically reduce communication
compared to previous approaches. The tridiagonalization
of a symmetric band matrix is a key kernel in solving
the symmetric eigenvalue problem for both full and band
matrices. In order to preserve structure,
tridiagonalization routines use annihilate-and-chase
procedures that previously have suffered from poor data
locality and high parallel latency cost. We improve
both by reorganizing the computation and obtain
asymptotic improvements. We also propose new algorithms
for reducing a full symmetric matrix to band form in a
communication-efficient manner. In this article, we
consider the cases of computing eigenvalues only and of
computing eigenvalues and all eigenvectors.",
acknowledgement = ack-nhfb,
articleno = "11",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Sack:2015:CAM,
author = "Paul Sack and William Gropp",
title = "Collective Algorithms for Multiported Torus Networks",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "12:1--12:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2686882",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Modern supercomputers with torus networks allow each
node to simultaneously pass messages on all of its
links. However, most collective algorithms are designed
to only use one link at a time. In this work, we
present novel multiported algorithms for the scatter,
gather, all-gather, and reduce-scatter operations. Our
algorithms can be combined to create multiported
reduce, all-reduce, and broadcast algorithms. Several
of these algorithms involve a new technique where we
relax the MPI message-ordering constraints to achieve
high performance and restore the correct ordering using
an additional stage of redundant communication.
According to our models, on an $n$-dimensional torus,
our algorithms should allow for nearly a $ 2 n$-fold
improvement in communication performance compared to
known, single-ported torus algorithms. In practice, we
have achieved nearly $ 6 \times $ better performance on
a 32k-node 3-dimensional torus.",
acknowledgement = ack-nhfb,
articleno = "12",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Dice:2015:LCG,
author = "David Dice and Virendra J. Marathe and Nir Shavit",
title = "Lock Cohorting: a General Technique for Designing
{NUMA} Locks",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "13:1--13:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2686884",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Multicore machines are quickly shifting to NUMA and
CC-NUMA architectures, making scalable NUMA-aware
locking algorithms, ones that take into account the
machine's nonuniform memory and caching hierarchy, ever
more important. This article presents lock cohorting, a
general new technique for designing NUMA-aware locks
that is as simple as it is powerful. Lock cohorting
allows one to transform any spin-lock algorithm, with
minimal nonintrusive changes,into a scalable NUMA-aware
spin-lock. Our new cohorting technique allows us to
easily create NUMA-aware versions of the TATAS-Backoff,
CLH, MCS, and ticket locks, to name a few. Moreover, it
allows us to derive a CLH-based cohort abortable lock,
the first NUMA-aware queue lock to support
abortability. We empirically compared the performance
of cohort locks with prior NUMA-aware and classic
NUMA-oblivious locks on a synthetic micro-benchmark, a
real world key-value store application memcached, as
well as the libc memory allocator. Our results
demonstrate that cohort locks perform as well or better
than known locks when the load is low and significantly
out-perform them as the load increases.",
acknowledgement = ack-nhfb,
articleno = "13",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Merrill:2015:HPS,
author = "Duane Merrill and Michael Garland and Andrew
Grimshaw",
title = "High-Performance and Scalable {GPU} Graph Traversal",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "14:1--14:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2717511",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Breadth-First Search (BFS) is a core primitive for
graph traversal and a basis for many higher-level graph
analysis algorithms. It is also representative of a
class of parallel computations whose memory accesses
and work distribution are both irregular and data
dependent. Recent work has demonstrated the
plausibility of GPU sparse graph traversal, but has
tended to focus on asymptotically inefficient
algorithms that perform poorly on graphs with
nontrivial diameter. We present a BFS parallelization
focused on fine-grained task management constructed
from efficient prefix sum computations that achieves an
asymptotically optimal O(|V| + |E|) gd work complexity.
Our implementation delivers excellent performance on
diverse graphs, achieving traversal rates in excess of
3.3 billion and 8.3 billion traversed edges per second
using single- and quad-GPU configurations,
respectively. This level of performance is several
times faster than state-of-the-art implementations on
both CPU and GPU platforms.",
acknowledgement = ack-nhfb,
articleno = "14",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kramer:2015:SET,
author = "Stephan C. Kramer and Johannes Hagemann",
title = "{SciPAL}: Expression Templates and Composition Closure
Objects for High Performance Computational Physics with
{CUDA} and {OpenMP}",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "15:1--15:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2686886",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present SciPAL (scientific parallel algorithms
library), a C ++-based, hardware-independent
open-source library. Its core is a domain-specific
embedded language for numerical linear algebra. The
main fields of application are finite element
simulations, coherent optics and the solution of
inverse problems. Using SciPAL algorithms can be stated
in a mathematically intuitive way in terms of matrix
and vector operations. Existing algorithms can easily
be adapted to GPU-based computing by proper template
specialization. Our library is compatible with the
finite element library deal.II and provides a port of
deal.II's most frequently used linear algebra classes
to CUDA (NVidia's extension of the programming
languages C and C ++ for programming their GPUs).
SciPAL 's operator-based API for BLAS operations
particularly aims at simplifying the usage of NVidia's
CUBLAS. For non-BLAS array arithmetic SciPAL 's
expression templates are able to generate CUDA kernels
at compile time. We demonstrate the benefits of SciPAL
using the iterative principal component analysis as
example which is the core algorithm for the
spike-sorting problem in neuroscience.",
acknowledgement = ack-nhfb,
articleno = "15",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Totoni:2015:PME,
author = "Ehsan Totoni and Nikhil Jain and Laxmikant V. Kale",
title = "Power Management of Extreme-Scale Networks with
On\slash Off Links in Runtime Systems",
journal = j-TOPC,
volume = "1",
number = "2",
pages = "16:1--16:??",
month = jan,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2687001",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Feb 18 16:46:00 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Networks are among major power consumers in
large-scale parallel systems. During execution of
common parallel applications, a sizeable fraction of
the links in the high-radix interconnects are either
never used or are underutilized. We propose a runtime
system based adaptive approach to turn off unused
links, which has various advantages over the previously
proposed hardware and compiler based approaches. We
discuss why the runtime system is the best system
component to accomplish this task, and test the
effectiveness of our approach using real applications
(including NAMD, MILC), and application benchmarks
(including NAS Parallel Benchmarks, Stencil). These
codes are simulated on representative topologies such
as 6-D Torus and multilevel directly connected network
(similar to IBM PERCS in Power 775 and Dragonfly in
Cray Aries). For common applications with near-neighbor
communication pattern, our approach can save up to 20\%
of total machine's power and energy, without any
performance penalty.",
acknowledgement = ack-nhfb,
articleno = "16",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Herlihy:2015:GEI,
author = "Maurice Herlihy",
title = "{Guest Editor} Introduction",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "1:1--1:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2716306",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "1",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Degener:2015:LCS,
author = "Bastian Degener and Barbara Kempkes and Peter Kling
and Friedhelm {Meyer Auf Der Heide}",
title = "Linear and Competitive Strategies for Continuous Robot
Formation Problems",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "2:1--2:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2742341",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We study a scenario in which n mobile robots with a
limited viewing range are distributed in the Euclidean
plane and have to solve a formation problem. The
formation problems we consider are the Gathering
problem and the Chain-Formation problem. In the
Gathering problem, the robots have to gather in one
(not predefined) point, while in the Chain-Formation
problem they have to form a connected communication
chain of minimal length between two stationary base
stations. Each robot may base its decisions where to
move only on the current relative positions of
neighboring robots (that are within its viewing range);
that is, besides having a limited viewing range, the
robots are oblivious (they do not use information from
the past), have none or only very limited identities,
and they do not have a common sense of direction.
Variants of these problems (especially for the
Gathering problem) have been studied extensively in
different discrete time models. In contrast, our work
focuses on a continuous time model; that is, the robots
continuously sense the positions of other robots within
their viewing range and continuously adapt their speed
and direction according to some simple, local rules.
Hereby, we assume that the robots have a maximum
movement speed of one. We show that this idealized idea
of continuous sensing allows us to solve the mentioned
formation problems in linear time $ O(n) $ (which,
given the maximum speed of one, immediately yields a
maximum traveled distance of $ O(n)$). Note that in the
more classical discrete time models, the best known
strategies need at least $ \Theta (n^2)$ or even $
\Theta (n^2 \log n)$ timesteps to solve these problems.
For the Gathering problem, our analysis solves a
problem left open by Gordon et al. [2004], where the
authors could prove that gathering in a continuous
model is possible in finite time, but were not able to
give runtime bounds. Apart from these linear bounds, we
also provide runtime bounds for both formation problems
that relate the runtime of our strategies to the
runtime of an optimal, global algorithm. Specifically,
we show that our strategy for the Gathering problem is
log OPT-competitive and the strategy for the
Chain-Formation problem is $ \log n$ competitive. Here,
by $c$-competitive, we mean that our (local) strategy
is asymptotically by at most a factor of $c$ slower
than an optimal, global strategy.",
acknowledgement = ack-nhfb,
articleno = "2",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Jain:2015:NOS,
author = "Navendu Jain and Ishai Menache and Joseph (Seffi) Naor
and Jonathan Yaniv",
title = "Near-Optimal Scheduling Mechanisms for
Deadline-Sensitive Jobs in Large Computing Clusters",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "3:1--3:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2742343",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We consider a market-based resource allocation model
for batch jobs in cloud computing clusters. In our
model, we incorporate the importance of the due date of
a job rather than the number of servers allocated to it
at any given time. Each batch job is characterized by
the work volume of total computing units (e.g., CPU
hours) along with a bound on maximum degree of
parallelism. Users specify, along with these job
characteristics, their desired due date and a value for
finishing the job by its deadline. Given this
specification, the primary goal is to determine the
scheduling of cloud computing instances under capacity
constraints in order to maximize the social welfare
(i.e., sum of values gained by allocated users). Our
main result is a new (CC-kcss-1)-approximation
algorithm for this objective, where $C$ denotes cloud
capacity, $k$ is the maximal bound on parallelized
execution (in practical settings, $ k < C$) and $s$ is
the slackness on the job completion time, that is, the
minimal ratio between a specified deadline and the
earliest finish time of a job. Our algorithm is based
on utilizing dual fitting arguments over a strengthened
linear program to the problem. Based on the new
approximation algorithm, we construct truthful
allocation and pricing mechanisms, in which reporting
the true value and other properties of the job
(deadline, work volume, and the parallelism bound) is a
dominant strategy for all users. To that end, we extend
known results for single-value settings to provide a
general framework for transforming allocation
algorithms into truthful mechanisms in domains of
single-value and multi-properties. We then show that
the basic mechanism can be extended under proper
Bayesian assumptions to the objective of maximizing
revenues, which is important for public clouds. We
empirically evaluate the benefits of our approach
through simulations on data-center job traces, and show
that the revenues obtained under our mechanism are
comparable with an ideal fixed-price mechanism, which
sets an on-demand price using oracle knowledge of
users' valuations. Finally, we discuss how our model
can be extended to accommodate uncertainties in job
work volumes, which is a practical challenge in cloud
settings.",
acknowledgement = ack-nhfb,
articleno = "3",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Feldman:2015:HCG,
author = "Moran Feldman and Liane Lewin-Eytan and Joseph (Seffi)
Naor",
title = "Hedonic Clustering Games",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "4:1--4:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2742345",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Clustering, the partitioning of objects with respect
to a similarity measure, has been extensively studied
as a global optimization problem. We investigate
clustering from a game-theoretic approach, and consider
the class of hedonic clustering games. Here, a
self-organized clustering is obtained via decisions
made by independent players, corresponding to the
elements clustered. Being a hedonic setting, the
utility of each player is determined by the identity of
the other members of her cluster. This class of games
seems to be quite robust, as it fits with rather
different, yet commonly used, clustering criteria.
Specifically, we investigate hedonic clustering games
in two different models: fixed clustering, which
subdivides into $k$-median and $k$-center, and
correlation clustering. We provide a thorough analysis
of these games, characterizing Nash equilibria, and
proving upper and lower bounds on the price of anarchy
and price of stability. For fixed clustering we focus
on the existence of a Nash equilibrium, as it is a
rather nontrivial issue in this setting. We study it
both for general metrics and special cases, such as
line and tree metrics. In the correlation clustering
model, we study both minimization and maximization
variants, and provide almost tight bounds on both the
price of anarchy and price of stability.",
acknowledgement = ack-nhfb,
articleno = "4",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Jahn:2015:RRA,
author = "Janmartin Jahn and Santiago Pagani and Sebastian Kobbe
and Jian-Jia Chen and J{\"o}rg Henkel",
title = "Runtime Resource Allocation for Software Pipelines",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "5:1--5:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2742347",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Efficiently allocating the computational resources of
many-core systems is one of the most prominent
challenges, especially when resource requirements may
vary unpredictably at runtime. This is even more
challenging when facing unreliable cores --- a scenario
that becomes common as the number of cores increases
and integration sizes shrink. To address this
challenge, this article presents an optimal method for
the allocation of the resources to software-pipelined
applications. Here we show how runtime observations of
the resource requirements of tasks can be used to adapt
resource allocations. Furthermore, we show how the
optimum can be traded for a high degree of scalability
by clustering applications in a distributed,
hierarchical manner. To diminish the negative effects
of unreliable cores, this article shows how
self-organization can effectively restore the integrity
of such a hierarchy when it is corrupted by a failing
core. Experiments on Intel's 48-core Single-Chip Cloud
Computer and in a many-core simulator show that a
significant improvement in system throughput can be
achieved over the current state of the art.",
acknowledgement = ack-nhfb,
articleno = "5",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Xu:2015:SVC,
author = "Yi Xu and Bo Zhao and Youtao Zhang and Jun Yang",
title = "Simple Virtual Channel Allocation for High-Throughput
and High-Frequency On-Chip Routers",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "6:1--6:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2742349",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Packet-switched network-on-chip (NoC) has provided a
scalable solution to the communications for tiled
multicore processors. However, the virtual channel (VC)
buffers in the NoC consume significant dynamic and
leakage power. To improve the energy efficiency of the
router design, it is advantageous to use small buffer
sizes while still maintaining throughput of the
network. This article proposes two new virtual channel
allocation (VA) mechanisms, termed fixed VC assignment
with dynamic VC allocation (FVADA) and adjustable VC
assignment with dynamic VC allocation (AVADA). VCs are
designated to output ports and allocated to packets
according to such assignment. This can help to reduce
the head-of-line blocking. Such VC-output port
assignment can also be adjusted dynamically to
accommodate traffic changes. Simulation results show
that both mechanisms can improve network throughput by
41\% on average. Real traffic evaluation shows a
network latency reduction of up to 66\%. In addition,
AVADA can outperform the baseline in throughput with
only half of the buffer size. Finally, we are able to
achieve comparable or better throughput than a previous
dynamic VC allocator while reducing its critical path
delay by 57\%. Hence, the proposed VA mechanisms are
suitable for low-power, high-throughput, and
high-frequency NoC designs.",
acknowledgement = ack-nhfb,
articleno = "6",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Hammouda:2015:NTE,
author = "Adam Hammouda and Andrew R. Siegel and Stephen F.
Siegel",
title = "Noise-Tolerant Explicit Stencil Computations for
Nonuniform Process Execution Rates",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "7:1--7:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2742351",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Next-generation HPC computing platforms are likely to
be characterized by significant, unpredictable
nonuniformities in execution time among compute nodes
and cores. The resulting load imbalances from this
nonuniformity are expected to arise from a variety of
sources-manufacturing discrepancies, dynamic power
management, runtime component failure, OS jitter,
software-mediated resiliency, and TLB/- cache
performance variations, for example. It is well
understood that existing algorithms with frequent
points of bulk synchronization will perform relatively
poorly in the presence of these sources of process
nonuniformity. Thus, recasting classic bulk synchronous
algorithms into more asynchronous, coarse-grained
parallelism is a critical area of research for
next-generation computing. We propose a class of
parallel algorithms for explicit stencil computations
that can tolerate these nonuniformities by decoupling
per process communication and computation in order for
each process to progress asynchronously while
maintaining solution correctness. These algorithms are
benchmarked with a $1$D domain decomposed (``slabbed'')
implementation of the $2$D heat equation as a model
problem, and are tested in the presence of simulated
nonuniform process execution rates. The resulting
performance is compared to a classic bulk synchronous
implementation of the model problem. Results show that
the runtime of this article's algorithm on a machine
with simulated process nonuniformities is 5--99\%
slower than the runtime of its classic counterpart on a
machine free of nonuniformities. However, when both
algorithms are run on a machine with comparable
synthetic process nonuniformities, this article's
algorithm is 1--37 times faster than its classic
counterpart.",
acknowledgement = ack-nhfb,
articleno = "7",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{McCreesh:2015:SST,
author = "Ciaran McCreesh and Patrick Prosser",
title = "The Shape of the Search Tree for the Maximum Clique
Problem and the Implications for Parallel Branch and
Bound",
journal = j-TOPC,
volume = "2",
number = "1",
pages = "8:1--8:??",
month = may,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2742359",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu May 21 16:27:00 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Finding a maximum clique in a given graph is one of
the fundamental NP-hard problems. We compare two
multicore thread-parallel adaptations of a
state-of-the-art branch-and-bound algorithm for the
maximum clique problem and provide a novel explanation
as to why they are successful. We show that load
balance is sometimes a problem but that the interaction
of parallel search order and the most likely location
of solutions within the search space is often the
dominating consideration. We use this explanation to
propose a new low-overhead, scalable work-splitting
mechanism. Our approach uses explicit early diversity
to avoid strong commitment to the weakest heuristic
advice and late resplitting for balance. More
generally, we argue that, for branch-and-bound,
parallel algorithm design should not be performed
independently of the underlying sequential algorithm.",
acknowledgement = ack-nhfb,
articleno = "8",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Hoefler:2015:RMA,
author = "Torsten Hoefler and James Dinan and Rajeev Thakur and
Brian Barrett and Pavan Balaji and William Gropp and
Keith Underwood",
title = "Remote Memory Access Programming in {MPI-3}",
journal = j-TOPC,
volume = "2",
number = "2",
pages = "9:1--9:??",
month = jul,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2780584",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Aug 7 10:22:35 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/pvm.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The Message Passing Interface (MPI) 3.0 standard,
introduced in September 2012, includes a significant
update to the one-sided communication interface, also
known as remote memory access (RMA). In particular, the
interface has been extended to better support popular
one-sided and global-address-space parallel programming
models to provide better access to hardware performance
features and enable new data-access modes. We present
the new RMA interface and specify formal axiomatic
models for data consistency and access semantics. Such
models can help users reason about details of the
semantics that are hard to extract from the English
prose in the standard. It also fosters the development
of tools and compilers, enabling them to automatically
analyze, optimize, and debug RMA programs.",
acknowledgement = ack-nhfb,
articleno = "9",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Maldonado:2015:STB,
author = "Walther Maldonado and Patrick Marlier and Pascal
Felber and Julia Lawall and Gilles Muller and Etienne
Rivi{\`e}re",
title = "Supporting Time-Based {QoS} Requirements in Software
Transactional Memory",
journal = j-TOPC,
volume = "2",
number = "2",
pages = "10:1--10:??",
month = jul,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2779621",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Aug 7 10:22:35 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Software transactional memory (STM) is an optimistic
concurrency control mechanism that simplifies parallel
programming. However, there has been little interest in
its applicability to reactive applications in which
there is a required response time for certain
operations. We propose supporting such applications by
allowing programmers to associate time with atomic
blocks in the form of deadlines and quality-of-service
(QoS) requirements. Based on statistics of past
executions, we adjust the execution mode of
transactions by decreasing the level of optimism as the
deadline approaches. In the presence of concurrent
deadlines, we propose different conflict resolution
policies. Execution mode switching mechanisms allow the
meeting of multiple deadlines in a consistent manner,
with potential QoS degradations being split fairly
among several threads as contention increases, and
avoiding starvation. Our implementation consists of
extensions to an STM runtime that allow gathering
statistics and switching execution modes. We also
propose novel contention managers adapted to
transactional workloads subject to deadlines. The
experimental evaluation shows that our approaches
significantly improve the likelihood of a transaction
meeting its deadline and QoS requirement, even in cases
where progress is hampered by conflicts and other
concurrent transactions with deadlines.",
acknowledgement = ack-nhfb,
articleno = "10",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kestor:2015:TPD,
author = "Gokcen Kestor and Osman S. Unsal and Adrian Cristal
and Serdar Tasiran",
title = "{TRADE}: Precise Dynamic Race Detection for Scalable
Transactional Memory Systems",
journal = j-TOPC,
volume = "2",
number = "2",
pages = "11:1--11:??",
month = jul,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2786021",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Aug 7 10:22:35 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/multithreading.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "As other multithreaded programs, transactional memory
(TM) programs are prone to race conditions. Previous
work focuses on extending existing definitions of data
race for lock-based applications to TM applications,
which requires all transactions to be totally ordered
``as if'' serialized by a global lock. This approach
poses implementation constraints on the STM that
severely limits TM applications' performance. This
article shows that forcing total ordering among all
running transactions, while sufficient, is not
necessary. We introduce an alternative data race
definition, relaxed transactional data race, that
requires ordering of only conflicting transactions. The
advantages of our relaxed definition are twofold:
First, unlike the previous definition, this definition
can be applied to a wide range of TMs, including those
that do not enforce transaction total ordering. Second,
within a single execution, it exposes a higher number
of data races, which considerably reduces debugging
time. Based on this definition, we propose a novel and
precise race detection tool for C/C++ TM applications
(TRADE), which detects data races by tracking
happens-before edges among conflicting transactions.
Our experiments reveal that TRADE precisely detects
data races for STAMP applications running on modern
STMs with overhead comparable to state-of-the-art race
detectors for lock-based applications. Our experiments
also show that in a single run, TRADE identifies
several races not discovered by 10 separate runs of a
race detection tool based on the previous data race
definition.",
acknowledgement = ack-nhfb,
articleno = "11",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Diegues:2015:TWE,
author = "Nuno Diegues and Paolo Romano",
title = "{Time-Warp}: Efficient Abort Reduction in
Transactional Memory",
journal = j-TOPC,
volume = "2",
number = "2",
pages = "12:1--12:??",
month = jul,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2775435",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Aug 7 10:22:35 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The multicore revolution that took place one decade
ago has turned parallel programming into a major
concern for the mainstream software development
industry. In this context, Transactional Memory (TM)
has emerged as a simpler and attractive alternative to
that of lock-based synchronization, whose complexity
and error-proneness are widely recognized. The notion
of permissiveness in TM translates to only aborting a
transaction when it cannot be accepted in any history
that guarantees a target correctness criterion. This
theoretically powerful property is often neglected by
state-of-the-art TMs because it imposes considerable
algorithmic costs. Instead, these TMs opt to maximize
their implementation's efficiency by aborting
transactions under overly conservative conditions. As a
result, they risk rejecting a significant number of
safe executions. In this article, we seek to identify a
sweet spot between permissiveness and efficiency by
introducing the Time-Warp Multiversion (TWM) algorithm.
TWM is based on the key idea of allowing an update
transaction that has performed stale reads (i.e.,
missed the writes of concurrently committed
transactions) to be serialized by ``committing it in
the past,'' which we call a time-warp commit. At its
core, TWM uses a novel, lightweight validation
mechanism with little computational overhead. TWM also
guarantees that read-only transactions can never be
aborted. Further, TWM guarantees Virtual World
Consistency, a safety property that is deemed as
particularly relevant in the context of TM. We
demonstrate the practicality of this approach through
an extensive experimental study: we compare TWM with
five other TMs, representative of typical alternative
design choices, and on a wide variety of benchmarks.
This study shows an average performance improvement
across all considered workloads and TMs of 65\% in high
concurrency scenarios, with gains extending up to $ 9
\times $ with the most favorable benchmarks. These
results are a consequence of TWM's ability to achieve
drastic reduction of aborts in scenarios of nonminimal
contention, while introducing little overhead
(approximately 10\%) in worst-case, synthetically
designed scenarios (i.e., no contention or contention
patterns that cannot be optimized using TWM).",
acknowledgement = ack-nhfb,
articleno = "12",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Eyraud-Dubois:2015:PST,
author = "Lionel Eyraud-Dubois and Loris Marchal and Oliver
Sinnen and Fr{\'e}d{\'e}ric Vivien",
title = "Parallel Scheduling of Task Trees with Limited
Memory",
journal = j-TOPC,
volume = "2",
number = "2",
pages = "13:1--13:??",
month = jul,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2779052",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Aug 7 10:22:35 MDT 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "This article investigates the execution of tree-shaped
task graphs using multiple processors. Each edge of
such a tree represents some large data. A task can only
be executed if all input and output data fit into
memory, and a data can only be removed from memory
after the completion of the task that uses it as an
input data. Such trees arise in the multifrontal method
of sparse matrix factorization. The peak memory needed
for the processing of the entire tree depends on the
execution order of the tasks. With one processor, the
objective of the tree traversal is to minimize the
required memory. This problem was well studied, and
optimal polynomial algorithms were proposed. Here, we
extend the problem by considering multiple processors,
which is of obvious interest in the application area of
matrix factorization. With multiple processors comes
the additional objective to minimize the time needed to
traverse the tree-that is, to minimize the makespan.
Not surprisingly, this problem proves to be much harder
than the sequential one. We study the computational
complexity of this problem and provide
inapproximability results even for unit weight trees.
We design a series of practical heuristics achieving
different trade-offs between the minimization of peak
memory usage and makespan. Some of these heuristics are
able to process a tree while keeping the memory usage
under a given memory limit. The different heuristics
are evaluated in an extensive experimental evaluation
using realistic trees.",
acknowledgement = ack-nhfb,
articleno = "13",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Dinitz:2015:ISI,
author = "Michael Dinitz and Torsten Hoefler",
title = "Introduction to the Special Issue on {SPAA 2013}",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "14:1--14:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2809923",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "14e",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kumar:2015:FGA,
author = "Ravi Kumar and Benjamin Moseley and Sergei
Vassilvitskii and Andrea Vattani",
title = "Fast Greedy Algorithms in {MapReduce} and Streaming",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "14:1--14:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2809814",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Greedy algorithms are practitioners' best
friends---they are intuitive, are simple to implement,
and often lead to very good solutions. However,
implementing greedy algorithms in a distributed setting
is challenging since the greedy choice is inherently
sequential, and it is not clear how to take advantage
of the extra processing power. Our main result is a
powerful sampling technique that aids in
parallelization of sequential algorithms. Armed with
this primitive, we then adapt a broad class of greedy
algorithms to the MapReduce paradigm; this class
includes maximum cover and submodular maximization
subject to p -system constraint problems. Our method
yields efficient algorithms that run in a logarithmic
number of rounds while obtaining solutions that are
arbitrarily close to those produced by the standard
sequential greedy algorithm. We begin with algorithms
for modular maximization subject to a matroid
constraint and then extend this approach to obtain
approximation algorithms for submodular maximization
subject to knapsack or p -system constraints.",
acknowledgement = ack-nhfb,
articleno = "14",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Sanders:2015:WEM,
author = "Peter Sanders and Jochen Speck and Raoul Steffen",
title = "Work-Efficient Matrix Inversion in Polylogarithmic
Time",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "15:1--15:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2809812",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present an algorithm for inversion of symmetric
positive definite matrices that combines the practical
requirement of an optimal number of arithmetic
operations and the theoretical goal of a
polylogarithmic critical path length. The algorithm
reduces inversion to matrix multiplication. It uses
Strassen's recursion scheme, but on the critical path
it breaks the recursion early, switching to an
asymptotically inefficient yet fast use of Newton's
method. We also show that the algorithm is numerically
stable. Overall, we get a candidate for a massively
parallel algorithm that scales to exascale systems even
on relatively small inputs.",
acknowledgement = ack-nhfb,
articleno = "15",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Gilbert:2015:SBO,
author = "Seth Gilbert and Chaodong Zheng",
title = "{SybilCast}: Broadcast on the Open Airwaves",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "16:1--16:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2809810",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Consider a scenario where many wireless users are
attempting to download data from a single base station.
While most of the users are honest, some users may be
malicious and attempt to obtain more than their fair
share of the bandwidth. One possible strategy for
attacking the system is to simulate multiple fake
identities, each of which is given its own equal share
of the bandwidth. Such an attack is often referred to
as a sybil attack. To counter such behavior, we propose
SybilCast, a protocol for multichannel wireless
networks that limits the number of fake identities and,
in doing so, ensures that each honest user gets at
least a constant fraction of his or her fair share of
the bandwidth. As a result, each honest user can
complete his or her data download in asymptotically
optimal time. A key aspect of this protocol is
balancing the rate at which new identities are admitted
and the maximum number of fake identities that can
coexist while keeping the overhead low. Besides sybil
attacks, our protocol can also tolerate spoofing and
jamming.",
acknowledgement = ack-nhfb,
articleno = "16",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Lee:2015:FPP,
author = "I-Ting Angelina Lee and Charles E. Leiserson and Tao
B. Schardl and Zhunping Zhang and Jim Sukha",
title = "On-the-Fly Pipeline Parallelism",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "17:1--17:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2809808",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Pipeline parallelism organizes a parallel program as a
linear sequence of stages. Each stage processes
elements of a data stream, passing each processed data
element to the next stage, and then taking on a new
element before the subsequent stages have necessarily
completed their processing. Pipeline parallelism is
used especially in streaming applications that perform
video, audio, and digital signal processing. Three out
of 13 benchmarks in PARSEC, a popular software
benchmark suite designed for shared-memory
multiprocessors, can be expressed as pipeline
parallelism. Whereas most concurrency platforms that
support pipeline parallelism use a
``construct-and-run'' approach, this article
investigates ``on-the-fly'' pipeline parallelism, where
the structure of the pipeline emerges as the program
executes rather than being specified a priori.
On-the-fly pipeline parallelism allows the number of
stages to vary from iteration to iteration and
dependencies to be data dependent. We propose simple
linguistics for specifying on-the-fly pipeline
parallelism and describe a provably efficient
scheduling algorithm, the P iper algorithm, which
integrates pipeline parallelism into a work-stealing
scheduler, allowing pipeline and fork-join parallelism
to be arbitrarily nested. The Piper algorithm
automatically throttles the parallelism, precluding
``runaway'' pipelines. Given a pipeline computation
with $ T_1 $ work and $ T_\infty $ span (critical-path
length), Piper executes the computation on $P$
processors in $ T_P \leq T_1 / P + O(T \infty + \lg P)$
expected time. Piper also limits stack space, ensuring
that it does not grow unboundedly with running time. We
have incorporated on-the-fly pipeline parallelism into
a Cilk-based work-stealing runtime system. Our
prototype Cilk-P implementation exploits optimizations
such as ``lazy enabling'' and ``dependency folding.''
We have ported the three PARSEC benchmarks that exhibit
pipeline parallelism to run on Cilk-P. One of these,
x264, cannot readily be executed by systems that
support only construct-and-run pipeline parallelism.
Benchmark results indicate that Cilk-P has low serial
overhead and good scalability. On x264, for example,
Cilk-P exhibits a speedup of 13.87 over its respective
serial counterpart when running on 16 processors.",
acknowledgement = ack-nhfb,
articleno = "17",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Eikel:2015:IRI,
author = "Martina Eikel and Christian Scheideler",
title = "{IRIS}: a Robust Information System Against Insider
{DoS} Attacks",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "18:1--18:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2809806",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "In this work, we present the first scalable
distributed information system, that is, a system with
low storage overhead, that is provably robust against
denial-of-service (DoS) attacks by a current insider.
We allow a current insider to have complete knowledge
about the information system and to have the power to
block any \varepsilon-fraction of its servers by a DoS
attack, where \varepsilon can be chosen up to a
constant. The task of the system is to serve any
collection of lookup requests with at most one per
nonblocked server in an efficient way despite this
attack. Previously, scalable solutions were only known
for DoS attacks of past insiders, where a past insider
only has complete knowledge about some past time point
$ t_0 $ of the information system. Scheideler et al.
[Awerbuch and Scheideler 2007; Baumgart et al. 2009]
showed that in this case, it is possible to design an
information system so that any information that was
inserted or last updated after $ t_0 $ is safe against
a DoS attack. But their constructions would not work at
all for a current insider. The key idea behind our IRIS
system is to make extensive use of coding. More
precisely, we present two alternative distributed
coding strategies with an at most logarithmic storage
overhead that can handle up to a constant fraction of
blocked servers.",
acknowledgement = ack-nhfb,
articleno = "18",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kling:2015:PSM,
author = "Peter Kling and Peter Pietrzyk",
title = "Profitable Scheduling on Multiple Speed-Scalable
Processors",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "19:1--19:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2809872",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present a new online algorithm for profit-oriented
scheduling on multiple speed-scalable processors and
provide a tight analysis of the algorithm's
competitiveness. Our results generalize and improve
upon work by Chan et al. [2010], which considers a
single speed-scalable processor. Using significantly
different techniques, we can not only extend their
model to multiprocessors but also prove an enhanced and
tight competitive ratio for our algorithm. In our
scheduling problem, jobs arrive over time and are
preemptable. They have different workloads, values, and
deadlines. The scheduler may decide not to finish a job
but instead to suffer a loss equaling the job's value.
However, to process a job's workload until its deadline
the scheduler must invest a certain amount of energy.
The cost of a schedule is the sum of lost values and
invested energy. In order to finish a job, the
scheduler has to determine which processors to use and
set their speeds accordingly. A processor's energy
consumption is power $ P_\alpha (s) $ integrated over
time, where $ P_\alpha (s) = s^\alpha $ is the power
consumption when running at speed $s$. Since we
consider the online variant of the problem, the
scheduler has no knowledge about future jobs. This
problem was introduced by Chan et al. [2010] for the
case of a single processor. They presented an online
algorithm that is $ \alpha^\alpha + 2 e \alpha
$-competitive. We provide an online algorithm for the
case of multiple processors with an improved
competitive ratio of $ \alpha^\alpha $.",
acknowledgement = ack-nhfb,
articleno = "19",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Dutta:2015:CBR,
author = "Chinmoy Dutta and Gopal Pandurangan and Rajmohan
Rajaraman and Scott Roche",
title = "Coalescing-Branching Random Walks on Graphs",
journal = j-TOPC,
volume = "2",
number = "3",
pages = "20:1--20:??",
month = oct,
year = "2015",
CODEN = "????",
DOI = "https://doi.org/10.1145/2817830",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Nov 3 07:30:42 MST 2015",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We study a distributed randomized information
propagation mechanism in networks we call the
coalescing-branching random walk (cobra walk, for
short). A cobra walk is a generalization of the
well-studied ``standard'' random walk, and is useful in
modeling and understanding the Susceptible-Infected-
Susceptible (SIS)-type of epidemic processes in
networks. It can also be helpful in performing
light-weight information dissemination in
resource-constrained networks. A cobra walk is
parameterized by a branching factor $k$. The process
starts from an arbitrary vertex, which is labeled
active for step 1. In each step of a cobra walk, each
active vertex chooses $k$ random neighbors to become
active for the next step (``branching''). A vertex is
active for step $ t + 1$ only if it is chosen by an
active vertex in step $t$ (``coalescing''). This
results in a stochastic process in the underlying
network with properties that are quite different from
both the standard random walk (which is equivalent to
the cobra walk with branching factor 1) as well as
other gossip-based rumor spreading mechanisms. We focus
on the cover time of the cobra walk, which is the
number of steps for the walk to reach all the vertices,
and derive almost-tight bounds for various graph
classes. We show an $ O(\log^2 n)$ high probability
bound for the cover time of cobra walks on expanders,
if either the expansion factor or the branching factor
is sufficiently large; we also obtain an $ O(\log n)$
high probability bound for the partial cover time,
which is the number of steps needed for the walk to
reach at least a constant fraction of the vertices. We
also show that the cover time of the cobra walk is,
with high probability, $ O(n \log n)$ on any $n$-vertex
tree for $ k \geq 2$, $ {\~ O}(n^{1 / d})$ on a
$d$-dimensional grid for $ k \geq 2$, and $ O(\log n)$
on the complete graph.",
acknowledgement = ack-nhfb,
articleno = "20",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Larus:2016:ISI,
author = "James Larus and Sandhya Dwarkadas and Jos{\'e} Moreira
and Andrew Lumsdaine",
title = "Introduction to the Special Issue on {PPoPP'14}",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "21:1--21:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2856513",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "21",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{Herlihy:2016:WSF,
author = "Maurice Herlihy and Zhiyu Liu",
title = "Well-Structured Futures and Cache Locality",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "22:1--22:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2858650",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "In fork-join parallelism, a sequential program is
split into a directed acyclic graph of tasks linked by
directed dependency edges, and the tasks are executed,
possibly in parallel, in an order consistent with their
dependencies. A popular and effective way to extend
fork-join parallelism is to allow threads to create
futures. A thread creates a future to hold the results
of a computation, which may or may not be executed in
parallel. That result is returned when some thread
touches that future, blocking if necessary until the
result is ready. Recent research has shown that
although futures can, of course, enhance parallelism in
a structured way, they can have a deleterious effect on
cache locality. In the worst case, futures can incur $
\Omega (P T_\infty + t T_\infty) $ deviations, which
implies $ \Omega (C P T_\infty + C t T_\infty) $
additional cache misses, where $C$ is the number of
cache lines, $P$ is the number of processors, $t$ is
the number of touches, and $ T_\infty $ is the
computation span. Since cache locality has a large
impact on software performance on modern multicores,
this result is troubling. In this article, we show that
if futures are used in a simple, disciplined way, then
the situation is much better: if each future is touched
only once, either by the thread that created it or by a
later descendant of the thread that created it, then
parallel executions with work stealing can incur at
most $ O(C P T^2_\infty)$ additional cache misses-a
substantial improvement. This structured use of futures
is characteristic of many (but not all) parallel
applications.",
acknowledgement = ack-nhfb,
articleno = "22",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{Thomson:2016:CTU,
author = "Paul Thomson and Alastair F. Donaldson and Adam
Betts",
title = "Concurrency Testing Using Controlled Schedulers: an
Empirical Study",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "23:1--23:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2858651",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present an independent empirical study on
concurrency testing using controlled schedulers. We
have gathered 49 buggy concurrent software benchmarks,
drawn from public code bases, which we call SCTBench.
We applied a modified version of an existing
concurrency testing tool to SCTBench, testing five
controlled scheduling techniques: depth-first search,
preemption bounding, delay bounding, a controlled
random scheduler, and probabilistic concurrency testing
(PCT). We attempt to answer several research questions:
Which technique performs the best, in terms of bug
finding ability? How effective are the two main
schedule bounding techniques-preemption bounding and
delay bounding-at finding bugs? What challenges are
associated with applying concurrency testing techniques
to existing code? Can we classify certain benchmarks as
trivial or nontrivial? Overall, we found that PCT (with
parameter d = 3) was the most effective technique in
terms of bug finding; it found all bugs found by the
other techniques, plus an additional three, and it
missed only one bug. Surprisingly, we found that the
naive controlled random scheduler, which randomly
chooses one thread to execute at each scheduling point,
performs well, finding more bugs than preemption
bounding and just two fewer bugs than delay bounding.
Our findings confirm that delay bounding is superior to
preemption bounding and that schedule bounding is
superior to an unbounded depth-first search. The
majority of bugs in SCTBench can be exposed using a
small schedule bound (1--2), supporting previous
claims, although one benchmark requires five
preemptions. We found that the need to remove
nondeterminism and control all synchronization (as is
required for systematic concurrency testing) can be
nontrivial. There were eight distinct programs that
could not easily be included in out study, such as
those that perform network and interprocess
communication. We report various properties about the
benchmarks tested, such as the fact that the bugs in 18
benchmarks were exposed 50\% of the time when using
random scheduling. We note that future work should not
use the benchmarks that we classify as trivial when
presenting new techniques, other than as a minimum
baseline. We have made SCTBench and our tools publicly
available for reproducibility and use in future work.",
acknowledgement = ack-nhfb,
articleno = "23",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{Petrovic:2016:LHM,
author = "Darko Petrovi{\'c} and Thomas Ropars and Andr{\'e}
Schiper",
title = "Leveraging Hardware Message Passing for Efficient
Thread Synchronization",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "24:1--24:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2858652",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "As the level of parallelism in manycore processors
keeps increasing, providing efficient mechanisms for
thread synchronization in concurrent programs is
becoming a major concern. On cache-coherent
shared-memory processors, synchronization efficiency is
ultimately limited by the performance of the underlying
cache coherence protocol. This article studies how
hardware support for message passing can improve
synchronization performance. Considering the ubiquitous
problem of mutual exclusion, we devise novel algorithms
for (i) classic locking, where application threads
obtain exclusive access to a shared resource prior to
executing their critical sections (CSes), and (ii)
delegation, where CSes are executed by special threads.
For classic locking, our HybLock algorithm uses a mix
of shared memory and hardware message passing, which
introduces the idea of hybrid synchronization
algorithms. For delegation, we propose mp-server and
HybComb: the former is a straightforward adaptation of
the server approach to hardware message passing,
whereas the latter is a novel hybrid combining
algorithm. Evaluation on Tilera's TILE-Gx processor
shows that HybLock outperforms the best known classic
locks. Furthermore, mp-server can execute contended
CSes with unprecedented throughput, as stalls related
to cache coherence are removed from the critical path.
HybComb can achieve comparable performance while
avoiding the need to dedicate server cores.
Consequently, our queue and stack implementations,
based on the new synchronization algorithms, largely
outperform their most efficient shared-memory-only
counterparts.",
acknowledgement = ack-nhfb,
articleno = "24",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{Tardieu:2016:XAP,
author = "Olivier Tardieu and Benjamin Herta and David
Cunningham and David Grove and Prabhanjan Kambadur and
Vijay Saraswat and Avraham Shinnar and Mikio Takeuchi
and Mandana Vaziri and Wei Zhang",
title = "{X10} and {APGAS} at Petascale",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "25:1--25:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2894746",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "X10 is a high-performance, high-productivity
programming language aimed at large-scale distributed
and shared-memory parallel applications. It is based on
the Asynchronous Partitioned Global Address Space
(APGAS) programming model, supporting the same
fine-grained concurrency mechanisms within and across
shared-memory nodes. We demonstrate that X10 delivers
solid performance at petascale by running (weak
scaling) eight application kernels on an IBM Power--775
supercomputer utilizing up to 55,680 Power7 cores (for
1.7Pflop/s of theoretical peak performance). For the
four HPC Class 2 Challenge benchmarks, X10 achieves
41\% to 87\% of the system's potential at scale (as
measured by IBM's HPCC Class 1 optimized runs). We also
implement K-Means, Smith-Waterman, Betweenness
Centrality, and Unbalanced Tree Search (UTS) for
geometric trees. Our UTS implementation is the first to
scale to petaflop systems. We describe the advances in
distributed termination detection, distributed load
balancing, and use of high-performance interconnects
that enable X10 to scale out to tens of thousands of
cores. We discuss how this work is driving the
evolution of the X10 language, core class libraries,
and runtime systems.",
acknowledgement = ack-nhfb,
articleno = "25",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{Maleki:2016:LRM,
author = "Saeed Maleki and Madanlal Musuvathi and Todd
Mytkowicz",
title = "Low-Rank Methods for Parallelizing Dynamic Programming
Algorithms",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "26:1--26:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2884065",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "This article proposes efficient parallel methods for
an important class of dynamic programming problems that
includes Viterbi, Needleman--Wunsch, Smith--Waterman,
and Longest Common Subsequence. In dynamic programming,
the subproblems that do not depend on each other, and
thus can be computed in parallel, form stages or
wavefronts. The methods presented in this article
provide additional parallelism allowing multiple stages
to be computed in parallel despite dependencies among
them. The correctness and the performance of the
algorithm relies on rank convergence properties of
matrix multiplication in the tropical semiring, formed
with plus as the multiplicative operation and max as
the additive operation. This article demonstrates the
efficiency of the parallel algorithm by showing
significant speedups on a variety of important dynamic
programming problems. In particular, the parallel
Viterbi decoder is up to $ 24 \times $ faster (with 64
processors) than a highly optimized commercial
baseline.",
acknowledgement = ack-nhfb,
articleno = "26",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{Yuan:2016:FCN,
author = "Xin Yuan and Wickus Nienaber and Santosh Mahapatra",
title = "On Folded-{Clos} Networks with Deterministic
Single-Path Routing",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "27:1--27:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2858654",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Folded-Clos networks, also known as fat-trees, have
been widely used as interconnects in large-scale
high-performance computing clusters. Although users
often treat such interconnects as replacements of
nonblocking crossbar switches that can carry out any
permutation communication without contention, the
networking capability of such interconnects without a
centralized controller in computer communication
environments is not well understood. In this article,
we investigate nonblocking two-level folded-Clos
networks with deterministic single-path routing, but no
centralized controller, and establish the nonblocking
condition. The results indicate that nonblocking
two-level folded-Clos networks without a centralized
controller are much more expensive to construct than
the traditional nonblocking networks in the
telecommunication environment. Practical two-level
folded-Clos based interconnects are blocking. For such
interconnects, we establish the lower bound for
worst-case contention for permutations with any
deterministic single-path routing scheme, show that
existing routing schemes perform poorly in terms of
worst-case contention for permutations, present a
routing scheme that achieves the theoretical optimal,
and empirically compare the performance of existing
schemes with the optimal routing scheme. The techniques
developed for two-level folded-Clos networks are
further extended for the general fat-trees of any
heights.",
acknowledgement = ack-nhfb,
articleno = "27",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{Sandes:2016:MMA,
author = "Edans F. De O. Sandes and Guillermo Miranda and Xavier
Martorell and Eduard Ayguade and George Teodoro and
Alba C. M. A. {De Melo}",
title = "{MASA}: a Multiplatform Architecture for Sequence
Aligners with Block Pruning",
journal = j-TOPC,
volume = "2",
number = "4",
pages = "28:1--28:??",
month = mar,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2858656",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 19 08:11:13 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Biological sequence alignment is a very popular
application in Bioinformatics, used routinely
worldwide. Many implementations of biological sequence
alignment algorithms have been proposed for multicores,
GPUs, FPGAs and CellBEs. These implementations are
platform-specific; porting them to other systems
requires considerable programming effort. This article
proposes and evaluates MASA, a flexible and
customizable software architecture that enables the
execution of biological sequence alignment applications
with three variants (local, global, and semiglobal) in
multiple hardware/software platforms with block
pruning, which is able to reduce significantly the
amount of data processed. To attain our flexibility
goals, we also propose a generic version of block
pruning and developed multiple parallelization
strategies as building blocks, including a new
asynchronous dataflow-based parallelization, which may
be combined to implement efficient aligners in
different platforms. We provide four MASA aligner
implementations for multicores (OmpSs and OpenMP), GPU
(CUDA), and Intel Phi (OpenMP), showing that MASA is
very flexible. The evaluation of our generic block
pruning strategy shows that it significantly
outperforms the previously proposed block pruning,
being able to prune up to 66.5\% of the cells when
using the new dataflow-based parallelization
strategy.",
acknowledgement = ack-nhfb,
articleno = "28",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
remark = "Special Issue on PPoPP'14 conference.",
}
@Article{MeyeraufderHeide:2016:ISI,
author = "Friedhelm {Meyer auf der Heide} and Peter Sanders and
Nodari Sitchinava",
title = "Introduction to the Special Issue on {SPAA 2014}",
journal = j-TOPC,
volume = "3",
number = "1",
pages = "1:1--1:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2936716",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:51 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "1",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kaler:2016:EDD,
author = "Tim Kaler and William Hasenplaugh and Tao B. Schardl
and Charles E. Leiserson",
title = "Executing Dynamic Data-Graph Computations
Deterministically Using Chromatic Scheduling",
journal = j-TOPC,
volume = "3",
number = "1",
pages = "2:1--2:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2896850",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:51 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "A data-graph computation-popularized by such
programming systems as Galois, Pregel, GraphLab,
PowerGraph, and GraphChi-is an algorithm that performs
local updates on the vertices of a graph. During each
round of a data-graph computation, an update function
atomically modifies the data associated with a vertex
as a function of the vertex's prior data and that of
adjacent vertices. A dynamic data-graph computation
updates only an active subset of the vertices during a
round, and those updates determine the set of active
vertices for the next round. This article introduces
Prism, a chromatic-scheduling algorithm for executing
dynamic data-graph computations. Prism uses a vertex
coloring of the graph to coordinate updates performed
in a round, precluding the need for mutual-exclusion
locks or other nondeterministic data synchronization. A
multibag data structure is used by Prism to maintain a
dynamic set of active vertices as an unordered set
partitioned by color. We analyze Prism using work-span
analysis. Let $ G = (V, E) $ be a degree-$ \Delta $
graph colored with \chi colors, and suppose that $ Q
\subseteq V $ is the set of active vertices in a round.
Define $ {\rm size} (Q) = | Q | + \Sigma_{v \in Q} {\rm
deg}(v) $, which is proportional to the space required
to store the vertices of $Q$ using a sparse-graph
layout. We show that a $P$-processor execution of Prism
performs updates in $Q$ using $ O(\chi (l g (Q / \chi)
+ l g \Delta)) + l g P$ span and $ \Theta (s i z e (Q)
+ P)$ work. These theoretical guarantees are matched by
good empirical performance. To isolate the effect of
the scheduling algorithm on performance, we modified
GraphLab to incorporate Prism and studied seven
application benchmarks on a 12-core multicore machine.
Prism executes the benchmarks 1.2 to 2.1 times faster
than GraphLab's nondeterministic lock-based scheduler
while providing deterministic behavior. This article
also presents Prism-R, a variation of Prism that
executes dynamic data-graph computations
deterministically even when updates modify global
variables with associative operations. Prism-R
satisfies the same theoretical bounds as Prism, but its
implementation is more involved, incorporating a
multivector data structure to maintain a
deterministically ordered set of vertices partitioned
by color. Despite its additional complexity, Prism-R is
only marginally slower than Prism. On the seven
application benchmarks studied, Prism-R incurs a 7\%
geometric mean overhead relative to Prism.",
acknowledgement = ack-nhfb,
articleno = "2",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Im:2016:CST,
author = "Sungjin Im and Benjamin Moseley and Kirk Pruhs and
Eric Torng",
title = "Competitively Scheduling Tasks with Intermediate
Parallelizability",
journal = j-TOPC,
volume = "3",
number = "1",
pages = "4:1--4:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2938378",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:51 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We introduce a scheduling algorithm Intermediate-SRPT,
and show that it is $ O (\log P)$-competitive with
respect to average flow time when scheduling jobs whose
parallelizability is intermediate between being fully
parallelizable and sequential. Here, the parameter P
denotes the ratio between the maximum job size to the
minimum. We also show a general matching lower bound on
the competitive ratio. Our analysis builds on an
interesting combination of potential function and local
competitiveness arguments.",
acknowledgement = ack-nhfb,
articleno = "4",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bercea:2016:CMI,
author = "Ioana O. Bercea and Navin Goyal and David G. Harris
and Aravind Srinivasan",
title = "On Computing Maximal Independent Sets of Hypergraphs
in Parallel",
journal = j-TOPC,
volume = "3",
number = "1",
pages = "5:1--5:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2938436",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:51 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Whether or not the problem of finding maximal
independent sets (MIS) in hypergraphs is in (R)NC is
one of the fundamental problems in the theory of
parallel computing. Essentially, the challenge is to
design (randomized) algorithms in which the number of
processors used is polynomial and the (expected)
runtime is polylogarithmic in the size of the input.
Unlike the well-understood case of MIS in graphs, for
the hypergraph problem, our knowledge is quite limited
despite considerable work. It is known that the problem
is in RNC when the edges of the hypergraph have
constant size. For general hypergraphs with n vertices
and m edges, the fastest previously known algorithm
works in time $ O(\sqrt n) $ with $ \poly (m, n) $
processors. In this article, we give an EREW PRAM
randomized algorithm that works in time $ n^{o (1)} $
with $ O(n + m \log n) $ processors on general
hypergraphs satisfying $ m \leq n^{o (1) \log \log n /
\log \log \log n} $. We also give an EREW PRAM
deterministic algorithm that runs in time $ n^\epsilon
$ on a graph with $ m \leq n^{1 / \delta } $ edges, for
any constants $ \delta $, $ \epsilon $; the number of
processors is polynomial in $m$, $n$ for a fixed choice
of $ \delta $, $ \epsilon $. Our algorithms are based
on a sampling idea that reduces the dimension of the
hypergraph and employs the algorithm for constant
dimension hypergraphs as a subroutine.",
acknowledgement = ack-nhfb,
articleno = "5",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bilo:2016:LBN,
author = "Davide Bil{\`o} and Luciano Gual{\`a} and Stefano
Leucci and Guido Proietti",
title = "Locality-Based Network Creation Games",
journal = j-TOPC,
volume = "3",
number = "1",
pages = "6:1--6:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2938426",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:51 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Network creation games have been extensively studied,
both by economists and computer scientists, due to
their versatility in modeling individual-based
community formation processes. These processes, in
turn, are the theoretical counterpart of several
economics, social, and computational applications on
the Internet. In their several variants, these games
model the tension of a player between the player's two
antagonistic goals: to be as close as possible to the
other players and to activate a cheapest possible set
of links. However, the generally adopted assumption is
that players have a common and complete information
about the ongoing network, which is quite unrealistic
in practice. In this article, we consider a more
compelling scenario in which players have only limited
information about the network in which they are
embedded. More precisely, we explore the game-theoretic
and computational implications of assuming that players
have a complete knowledge of the network structure only
up to a given radius k, which is one of the most
qualified local-knowledge models used in distributed
computing. In this respect, we define a suitable
equilibrium concept, and we provide a comprehensive set
of upper and lower bounds to the price of anarchy for
the entire range of values of k and for the two classic
variants of the game, namely, those in which a player's
cost-besides the activation cost of the owned
links-depends on the maximum/sum of all distances to
the other nodes in the network, respectively. These
bounds are assessed through an extensive set of
experiments.",
acknowledgement = ack-nhfb,
articleno = "6",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Jiang:2016:PPA,
author = "Jiayang Jiang and Michael Mitzenmacher and Justin
Thaler",
title = "Parallel Peeling Algorithms",
journal = j-TOPC,
volume = "3",
number = "1",
pages = "7:1--7:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2938412",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:51 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The analysis of several algorithms and data structures
can be framed as a peeling process on a random
hypergraph: vertices with degree less than k are
removed until there are no vertices of degree less than
k left. The remaining hypergraph is known as the k
core. In this article, we analyze parallel peeling
processes, in which in each round, all vertices of
degree less than k are removed. It is known that, below
a specific edge-density threshold, the k -core is empty
with high probability. We show that, with high
probability, below this threshold, only $ 1 / \log ((k
- 1) (r - 1)) \log \log n + O (1) $ rounds of peeling
are needed to obtain the empty $k$-core for $ 4 b
r$-uniform hypergraphs. This bound is tight up to an
additive constant. Interestingly, we show that, above
this threshold, $ \Omega (\log n)$ rounds of peeling
are required to find the nonempty $k$-core. Since most
algorithms and data structures aim to peel to an empty
$k$-core, this asymmetry appears fortunate. We verify
the theoretical results both with simulation and with a
parallel implementation using graphics processing units
(GPUs). Our implementation provides insights into how
to structure parallel peeling algorithms for efficiency
in practice.",
acknowledgement = ack-nhfb,
articleno = "7",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Simhadri:2016:EAS,
author = "Harsha Vardhan Simhadri and Guy E. Blelloch and Jeremy
T. Fineman and Phillip B. Gibbons and Aapo Kyrola",
title = "Experimental Analysis of Space-Bounded Schedulers",
journal = j-TOPC,
volume = "3",
number = "1",
pages = "8:1--8:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2938389",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:51 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The running time of nested parallel programs on
shared-memory machines depends in significant part on
how well the scheduler mapping the program to the
machine is optimized for the organization of caches and
processor cores on the machine. Recent work proposed
``space-bounded schedulers'' for scheduling such
programs on the multilevel cache hierarchies of current
machines. The main benefit of this class of schedulers
is that they provably preserve locality of the program
at every level in the hierarchy, which can result in
fewer cache misses and better use of bandwidth than the
popular work-stealing scheduler. On the other hand,
compared to work stealing, space-bounded schedulers are
inferior at load balancing and may have greater
scheduling overheads, raising the question as to the
relative effectiveness of the two schedulers in
practice. In this article, we provide the first
experimental study aimed at addressing this question.
To facilitate this study, we built a flexible
experimental framework with separate interfaces for
programs and schedulers. This enables a head-to-head
comparison of the relative strengths of schedulers in
terms of running times and cache miss counts across a
range of benchmarks. (The framework is validated by
comparisons with the Intel{\reg} CilkTM Plus
work-stealing scheduler.) We present experimental
results on a 32-core Xeon{\reg} 7560 comparing work
stealing, hierarchy-minded work stealing, and two
variants of space-bounded schedulers on both
divide-and-conquer microbenchmarks and some popular
algorithmic kernels. Our results indicate that
space-bounded schedulers reduce the number of L3 cache
misses compared to work-stealing schedulers by 25\% to
65\% for most of the benchmarks, but incur up to 27\%
additional scheduler and load-imbalance overhead. Only
for memory-intensive benchmarks can the reduction in
cache misses overcome the added overhead, resulting in
up to a 25\% improvement in running time for synthetic
benchmarks and about 20\% improvement for algorithmic
kernels. We also quantify runtime improvements varying
the available bandwidth per core (the ``bandwidth
gap'') and show up to 50\% improvements in the running
times of kernels as this gap increases fourfold. As
part of our study, we generalize prior definitions of
space-bounded schedulers to allow for more practical
variants (while still preserving their guarantees) and
explore implementation tradeoffs.",
acknowledgement = ack-nhfb,
articleno = "8",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Sheikh:2016:SHJ,
author = "Hafiz Fahad Sheikh and Ishfaq Ahmad",
title = "Sixteen Heuristics for Joint Optimization of
Performance, Energy, and Temperature in Allocating
Tasks to Multi-Cores",
journal = j-TOPC,
volume = "3",
number = "2",
pages = "9:1--9:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2948973",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:52 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Three-way joint optimization of performance ( P ),
energy ( E ), and temperature ( T ) in scheduling
parallel tasks to multiple cores poses a challenge that
is staggering in its computational complexity. The goal
of the PET optimized scheduling ( PETOS ) problem is to
minimize three quantities: the completion time of a
task graph, the total energy consumption, and the peak
temperature of the system. Algorithms based on
conventional multi-objective optimization techniques
can be designed for solving the PETOS problem. But
their execution times are exceedingly high and hence
their applicability is restricted merely to problems of
modest size. Exacerbating the problem is the solution
space that is typically a Pareto front since no single
solution can be strictly best along all three
objectives. Thus, not only is the absolute quality of
the solutions important but ``the spread of the
solutions'' along each objective and the distribution
of solutions within the generated tradeoff front are
also desired. A natural alternative is to design
efficient heuristic algorithms that can generate good
solutions as well as good spreads --- note that most of
the prior work in energy-efficient task allocation is
predominantly single- or dual-objective oriented. Given
a directed acyclic graph (DAG) representing a parallel
program, a heuristic encompasses policies as to what
tasks should go to what cores and at what frequency
should that core operate. Various policies, such as
greedy, iterative, and probabilistic, can be employed.
However, the choice and usage of these policies can
influence a heuristic towards a particular objective
and can also profoundly impact its performance. This
article proposes 16 heuristics that utilize various
methods for task-to-core allocation and frequency
selection. This article also presents a methodical
classification scheme which not only categorizes the
proposed heuristics but can also accommodate additional
heuristics. Extensive simulation experiments compare
these algorithms while shedding light on their
strengths and tradeoffs.",
acknowledgement = ack-nhfb,
articleno = "9",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Blanchard:2016:SMO,
author = "Jeffrey D. Blanchard and Erik Opavsky and Emircan
Uysaler",
title = "Selecting Multiple Order Statistics with a Graphics
Processing Unit",
journal = j-TOPC,
volume = "3",
number = "2",
pages = "10:1--10:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2948974",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:52 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Extracting a set of multiple order statistics from a
huge data set provides important information about the
distribution of the values in the full set of data.
This article introduces an algorithm,
bucketMultiSelect, for simultaneously selecting
multiple order statistics with a graphics processing
unit (GPU). Typically, when a large set of order
statistics is desired, the vector is sorted. When the
sorted version of the vector is not needed,
bucketMultiSelect significantly reduces computation
time by eliminating a large portion of the unnecessary
operations involved in sorting. For large vectors,
bucketMultiSelect returns thousands of order statistics
in less time than sorting the vector while typically
using less memory. For vectors containing $ 2^{28} $
values of type double, bucketMultiSelect selects the
101 percentile order statistics in less than 95ms and
is more than $ 8 \times $ faster than sorting the
vector with a GPU optimized merge sort.",
acknowledgement = ack-nhfb,
articleno = "10",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bohme:2016:IRC,
author = "David B{\"o}hme and Markus Geimer and Lukas Arnold and
Felix Voigtlaender and Felix Wolf",
title = "Identifying the Root Causes of Wait States in
Large-Scale Parallel Applications",
journal = j-TOPC,
volume = "3",
number = "2",
pages = "11:1--11:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2934661",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:52 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Driven by growing application requirements and
accelerated by current trends in microprocessor design,
the number of processor cores on modern supercomputers
is increasing from generation to generation. However,
load or communication imbalance prevents many codes
from taking advantage of the available parallelism, as
delays of single processes may spread wait states
across the entire machine. Moreover, when employing
complex point-to-point communication patterns, wait
states may propagate along far-reaching cause-effect
chains that are hard to track manually and that
complicate an assessment of the actual costs of an
imbalance. Building on earlier work by Meira, Jr., et
al., we present a scalable approach that identifies
program wait states and attributes their costs in terms
of resource waste to their original cause. By replaying
event traces in parallel both forward and backward, we
can identify the processes and call paths responsible
for the most severe imbalances, even for runs with
hundreds of thousands of processes.",
acknowledgement = ack-nhfb,
articleno = "11",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Dathathri:2016:CAL,
author = "Roshan Dathathri and Ravi Teja Mullapudi and Uday
Bondhugula",
title = "Compiling Affine Loop Nests for a Dynamic Scheduling
Runtime on Shared and Distributed Memory",
journal = j-TOPC,
volume = "3",
number = "2",
pages = "12:1--12:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2948975",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:52 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Current de-facto parallel programming models like
OpenMP and MPI make it difficult to extract task-level
dataflow parallelism as opposed to bulk-synchronous
parallelism. Task parallel approaches that use
point-to-point synchronization between dependent tasks
in conjunction with dynamic scheduling dataflow
runtimes are thus becoming attractive. Although good
performance can be extracted for both shared and
distributed memory using these approaches, there is
little compiler support for them. In this article, we
describe the design of compiler--runtime interaction to
automatically extract coarse-grained dataflow
parallelism in affine loop nests for both shared and
distributed-memory architectures. We use techniques
from the polyhedral compiler framework to extract tasks
and generate components of the runtime that are used to
dynamically schedule the generated tasks. The runtime
includes a distributed decentralized scheduler that
dynamically schedules tasks on a node. The schedulers
on different nodes cooperate with each other through
asynchronous point-to-point communication, and all of
this is achieved by code automatically generated by the
compiler. On a set of six representative affine loop
nest benchmarks, while running on 32 nodes with 8
threads each, our compiler-assisted runtime yields a
geometric mean speedup of $ 143.6 \times $ ($ 70.3
\times $ to $ 474.7 \times $) over the sequential
version and a geometric mean speedup of $ 1.64 \times $
($ 1.04 \times $ to $ 2.42 \times $) over the
state-of-the-art automatic parallelization approach
that uses bulk synchronization. We also compare our
system with past work that addresses some of these
challenges on shared memory, and an emerging runtime
(Intel Concurrent Collections) that demands higher
programmer input and effort in parallelizing. To the
best of our knowledge, ours is also the first automatic
scheme that allows for dynamic scheduling of affine
loop nests on a cluster of multicores.",
acknowledgement = ack-nhfb,
articleno = "12",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Benoit:2016:AGP,
author = "Anne Benoit and Aur{\'e}lien Cavelan and Yves Robert
and Hongyang Sun",
title = "Assessing General-Purpose Algorithms to Cope with
Fail-Stop and Silent Errors",
journal = j-TOPC,
volume = "3",
number = "2",
pages = "13:1--13:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2897189",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:52 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "In this article, we combine the traditional
checkpointing and rollback recovery strategies with
verification mechanisms to cope with both fail-stop and
silent errors. The objective is to minimize makespan
and/or energy consumption. For divisible load
applications, we use first-order approximations to find
the optimal checkpointing period to minimize execution
time, with an additional verification mechanism to
detect silent errors before each checkpoint, hence
extending the classical formula by Young and Daly for
fail-stop errors only. We further extend the approach
to include intermediate verifications, and to consider
a bicriteria problem involving both time and energy
(linear combination of execution time and energy
consumption). Then, we focus on application workflows
whose dependence graph is a linear chain of tasks.
Here, we determine the optimal checkpointing and
verification locations, with or without intermediate
verifications, for the bicriteria problem. Rather than
using a single speed during the whole execution, we
further introduce a new execution scenario, which
allows for changing the execution speed via Dynamic
Voltage and Frequency Scaling (DVFS). In this latter
scenario, we determine the optimal checkpointing and
verification locations, as well as the optimal speed
pairs for each task segment between any two consecutive
checkpoints. Finally, we conduct an extensive set of
simulations to support the theoretical study, and to
assess the performance of each algorithm, showing that
the best overall performance is achieved under the most
flexible scenario using intermediate verifications and
different speeds.",
acknowledgement = ack-nhfb,
articleno = "13",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Koutis:2016:SPD,
author = "Ioannis Koutis and Shen Chen Xu",
title = "Simple Parallel and Distributed Algorithms for
Spectral Graph Sparsification",
journal = j-TOPC,
volume = "3",
number = "2",
pages = "14:1--14:??",
month = aug,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2948062",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Sep 23 15:24:52 MDT 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We describe simple algorithms for spectral graph
sparsification, based on iterative computations of
weighted spanners and sampling. Leveraging the
algorithms of Baswana and Sen for computing spanners,
we obtain the first distributed spectral sparsification
algorithm in the CONGEST model. We also obtain a
parallel algorithm with improved work and time
guarantees, as well as other natural distributed
implementations. Combining this algorithm with the
parallel framework of Peng and Spielman for solving
symmetric diagonally dominant linear systems, we get a
parallel solver that is significantly more efficient in
terms of the total work.",
acknowledgement = ack-nhfb,
articleno = "14",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Dorier:2016:DAP,
author = "Matthieu Dorier and Gabriel Antoniu and Franck
Cappello and Marc Snir and Robert Sisneros and Orcun
Yildiz and Shadi Ibrahim and Tom Peterka and Leigh
Orf",
title = "{Damaris}: Addressing Performance Variability in Data
Management for Post-Petascale Simulations",
journal = j-TOPC,
volume = "3",
number = "3",
pages = "15:1--15:??",
month = dec,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2987371",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Dec 26 17:40:41 MST 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "With exascale computing on the horizon, reducing
performance variability in data management tasks
(storage, visualization, analysis, etc.) is becoming a
key challenge in sustaining high performance. This
variability significantly impacts the overall
application performance at scale and its predictability
over time. In this article, we present Damaris, a
system that leverages dedicated cores in multicore
nodes to offload data management tasks, including I/O,
data compression, scheduling of data movements, in situ
analysis, and visualization. We evaluate Damaris with
the CM1 atmospheric simulation and the Nek5000
computational fluid dynamic simulation on four
platforms, including NICS's Kraken and NCSA's Blue
Waters. Our results show that (1) Damaris fully hides
the I/O variability as well as all I/O-related costs,
thus making simulation performance predictable; (2) it
increases the sustained write throughput by a factor of
up to 15 compared with standard I/O approaches; (3) it
allows almost perfect scalability of the simulation up
to over 9,000 cores, as opposed to state-of-the-art
approaches that fail to scale; and (4) it enables a
seamless connection to the VisIt visualization software
to perform in situ analysis and visualization in a way
that impacts neither the performance of the simulation
nor its variability. In addition, we extended our
implementation of Damaris to also support the use of
dedicated nodes and conducted a thorough comparison of
the two approaches-dedicated cores and dedicated
nodes-for I/O tasks with the aforementioned
applications.",
acknowledgement = ack-nhfb,
articleno = "15",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Gao:2016:AOM,
author = "Jiaquan Gao and Yu Wang and Jun Wang and Ronghua
Liang",
title = "Adaptive Optimization Modeling of Preconditioned
Conjugate Gradient on Multi-{GPUs}",
journal = j-TOPC,
volume = "3",
number = "3",
pages = "16:1--16:??",
month = dec,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/2990849",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Dec 26 17:40:41 MST 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The preconditioned conjugate gradient (PCG) algorithm
is a well-known iterative method for solving sparse
linear systems in scientific computations.
GPU-accelerated PCG algorithms for large-sized problems
have attracted considerable attention recently.
However, on a specific multi-GPU platform, producing a
highly parallel PCG implementation for any large-sized
problem requires significant time because several
manual steps are involved in adjusting the related
parameters and selecting an appropriate storage format
for the matrix block that is assigned to each GPU. This
motivates us to propose adaptive optimization modeling
of PCG on multi-GPUs, which mainly involves the
following parts: (1) an optimization multi-GPU parallel
framework of PCG and (2) the profile-based optimization
modeling for each one of the main components of the PCG
algorithm, including vector operation, inner product,
and sparse matrix-vector multiplication (SpMV). Our
model does not construct a new storage format or kernel
but automatically and rapidly generates an optimal
parallel PCG algorithm for any problem on a specific
multi-GPU platform by integrating existing storage
formats and kernels. We take a vector operation kernel,
an inner-product kernel, and five popular SpMV kernels
for an example to present the idea of constructing the
model. Given that our model is general, independent of
the problems, and dependent on the resources of
devices, this model is constructed only once for each
type of GPU. The experiments validate the high
efficiency of our proposed model.",
acknowledgement = ack-nhfb,
articleno = "16",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Creech:2016:TSS,
author = "Timothy Creech and Rajeev Barua",
title = "Transparently Space Sharing a Multicore Among Multiple
Processes",
journal = j-TOPC,
volume = "3",
number = "3",
pages = "17:1--17:??",
month = dec,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/3001910",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Dec 26 17:40:41 MST 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "As hardware becomes increasingly parallel and the
availability of scalable parallel software improves,
the problem of managing multiple multithreaded
applications (processes) becomes important. Malleable
processes, which can vary the number of threads used as
they run, enable sophisticated and flexible resource
management. Although many existing applications
parallelized for SMPs with parallel runtimes are in
fact already malleable, deployed runtime environments
provide no interface nor any strategy for intelligently
allocating hardware threads or even preventing
oversubscription. Prior research methods either depend
on profiling applications ahead of time to make good
decisions about allocations or do not account for
process efficiency at all, leading to poor performance.
None of these prior methods have been adapted widely in
practice. This article presents the Scheduling and
Allocation with Feedback (SCAF) system: a drop-in
runtime solution that supports existing malleable
applications in making intelligent allocation decisions
based on observed efficiency without any changes to
semantics, program modification, offline profiling, or
even recompilation. Our existing implementation can
control most unmodified OpenMP applications. Other
malleable threading libraries can also easily be
supported with small modifications without requiring
application modification or recompilation. In this
work, we present the SCAF daemon and a SCAF-aware port
of the GNU OpenMP runtime. We present a new technique
for estimating process efficiency purely at runtime
using available hardware counters and demonstrate its
effectiveness in aiding allocation decisions. We
evaluated SCAF using NAS NPB parallel benchmarks on
five commodity parallel platforms, enumerating
architectural features and their effects on our scheme.
We measured the benefit of SCAF in terms of sum of
speedups improvement (a common metric for
multiprogrammed environments) when running all
benchmark pairs concurrently compared to
equipartitioning-the best existing competing scheme in
the literature. We found that SCAF improves on
equipartitioning on four out of five machines, showing
a mean improvement factor in sum of speedups of 1.04 to
1.11x for benchmark pairs, depending on the machine,
and 1.09x on average. Since we are not aware of any
widely available tool for equipartitioning, we also
compare SCAF against multiprogramming using unmodified
OpenMP, which is the only environment available to end
users today. SCAF improves on the unmodified OpenMP
runtimes for all five machines, with a mean improvement
of 1.08 to 2.07x, depending on the machine, and 1.59x
on average.",
acknowledgement = ack-nhfb,
articleno = "17",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ballard:2016:HPS,
author = "Grey Ballard and Alex Druinsky and Nicholas Knight and
Oded Schwartz",
title = "Hypergraph Partitioning for Sparse Matrix--Matrix
Multiplication",
journal = j-TOPC,
volume = "3",
number = "3",
pages = "18:1--18:??",
month = dec,
year = "2016",
CODEN = "????",
DOI = "https://doi.org/10.1145/3015144",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Dec 26 17:40:41 MST 2016",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We propose a fine-grained hypergraph model for sparse
matrix-matrix multiplication (SpGEMM), a key
computational kernel in scientific computing and data
analysis whose performance is often communication
bound. This model correctly describes both the
interprocessor communication volume along a critical
path in a parallel computation and also the volume of
data moving through the memory hierarchy in a
sequential computation. We show that identifying a
communication-optimal algorithm for particular input
matrices is equivalent to solving a hypergraph
partitioning problem. Our approach is nonzero structure
dependent, meaning that we seek the best algorithm for
the given input matrices. In addition to our
three-dimensional fine-grained model, we also propose
coarse-grained one-dimensional and two-dimensional
models that correspond to simpler SpGEMM algorithms. We
explore the relations between our models theoretically,
and we study their performance experimentally in the
context of three applications that use SpGEMM as a key
computation. For each application, we find that at
least one coarse-grained model is as communication
efficient as the fine-grained model. We also observe
that different applications have affinities for
different algorithms. Our results demonstrate that
hypergraphs are an accurate model for reasoning about
the communication costs of SpGEMM as well as a
practical tool for exploring the SpGEMM algorithm
design space.",
acknowledgement = ack-nhfb,
articleno = "18",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Grove:2017:ISS,
author = "David Grove",
title = "Introduction to the Special Section on {PPoPP'15}",
journal = j-TOPC,
volume = "3",
number = "4",
pages = "19:1--19:??",
month = mar,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3040224",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 25 07:55:06 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "19",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Majo:2017:LPC,
author = "Zoltan Majo and Thomas R. Gross",
title = "A Library for Portable and Composable Data Locality
Optimizations for {NUMA} Systems",
journal = j-TOPC,
volume = "3",
number = "4",
pages = "20:1--20:??",
month = mar,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3040222",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 25 07:55:06 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Many recent multiprocessor systems are realized with a
nonuniform memory architecture (NUMA) and accesses to
remote memory locations take more time than local
memory accesses. Optimizing NUMA memory system
performance is difficult and costly for three principal
reasons: (1) Today's programming languages/libraries
have no explicit support for NUMA systems, (2) NUMA
optimizations are not portable, and (3) optimizations
are not composable (i.e., they can become ineffective
or worsen performance in environments that support
composable parallel software). This article presents
TBB-NUMA, a parallel programming library based on Intel
Threading Building Blocks (TBB) that supports portable
and composable NUMA-aware programming. TBB-NUMA
provides a model of task affinity that captures a
programmer's insights on mapping tasks to resources.
NUMA-awareness affects all layers of the library (i.e.,
resource management, task scheduling, and high-level
parallel algorithm templates) and requires close
coupling between all these layers. Optimizations
implemented with TBB-NUMA (for a set of standard
benchmark programs) result in up to 44\% performance
improvement over standard TBB. But more important,
optimized programs are portable across different NUMA
architectures and preserve data locality also when
composed with other parallel computations sharing the
same resource management layer.",
acknowledgement = ack-nhfb,
articleno = "20",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Golan-Gueta:2017:ASA,
author = "Guy Golan-Gueta and G. Ramalingam and Mooly Sagiv and
Eran Yahav",
title = "Automatic Scalable Atomicity via Semantic Locking",
journal = j-TOPC,
volume = "3",
number = "4",
pages = "21:1--21:??",
month = mar,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3040223",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 25 07:55:06 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "In this article, we consider concurrent programs in
which the shared state consists of instances of
linearizable abstract data types (ADTs). We present an
automated approach to concurrency control that
addresses a common need: the need to atomically execute
a code fragment, which may contain multiple ADT
operations on multiple ADT instances. We present a
synthesis algorithm that automatically enforces
atomicity of given code fragments (in a client program)
by inserting pessimistic synchronization that
guarantees atomicity and deadlock-freedom (without
using any rollback mechanism). Our algorithm takes a
commutativity specification as an extra input. This
specification indicates for every pair of ADT
operations the conditions under which the operations
commute. Our algorithm enables greater parallelism by
permitting commuting operations to execute
concurrently. We have implemented the synthesis
algorithm in a Java compiler and applied it to several
Java programs. Our results show that our approach
produces efficient and scalable synchronization.",
acknowledgement = ack-nhfb,
articleno = "21",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Izraelevitz:2017:GSN,
author = "Joseph Izraelevitz and Michael L. Scott",
title = "Generality and Speed in Nonblocking Dual Containers",
journal = j-TOPC,
volume = "3",
number = "4",
pages = "22:1--22:??",
month = mar,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3040220",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 25 07:55:06 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Nonblocking dual data structures extend traditional
notions of nonblocking progress to accommodate partial
methods, both by bounding the number of steps that a
thread can execute after its preconditions have been
satisfied and by ensuring that a waiting thread
performs no remote memory accesses that could interfere
with the execution of other threads. A nonblocking dual
container, in particular, is designed to hold either
data or requests. An insert operation either adds data
to the container or removes and satisfies a request; a
remove operation either takes data out of the container
or inserts a request. We present the first
general-purpose construction for nonblocking dual
containers, allowing any nonblocking container for data
to be paired with almost any nonblocking container for
requests. We also present new custom algorithms, based
on the LCRQ of Morrison and Afek, that outperform the
fastest previously known dual containers by factors of
four to six.",
acknowledgement = ack-nhfb,
articleno = "22",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Cole:2017:ROS,
author = "Richard Cole and Vijaya Ramachandran",
title = "Resource Oblivious Sorting on Multicores",
journal = j-TOPC,
volume = "3",
number = "4",
pages = "23:1--23:??",
month = mar,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3040221",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Sat Mar 25 07:55:06 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present a deterministic sorting algorithm, Sample,
Partition, and Merge Sort (SPMS), that interleaves the
partitioning of a sample sort with merging.
Sequentially, it sorts $n$ elements in $ O(n \log n)$
time cache-obliviously with an optimal number of cache
misses. The parallel complexity (or critical path
length) of the algorithm is $ O(\log n \log \log n)$,
which improves on previous bounds for deterministic
sample sort. The algorithm also has low false sharing
costs. When scheduled by a work-stealing scheduler in a
multicore computing environment with a global shared
memory and p cores, each having a cache of size $M$
organized in blocks of size $B$, the costs of the
additional cache misses and false sharing misses due to
this parallel execution are bounded by the cost of $
O(S \cdot M / B)$ and $ O(S \cdot B)$ cache misses,
respectively, where $S$ is the number of steals
performed during the execution. Finally, SPMS is
resource oblivious in that the dependence on machine
parameters appear only in the analysis of its
performance and not within the algorithm itself.",
acknowledgement = ack-nhfb,
articleno = "23",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ballard:2017:GEIa,
author = "Grey Ballard and Mary Hall and Tim Harris and Brandon
Lucia",
title = "{Guest Editor} Introduction {PPoPP 2016}, Special
Issue 2 of 2",
journal = j-TOPC,
volume = "4",
number = "1",
pages = "1:1--1:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3108141",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "1",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ashkiani:2017:GME,
author = "Saman Ashkiani and Andrew Davidson and Ulrich Meyer
and John D. Owens",
title = "{GPU Multisplit}: an Extended Study of a Parallel
Algorithm",
journal = j-TOPC,
volume = "4",
number = "1",
pages = "2:1--2:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3108139",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Multisplit is a broadly useful parallel primitive that
permutes its input data into contiguous buckets or
bins, where the function that categorizes an element
into a bucket is provided by the programmer. Due to the
lack of an efficient multisplit on Graphics Processing
Units (GPUs), programmers often choose to implement
multisplit with a sort. One way is to first generate an
auxiliary array of bucket IDs and then sort input data
based on it. In case smaller indexed buckets possess
smaller valued keys, another way for multisplit is to
directly sort input data. Both methods are inefficient
and require more work than necessary: The former
requires more expensive data movements while the latter
spends unnecessary effort in sorting elements within
each bucket. In this work, we provide a parallel model
and multiple implementations for the multisplit
problem. Our principal focus is multisplit for a small
(up to 256) number of buckets. We use warp-synchronous
programming models and emphasize warpwide
communications to avoid branch divergence and reduce
memory usage. We also hierarchically reorder input
elements to achieve better coalescing of global memory
accesses. On a GeForce GTX 1080 GPU, we can reach a
peak throughput of 18.93Gkeys/s (or 11.68Gpairs/s) for
a key-only (or key-value) multisplit. Finally, we
demonstrate how multisplit can be used as a building
block for radix sort. In our multisplit-based sort
implementation, we achieve comparable performance to
the fastest GPU sort routines, sorting 32-bit keys (and
key-value pairs) with a throughput of 3.0Gkeys/s (and
2.1Gpair/s).",
acknowledgement = ack-nhfb,
articleno = "2",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Wang:2017:GGG,
author = "Yangzihao Wang and Yuechao Pan and Andrew Davidson and
Yuduo Wu and Carl Yang and Leyuan Wang and Muhammad
Osama and Chenshan Yuan and Weitang Liu and Andy T.
Riffel and John D. Owens",
title = "{Gunrock}: {GPU} Graph Analytics",
journal = j-TOPC,
volume = "4",
number = "1",
pages = "3:1--3:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3108140",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "For large-scale graph analytics on the GPU, the
irregularity of data access and control flow, and the
complexity of programming GPUs, have presented two
significant challenges to developing a programmable
high-performance graph library. ``Gunrock,'' our
graph-processing system designed specifically for the
GPU, uses a high-level, bulk-synchronous, data-centric
abstraction focused on operations on a vertex or edge
frontier. Gunrock achieves a balance between
performance and expressiveness by coupling
high-performance GPU computing primitives and
optimization strategies with a high-level programming
model that allows programmers to quickly develop new
graph primitives with small code size and minimal GPU
programming knowledge. We characterize the performance
of various optimization strategies and evaluate
Gunrock's overall performance on different GPU
architectures on a wide range of graph primitives that
span from traversal-based algorithms and ranking
algorithms, to triangle counting and
bipartite-graph-based algorithms. The results show that
on a single GPU, Gunrock has on average at least an
order of magnitude speedup over Boost and PowerGraph,
comparable performance to the fastest GPU hardwired
primitives and CPU shared-memory graph libraries, such
as Ligra and Galois, and better performance than any
other GPU high-level graph library.",
acknowledgement = ack-nhfb,
articleno = "3",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Chowdhury:2017:AAD,
author = "Rezaul Chowdhury and Pramod Ganapathi and Stephen
Tschudi and Jesmin Jahan Tithi and Charles Bachmeier
and Charles E. Leiserson and Armando Solar-Lezama and
Bradley C. Kuszmaul and Yuan Tang",
title = "{Autogen}: Automatic Discovery of Efficient Recursive
Divide-\&-Conquer Algorithms for Solving Dynamic
Programming Problems",
journal = j-TOPC,
volume = "4",
number = "1",
pages = "4:1--4:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3125632",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present Autogen --- an algorithm that for a wide
class of dynamic programming (DP) problems
automatically discovers highly efficient
cache-oblivious parallel recursive divide-and-conquer
algorithms from inefficient iterative descriptions of
DP recurrences. Autogen analyzes the set of DP table
locations accessed by the iterative algorithm when run
on a DP table of small size and automatically
identifies a recursive access pattern and a
corresponding provably correct recursive algorithm for
solving the DP recurrence. We use Autogen to
autodiscover efficient algorithms for several
well-known problems. Our experimental results show that
several autodiscovered algorithms significantly
outperform parallel looping and tiled loop-based
algorithms. Also, these algorithms are less sensitive
to fluctuations of memory and bandwidth compared with
their looping counterparts, and their running times and
energy profiles remain relatively more stable. To the
best of our knowledge, Autogen is the first algorithm
that can automatically discover new nontrivial
divide-and-conquer algorithms.",
acknowledgement = ack-nhfb,
articleno = "4",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Steele:2017:AAC,
author = "Guy L. {Steele Jr.} and Jean-Baptiste Tristan",
title = "Adding Approximate Counters",
journal = j-TOPC,
volume = "4",
number = "1",
pages = "5:1--5:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3132167",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We describe a general framework for adding the values
of two approximate counters to produce a new
approximate counter value whose expected estimated
value is equal to the sum of the expected estimated
values of the given approximate counters. (To the best
of our knowledge, this is the first published
description of any algorithm for adding two approximate
counters.) We then work out implementation details for
five different kinds of approximate counter and provide
optimized pseudocode. For three of them, we present
proofs that the variance of a counter value produced by
adding two counter values in this way is bounded, and
in fact is no worse, or not much worse, than the
variance of the value of a single counter to which the
same total number of increment operations have been
applied. Addition of approximate counters is useful in
massively parallel divide-and-conquer algorithms that
use a distributed representation for large arrays of
counters. We describe two machine-learning algorithms
for topic modeling that use millions of integer
counters and confirm that replacing the integer
counters with approximate counters is effective,
speeding up a GPU-based implementation by over 65\% and
a CPU-based implementation by nearly 50\%, as well as
reducing memory requirements, without degrading their
statistical effectiveness.",
acknowledgement = ack-nhfb,
articleno = "5",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ballard:2017:GEIb,
author = "Grey Ballard and Mary Hall and Tim Harris and Brandon
Lucia",
title = "{Guest Editor} Introduction {PPoPP 2016}, Special
Issue 2 of 2",
journal = j-TOPC,
volume = "4",
number = "2",
pages = "6:1--6:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3108142",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "6",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kalikar:2017:DNM,
author = "Saurabh Kalikar and Rupesh Nasre",
title = "{DomLock}: a New Multi-Granularity Locking Technique
for Hierarchies",
journal = j-TOPC,
volume = "4",
number = "2",
pages = "7:1--7:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3127584",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We present efficient locking mechanisms for
hierarchical data structures. Several applications work
on an abstract hierarchy of objects, and a parallel
execution on this hierarchy necessitates
synchronization across workers operating on different
parts of the hierarchy. Existing synchronization
mechanisms are too coarse, too inefficient, or too ad
hoc, resulting in reduced or unpredictable amount of
concurrency. We propose a new locking approach based on
the structural properties of the underlying hierarchy.
We show that the developed techniques are efficient
even when the hierarchy is an arbitrary graph.
Theoretically, we present our approach as a
locking-cost-minimizing instance of a generic algebraic
model of synchronization for hierarchies. Using
STMBench7, we illustrate considerable reduction in the
locking cost, resulting in an average throughput
improvement of 42\%.",
acknowledgement = ack-nhfb,
articleno = "7",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Haider:2017:LRA,
author = "Syed Kamran Haider and William Hasenplaugh and Dan
Alistarh",
title = "{Lease\slash Release}: Architectural Support for
Scaling Contended Data Structures",
journal = j-TOPC,
volume = "4",
number = "2",
pages = "8:1--8:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3132168",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "High memory contention is generally agreed to be a
worst-case scenario for concurrent data structures.
There has been a significant amount of research effort
spent investigating designs that minimize contention,
and several programming techniques have been proposed
to mitigate its effects. However, there are currently
few architectural mechanisms to allow scaling contended
data structures at high thread counts. In this article,
we investigate hardware support for scalable contended
data structures. We propose Lease/Release, a simple
addition to standard directory-based MESI cache
coherence protocols, allowing participants to lease
memory, at the granularity of cache lines, by delaying
coherence messages for a short, bounded period of time.
Our analysis shows that Lease/Release can significantly
reduce the overheads of contention for both
non-blocking (lock-free) and lock-based data structure
implementations while ensuring that no deadlocks are
introduced. We validate Lease/Release empirically on
the Graphite multiprocessor simulator on a range of
data structures, including queue, stack, and priority
queue implementations, as well as on transactional
applications. Results show that Lease/Release
consistently improves both throughput and energy usage,
by up to 5x, both for lock-free and lock-based data
structure designs.",
acknowledgement = ack-nhfb,
articleno = "8",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Cao:2017:HRD,
author = "Man Cao and Minjia Zhang and Aritra Sengupta and
Swarnendu Biswas and Michael D. Bond",
title = "Hybridizing and Relaxing Dependence Tracking for
Efficient Parallel Runtime Support",
journal = j-TOPC,
volume = "4",
number = "2",
pages = "9:1--9:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3108138",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "It is notoriously challenging to develop parallel
software systems that are both scalable and correct.
Runtime support for parallelism-such as multithreaded
record and replay, data race detectors, transactional
memory, and enforcement of stronger memory models-helps
achieve these goals, but existing commodity solutions
slow programs substantially to track (i.e., detect or
control) an execution's cross-thread dependencies
accurately. Prior work tracks cross-thread dependencies
either ``pessimistically,'' slowing every program
access, or ``optimistically,'' allowing for lightweight
instrumentation of most accesses but dramatically
slowing accesses that are conflicting (i.e., involved
in cross-thread dependencies). This article presents
two novel approaches that seek to improve the
performance of dependence tracking. Hybrid tracking
(HT) hybridizes pessimistic and optimistic tracking by
overcoming a fundamental mismatch between these two
kinds of tracking. HT uses an adaptive, profile-based
policy to make runtime decisions about switching
between pessimistic and optimistic tracking. Relaxed
tracking (RT) attempts to reduce optimistic tracking's
overhead on conflicting accesses by tracking
dependencies in a ``relaxed'' way-meaning that not all
dependencies are tracked accurately-while still
preserving both program semantics and runtime support's
correctness. To demonstrate the usefulness and
potential of HT and RT, we build runtime support based
on the two approaches. Our evaluation shows that both
approaches offer performance advantages over existing
approaches, but there exist challenges and
opportunities for further improvement. HT and RT are
distinct solutions to the same problem. It is easier to
build runtime support based on HT than on RT, although
RT does not incur the overhead of online profiling.
This article presents the two approaches together to
inform and inspire future designs for efficient
parallel runtime support.",
acknowledgement = ack-nhfb,
articleno = "9",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Chatzopoulos:2017:EES,
author = "Georgios Chatzopoulos and Aleksandar Dragojevi{\'c}
and Rachid Guerraoui",
title = "{ESTIMA}: Extrapolating {ScalabiliTy} of In-Memory
Applications",
journal = j-TOPC,
volume = "4",
number = "2",
pages = "10:1--10:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3108137",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "This article presents estima, an easy-to-use tool for
extrapolating the scalability of in-memory
applications. estima is designed to perform a simple
yet important task: Given the performance of an
application on a small machine with a handful of cores,
estima extrapolates its scalability to a larger machine
with more cores, while requiring minimum input from the
user. The key idea underlying estima is the use of
stalled cycles (e.g., cycles that the processor spends
waiting for missed cache line fetches or busy locks).
estima measures stalled cycles on a few cores and
extrapolates them to more cores, estimating the amount
of waiting in the system. estima can be effectively
used to predict the scalability of in-memory
applications for bigger execution machines. For
instance, using measurements of memcached and SQLite on
a desktop machine, we obtain accurate predictions of
their scalability on a server. Our extensive evaluation
shows the effectiveness of estima on a large number of
in-memory benchmarks.",
acknowledgement = ack-nhfb,
articleno = "10",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Gulisano:2017:EDS,
author = "Vincenzo Gulisano and Yiannis Nikolakopoulos and
Daniel Cederman and Marina Papatriantafilou and
Philippas Tsigas",
title = "Efficient Data Streaming Multiway Aggregation through
Concurrent Algorithmic Designs and New Abstract Data
Types",
journal = j-TOPC,
volume = "4",
number = "2",
pages = "11:1--11:??",
month = oct,
year = "2017",
CODEN = "????",
DOI = "https://doi.org/10.1145/3131272",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Tue Oct 10 17:42:07 MDT 2017",
bibsource = "http://topc.acm.org/;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Data streaming relies on continuous queries to process
unbounded streams of data in a real-time fashion. It is
commonly demanding in computation capacity, given that
the relevant applications involve very large volumes of
data. Data structures act as articulation points and
maintain the state of data streaming operators,
potentially supporting high parallelism and balancing
the work among them. Prompted by this fact, in this
work we study and analyze parallelization needs of
these articulation points, focusing on the problem of
streaming multiway aggregation, where large data
volumes are received from multiple input streams. The
analysis of the parallelization needs, as well as of
the use and limitations of existing aggregate designs
and their data structures, leads us to identify needs
for appropriate shared objects that can achieve
low-latency and high-throughput multiway aggregation.
We present the requirements of such objects as abstract
data types and we provide efficient lock-free
linearizable algorithmic implementations of them, along
with new multiway aggregate algorithmic designs that
leverage them, supporting both deterministic
order-sensitive and order-insensitive aggregate
functions. Furthermore, we point out future directions
that open through these contributions. The article
includes an extensive experimental study, based on a
variety of continuous aggregation queries on two large
datasets extracted from SoundCloud, a music social
network, and from a Smart Grid network. In all the
experiments, the proposed data structures and the
enhanced aggregate operators improved the processing
performance significantly, up to one order of
magnitude, in terms of both throughput and latency,
over the commonly used techniques based on queues.",
acknowledgement = ack-nhfb,
articleno = "11",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Malas:2018:MIP,
author = "Tareq M. Malas and Georg Hager and Hatem Ltaief and
David E. Keyes",
title = "Multidimensional Intratile Parallelization for
Memory-Starved Stencil Computations",
journal = j-TOPC,
volume = "4",
number = "3",
pages = "12:1--12:??",
month = apr,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3155290",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Optimizing the performance of stencil algorithms has
been the subject of intense research over the last two
decades. Since many stencil schemes have low arithmetic
intensity, most optimizations focus on increasing the
temporal data access locality, thus reducing the data
traffic through the main memory interface with the
ultimate goal of decoupling from this bottleneck. There
are, however, only a few approaches that explicitly
leverage the shared cache feature of modern multicore
chips. If every thread works on its private, separate
cache block, the available cache space can become too
small, and sufficient temporal locality may not be
achieved. We propose a flexible multidimensional
intratile parallelization method for stencil algorithms
on multicore CPUs with a shared outer-level cache. This
method leads to a significant reduction in the required
cache space without adverse effects from hardware
prefetching or TLB shortage. Our Girih framework
includes an autotuner to select optimal parameter
configurations on the target hardware. We conduct
performance experiments on two contemporary Intel
processors and compare with the state-of-the-art
stencil frameworks Pluto and Pochoir, using four
corner-case stencil schemes and a wide range of problem
sizes. Girih shows substantial performance advantages
and best arithmetic intensity at almost all problem
sizes, especially on low-intensity stencils with
variable coefficients. We study in detail the
performance behavior at varying grid sizes using
phenomenological performance modeling. Our analysis of
energy consumption reveals that our method can save
energy through reduced DRAM bandwidth usage even at a
marginal performance gain. It is thus well suited for
future architectures that will be strongly challenged
by the cost of data movement, be it in terms of
performance or energy consumption.",
acknowledgement = ack-nhfb,
articleno = "12",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Akbudak:2018:PMS,
author = "Kadir Akbudak and Oguz Selvitopi and Cevdet Aykanat",
title = "Partitioning Models for Scaling Parallel Sparse
Matrix--Matrix Multiplication",
journal = j-TOPC,
volume = "4",
number = "3",
pages = "13:1--13:??",
month = apr,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3155292",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We investigate outer-product--parallel,
inner-product--parallel, and
row-by-row-product--parallel formulations of sparse
matrix-matrix multiplication (SpGEMM) on distributed
memory architectures. For each of these three
formulations, we propose a hypergraph model and a
bipartite graph model for distributing SpGEMM
computations based on one-dimensional (1D) partitioning
of input matrices. We also propose a communication
hypergraph model for each formulation for distributing
communication operations. The computational graph and
hypergraph models adopted in the first phase aim at
minimizing the total message volume and balancing the
computational loads of processors, whereas the
communication hypergraph models adopted in the second
phase aim at minimizing the total message count and
balancing the message volume loads of processors. That
is, the computational partitioning models reduce the
bandwidth cost and the communication hypergraph models
reduce the latency cost. Our extensive parallel
experiments on up to 2048 processors for a wide range
of realistic SpGEMM instances show that although the
outer-product--parallel formulation scales better, the
row-by-row-product--parallel formulation is more viable
due to its significantly lower partitioning overhead
and competitive scalability. For computational
partitioning models, our experimental findings indicate
that the proposed bipartite graph models are attractive
alternatives to their hypergraph counterparts because
of their lower partitioning overhead. Finally, we show
that by reducing the latency cost besides the bandwidth
cost through using the communication hypergraph models,
the parallel SpGEMM time can be further improved up to
32\%.",
acknowledgement = ack-nhfb,
articleno = "13",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bilardi:2018:LBT,
author = "Gianfranco Bilardi and Michele Scquizzato and
Francesco Silvestri",
title = "A Lower Bound Technique for Communication in {BSP}",
journal = j-TOPC,
volume = "4",
number = "3",
pages = "14:1--14:??",
month = apr,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3181776",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Communication is a major factor determining the
performance of algorithms on current computing systems;
it is therefore valuable to provide tight lower bounds
on the communication complexity of computations. This
article presents a lower bound technique for the
communication complexity in the bulk-synchronous
parallel (BSP) model of a given class of DAG
computations. The derived bound is expressed in terms
of the switching potential of a DAG, that is, the
number of permutations that the DAG can realize when
viewed as a switching network. The proposed technique
yields tight lower bounds for the fast Fourier
transform (FFT), and for any sorting and permutation
network. A stronger bound is also derived for the
periodic balanced sorting network, by applying this
technique to suitable subnetworks. Finally, we
demonstrate that the switching potential captures
communication requirements even in computational models
different from BSP, such as the I/O model and the
LPRAM.",
acknowledgement = ack-nhfb,
articleno = "14",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Sahin:2018:CSC,
author = "Semih Sahin and Bugra Gedik",
title = "{C-Stream}: a Co-routine-Based Elastic Stream
Processing Engine",
journal = j-TOPC,
volume = "4",
number = "3",
pages = "15:1--15:??",
month = apr,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3184120",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/multithreading.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Stream processing is a computational paradigm for
on-the-fly processing of live data. This paradigm lends
itself to implementations that can provide high
throughput and low latency by taking advantage of
various forms of parallelism that are naturally
captured by the stream processing model of computation,
such as pipeline, task, and data parallelism. In this
article, we describe the design and implementation of
C-Stream, which is an elastic stream processing engine.
C-Stream encompasses three unique properties. First, in
contrast to the widely adopted event-based interface
for developing streaming operators, C-Stream provides
an interface wherein each operator has its own driver
loop and relies on data availability application
programming interfaces (APIs) to decide when to perform
its computations. This self-control-based model
significantly simplifies the development of operators
that require multiport synchronization. Second,
C-Stream contains a dynamic scheduler that manages the
multithreaded execution of the operators. The
scheduler, which is customizable via plug-ins, enables
the execution of the operators as co-routines, using
any number of threads. The base scheduler implements
back-pressure, provides data availability APIs, and
manages preemption and termination handling. Last,
C-Stream varies the degree of parallelism to resolve
bottlenecks by both dynamically changing the number of
threads used to execute an application and adjusting
the number of replicas of data-parallel operators. We
provide an experimental evaluation of C-Stream. The
results show that C-Stream is scalable, highly
customizable, and can resolve bottlenecks by
dynamically adjusting the level of data parallelism
used.",
acknowledgement = ack-nhfb,
articleno = "15",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Agrawal:2018:ISI,
author = "Kunal Agrawal and I-Ting Angelina Lee and Michael
Spear",
title = "Introduction to Special Issue on {SPAA'15}",
journal = j-TOPC,
volume = "4",
number = "4",
pages = "16:1--16:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3226041",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "16",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ahn:2018:ADN,
author = "Kook Jin Ahn and Sudipto Guha",
title = "Access to Data and Number of Iterations: Dual Primal
Algorithms for Maximum Matching under Resource
Constraints",
journal = j-TOPC,
volume = "4",
number = "4",
pages = "17:1--17:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3154855",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "In this article, we consider graph algorithms in
models of computation where the space usage (random
accessible storage, in addition to the read-only input)
is sublinear in the number of edges $m$ and the access
to input is constrained. These questions arise in many
natural settings, and in particular in the analysis of
streaming algorithms, MapReduce or similar algorithms,
or message passing distributed computing that model
constrained parallelism with sublinear central
processing. We focus on weighted nonbipartite maximum
matching in this article. For any constant $ p > 1$, we
provide an iterative sampling-based algorithm for
computing a $ (1 - \epsilon)$-approximation of the
weighted nonbipartite maximum matching that uses $ O(p
/ \epsilon)$ rounds of sampling, and $ O(n^{1 + 1 /
p})$ space. The results extend to $b$-Matching with
small changes. This article combines adaptive sketching
literature and fast primal-dual algorithms based on
relaxed Dantzig--Wolfe decision procedures. Each round
of sampling is implemented through linear sketches and
can be executed in a single round of streaming or two
rounds of MapReduce. The article also proves that
nonstandard linear relaxations of a problem, in
particular penalty-based formulations, are helpful in
reducing the adaptive dependence of the iterations.",
acknowledgement = ack-nhfb,
articleno = "17",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Alistarh:2018:TAS,
author = "Dan Alistarh and William Leiserson and Alexander
Matveev and Nir Shavit",
title = "{ThreadScan}: Automatic and Scalable Memory
Reclamation",
journal = j-TOPC,
volume = "4",
number = "4",
pages = "18:1--18:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3201897",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The concurrent memory reclamation problem is that of
devising a way for a deallocating thread to verify that
no other concurrent threads hold references to a memory
block being deallocated. To date, in the absence of
automatic garbage collection, there is no satisfactory
solution to this problem; existing tracking methods
like hazard pointers, reference counters, or
epoch-based techniques like RCU are either
prohibitively expensive or require significant
programming expertise to the extent that implementing
them efficiently can be worthy of a publication. None
of the existing techniques are automatic or even
semi-automated. In this article, we take a new approach
to concurrent memory reclamation. Instead of manually
tracking access to memory locations as done in
techniques like hazard pointers, or restricting shared
accesses to specific epoch boundaries as in RCU, our
algorithm, called ThreadScan, leverages operating
system signaling to automatically detect which memory
locations are being accessed by concurrent threads.
Initial empirical evidence shows that ThreadScan scales
surprisingly well and requires negligible programming
effort beyond the standard use of Malloc and Free.",
acknowledgement = ack-nhfb,
articleno = "18",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Dimitrov:2018:RDT,
author = "Dimitar Dimitrov and Martin Vechev and Vivek Sarkar",
title = "Race Detection in Two Dimensions",
journal = j-TOPC,
volume = "4",
number = "4",
pages = "19:1--19:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3264618",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Dynamic race detection is a program analysis technique
for detecting errors caused by undesired interleavings
of concurrent tasks. A primary challenge when designing
efficient race detection algorithms is to achieve
manageable space requirements. State-of-the-art
algorithms for unstructured parallelism require $
\Theta (n) $ space per monitored memory location, where
n is the total number of tasks. This is a serious
drawback when analyzing programs with many tasks. In
contrast, algorithms for programs with a
series-parallel (SP) structure require only $ \Theta
(1) $ space. Unfortunately, it is currently not well
understood if there are classes of parallelism beyond
SP that can also benefit from and be analyzed with $
\Theta (1) $ space complexity. In this work, we show
that structures richer than SP graphs, namely, that of
two-dimensional (2D) lattices, can also be analyzed in
$ \Theta (1) $ space. Toward that (a) we extend
Tarjan's algorithm for finding lowest common ancestors
to handle 2D lattices; (b) from that extension we
derive a serial algorithm for race detection that can
analyze arbitrary task graphs with a 2D lattice
structure; (c) we present a restriction to fork-join
that admits precisely the 2D lattices as task graphs
(e.g., it can express pipeline parallelism). Our work
generalizes prior work on structured race detection and
aims to provide a deeper understanding of the interplay
between structured parallelism and program analysis
efficiency.",
acknowledgement = ack-nhfb,
articleno = "19",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Lee:2018:ERD,
author = "I-Ting Angelina Lee and Tao B. Schardl",
title = "Efficient Race Detection for Reducer Hyperobjects",
journal = j-TOPC,
volume = "4",
number = "4",
pages = "20:1--20:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3205914",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:25 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/multithreading.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "A multithreaded Cilk program that is ostensibly
deterministic may nevertheless behave
nondeterministically due to programming errors in the
code. For a Cilk program that uses reducers-a general
reduction mechanism supported in various Cilk
dialects-such programming errors are especially
challenging to debug, because the errors can expose the
nondeterminism in how the Cilk runtime system manages
reducers. We identify two unique types of races that
arise from incorrect use of reducers in a Cilk program,
and we present two algorithms to catch these races. The
first algorithm, called the Peer-Set algorithm, detects
view-read races, which occur when the program attempts
to retrieve a value out of a reducer when the read may
result in a nondeterministic value, such as before all
previously spawned subcomputations that might update
the reducer have necessarily returned. The second
algorithm, called the SP+ algorithm, detects
determinacy races-instances where a write to a memory
location occurs logically in parallel with another
access to that location-even when the raced-on memory
locations relate to reducers. Both algorithms are
provably correct, asymptotically efficient, and can be
implemented efficiently in practice. We have
implemented both algorithms in our prototype race
detector, Rader. When running Peer-Set, Rader incurs a
geometric-mean multiplicative overhead of 2.56 over
running the benchmark without instrumentation. When
running SP+, Rader incurs a geometric-mean
multiplicative overhead of 16.94.",
acknowledgement = ack-nhfb,
articleno = "20",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Gilbert:2018:ISI,
author = "Seth Gilbert",
title = "Introduction to the Special Issue for {SPAA 2016}",
journal = j-TOPC,
volume = "5",
number = "1",
pages = "1:1--1:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3230677",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "1",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Mitzenmacher:2018:BBC,
author = "Michael Mitzenmacher and Rajmohan Rajaraman and Scott
Roche",
title = "Better Bounds for Coalescing-Branching Random Walks",
journal = j-TOPC,
volume = "5",
number = "1",
pages = "2:1--2:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3209688",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Coalescing-branching random walks, or cobra walks for
short, are a natural variant of random walks on graphs
that can model the spread of disease through contacts
or the spread of information in networks. In a k -cobra
walk, at each timestep, a subset of the vertices are
active; each active vertex chooses k random neighbors
(sampled independently and uniformly with replacement)
that become active at the next step, and these are the
only active vertices at the next step. A natural
quantity to study for cobra walks is the cover time,
which corresponds to the expected time when all nodes
have become infected or received the disseminated
information. In this article, we extend previous
results for cobra walks in multiple ways. We show that
the cover time for the 2-cobra walk on $ [0, n]^d $ is
$ O(n) $ (where the order notation hides constant
factors that depend on $d$); previous work had shown
the cover time was $ O(n \cdot \polylog (n))$. We show
that the cover time for a 2-cobra walk on an $n$-vertex
$d$ regular graph with conductance $ \phi_G$ is $ O(d^4
\phis^{-2}_G \log^2 n)$, significantly generalizing a
previous result that held only for expander graphs with
sufficiently high expansion. And, finally, we show that
the cover time for a 2-cobra walk on a graph with n
vertices and m edges is always $ O(m n^{3 / 4} \log
n)$; this is the first result showing that the bound of
$ \Theta (n^3)$ for the worst-case cover time for
random walks can be beaten using 2-cobra walks.",
acknowledgement = ack-nhfb,
articleno = "2",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Liu:2018:RAN,
author = "Mingmou Liu and Xiaoyin Pan and Yitong Yin",
title = "Randomized Approximate Nearest Neighbor Search with
Limited Adaptivity",
journal = j-TOPC,
volume = "5",
number = "1",
pages = "3:1--3:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3209884",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We study the complexity of parallel data structures
for approximate nearest neighbor search in
$d$-dimensional Hamming space $ \{ 0, 1 \}^d$. A
classic model for static data structures is the
cell-probe model [27]. We consider a cell-probe model
with limited adaptivity, where given a $ k \geq 1$, a
query is resolved by making at most $k$ rounds of
parallel memory accesses to the data structure. We give
two randomized algorithms that solve the approximate
nearest neighbor search using $k$ rounds of parallel
memory accesses: --- a simple algorithm with $ O(k
(\log d)^{1 / k})$ total number of memory accesses for
all $ k \geq 1$ --- an algorithm with $ O(k + (1 / k
\log d)^{O(1 / k)})$ total number of memory accesses
for all sufficiently large $k$. Both algorithms use
data structures of polynomial size. We prove an $
\Omega (1 / k (\log d)^{1 / k})$ lower bound for the
total number of memory accesses for any randomized
algorithm solving the approximate nearest neighbor
search within $ k \leq \log \log d / 2 \log \log \log
d$ rounds of parallel memory accesses on any data
structures of polynomial size. This lower bound shows
that our first algorithm is asymptotically optimal when
$ k = O(1)$. And our second algorithm achieves the
asymptotically optimal tradeoff between number of
rounds and total number of memory accesses. In the
extremal case, when $ k = O(\log \log d / \log \log
\log d)$ is big enough, our second algorithm matches
the $ \Theta (\log \log d / \log \log \log d)$ tight
bound for fully adaptive algorithms for approximate
nearest neighbor search in [11].",
acknowledgement = ack-nhfb,
articleno = "3",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Pandurangan:2018:FDA,
author = "Gopal Pandurangan and Peter Robinson and Michele
Scquizzato",
title = "Fast Distributed Algorithms for Connectivity and {MST}
in Large Graphs",
journal = j-TOPC,
volume = "5",
number = "1",
pages = "4:1--4:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3209689",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Motivated by the increasing need to understand the
algorithmic foundations of distributed large-scale
graph computations, we study a number of fundamental
graph problems in a message-passing model for
distributed computing where $ k \geq 2 $ machines
jointly perform computations on graphs with $n$ nodes
(typically, $ n \gg k$). The input graph is assumed to
be initially randomly partitioned among the $k$
machines, a common implementation in many real-world
systems. Communication is point-to-point, and the goal
is to minimize the number of communication rounds of
the computation. Our main result is an (almost) optimal
distributed randomized algorithm for graph
connectivity. Our algorithm runs in $ {\tilde O}(n /
k^2)$ rounds ($ {\tilde O}$ notation hides a $ \polylog
(n)$ factor and an additive $ \polylog (n)$ term). This
improves over the best previously known bound of $
{\tilde O}(n / k)$ [Klauck et al., SODA 2015] and is
optimal (up to a polylogarithmic factor) in light of an
existing lower bound of $ \tilde \Omega (n / k^2)$. Our
improved algorithm uses a bunch of techniques,
including linear graph sketching, that prove useful in
the design of efficient distributed graph algorithms.
Using the connectivity algorithm as a building block,
we then present fast randomized algorithms for
computing minimum spanning trees, (approximate)
min-cuts, and for many graph verification problems. All
these algorithms take $ {\tilde O}(n / k^2)$ rounds and
are optimal up to polylogarithmic factors. We also show
an almost matching lower bound of $ \tilde \Omega (n /
k^2)$ rounds for many graph verification problems by
leveraging lower bounds in random-partition
communication complexity.",
acknowledgement = ack-nhfb,
articleno = "4",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Korupolu:2018:RPF,
author = "Madhukar Korupolu and Rajmohan Rajaraman",
title = "Robust and Probabilistic Failure-Aware Placement",
journal = j-TOPC,
volume = "5",
number = "1",
pages = "5:1--5:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3210367",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Motivated by the growing complexity and heterogeneity
of modern data centers, and the prevalence of commodity
component failures, this article studies the
failure-aware placement problem of placing tasks of a
parallel job on machines in the data center with the
goal of increasing availability. We consider two models
of failures: adversarial and probabilistic. In the
adversarial model, each node has a weight (higher
weight implying higher reliability) and the adversary
can remove any subset of nodes of total weight at most
a given bound W and our goal is to find a placement
that incurs the least disruption against such an
adversary. In the probabilistic model, each node has a
probability of failure and we need to find a placement
that maximizes the probability that at least K out of N
tasks survive at any time. For adversarial failures, we
first show that (i) the problems are in $ \Sigma_2 $,
the second level of the polynomial hierarchy; (ii) a
variant of the problem that we call RobustFap (for
Robust Failure-Aware Placement) is co-NP-hard; and
(iii) an all-or-nothing version of RobustFap is $
\Sigma_2$-complete. We then give a polynomial-time
approximation scheme (PTAS) for RobustFap, a key
ingredient of which is a solution that we design for a
fractional version of RobustFap. We then study
HierRobustFap, which is the fractional RobustFap
problem over a hierarchical network, in which failures
can occur at any subset of nodes in the hierarchy, and
a failure at a node can adversely impact all of its
descendants in the hierarchy. To solve HierRobustFap,
we introduce a notion of hierarchical max-min fairness
and a novel Generalized Spreading algorithm, which is
simultaneously optimal for every upper bound W on the
total weight of nodes that an adversary can fail. These
generalize the classical notion of max-min fairness to
work with nodes of differing capacities, differing
reliability weights, and hierarchical structures. Using
randomized rounding, we extend this to give an
algorithm for integral HierRobustFap. For the
probabilistic version, we first give an algorithm that
achieves an additive $ \epsilon $ approximation in the
failure probability for the single level version,
called ProbFap, while giving up a $ (1 + \epsilon)$
multiplicative factor in the number of failures. We
then extend the result to the hierarchical version,
HierProbFap, achieving an \epsilon additive
approximation in failure probability while giving up an
$ (L + \epsilon)$ multiplicative factor in the number
of failures, where $L$ is the number of levels in the
hierarchy.",
acknowledgement = ack-nhfb,
articleno = "5",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Zhang:2018:LFT,
author = "Deli Zhang and Pierre Laborde and Lance Lebanoff and
Damian Dechev",
title = "Lock-Free Transactional Transformation for Linked Data
Structures",
journal = j-TOPC,
volume = "5",
number = "1",
pages = "6:1--6:??",
month = sep,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3209690",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/hash.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Nonblocking data structures allow scalable and
thread-safe access to shared data. They provide
individual operations that appear to execute
atomically. However, it is often desirable to execute
multiple operations atomically in a transactional
manner. Previous solutions, such as Software
Transactional Memory (STM) and transactional boosting,
manage transaction synchronization separately from the
underlying data structure's thread synchronization.
Although this reduces programming effort, it leads to
overhead associated with additional synchronization and
the need to rollback aborted transactions. In this
work, we present a new methodology for transforming
high-performance lock-free linked data structures into
high-performance lock-free transactional linked data
structures without revamping the data structures'
original synchronization design. Our approach leverages
the semantic knowledge of the data structure to
eliminate the overhead of false conflicts and
rollbacks. We encapsulate all operations, operands, and
transaction status in a transaction descriptor, which
is shared among the nodes accessed by the same
transaction. We coordinate threads to help finish the
remaining operations of delayed transactions based on
their transaction descriptors. When a transaction
fails, we recover the correct abstract state by
reversely interpreting the logical status of a node. We
also present an obstruction-free version of our
algorithm that can be applied to dynamic execution
scenarios and an example of our approach applied to a
hash map. In our experimental evaluation using
transactions with randomly generated operations, our
lock-free transactional data structures outperform the
transactional boosted ones by 70\% on average. They
also outperform the alternative STM-based approaches by
a factor of 2 to 13 across all scenarios. More
importantly, we achieve 4,700 to 915,000 times fewer
spurious aborts than the alternatives.",
acknowledgement = ack-nhfb,
articleno = "6",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Muller:2018:NHP,
author = "Michel M{\"u}ller and Takayuki Aoki",
title = "New High Performance {GPGPU} Code Transformation
Framework Applied to Large Production Weather
Prediction Code",
journal = j-TOPC,
volume = "5",
number = "2",
pages = "7:1--7:??",
month = jan,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3291523",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/fortran3.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We introduce ``Hybrid Fortran,'' a new approach that
allows a high-performance GPGPU port for structured
grid Fortran codes. This technique only requires
minimal changes for a CPU targeted codebase, which is a
significant advancement in terms of productivity. It
has been successfully applied to both dynamical core
and physical processes of ASUCA, a Japanese mesoscale
weather prediction model with more than 150k lines of
code. By means of a minimal weather application that
resembles ASUCA's code structure, Hybrid Fortran is
compared to both a performance model as well as today's
commonly used method, OpenACC. As a result, the Hybrid
Fortran implementation is shown to deliver the same or
better performance than OpenACC, and its performance
agrees with the model both on CPU and GPU. In a
full-scale production run, using an ASUCA grid with
1581 $ \times $ 1301 $ \times $ 58 cells and real-world
weather data in 2km resolution, 24 NVIDIA Tesla P100
running the Hybrid Fortran-based GPU port are shown to
replace more than fifty 18-core Intel Xeon Broadwell
E5-2695 v4 running the reference implementation-an
achievement comparable to more invasive GPGPU rewrites
of other weather models.",
acknowledgement = ack-nhfb,
articleno = "7",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Burtscher:2018:HQF,
author = "Martin Burtscher and Sindhu Devale and Sahar Azimi and
Jayadharini Jaiganesh and Evan Powers",
title = "A High-Quality and Fast Maximal Independent Set
Implementation for {GPUs}",
journal = j-TOPC,
volume = "5",
number = "2",
pages = "8:1--8:??",
month = jan,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3291525",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Computing a maximal independent set is an important
step in many parallel graph algorithms. This article
introduces ECL-MIS, a maximal independent set
implementation that works well on GPUs. It includes key
optimizations to speed up computation, reduce the
memory footprint, and increase the set size. Its CUDA
implementation requires fewer than 30 kernel
statements, runs asynchronously, and produces a
deterministic result. It outperforms the maximal
independent set implementations of Pannotia, CUSP, and
IrGL on each of the 16 tested graphs of various types
and sizes. On a Titan X GPU, ECL-MIS is between 3.9 and
100 times faster (11.5 times, on average). ECL-MIS
running on the GPU is also faster than the parallel CPU
codes Ligra, Ligra+, and PBBS running on 20 Xeon cores,
which it outperforms by 4.1 times, on average. At the
same time, ECL-MIS produces maximal independent sets
that are up to 52\% larger (over 10\%, on average)
compared to these preexisting CPU and GPU
implementations. Whereas these codes produce maximal
independent sets that are, on average, about 15\%
smaller than the largest possible such sets, ECL-MIS
sets are less than 6\% smaller than the maximum
independent sets.",
acknowledgement = ack-nhfb,
articleno = "8",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Liu:2018:APR,
author = "Junhong Liu and Guangming Tan and Yulong Luo and
Jiajia Li and Zeyao Mo and Ninghui Sun",
title = "An Autotuning Protocol to Rapidly Build Autotuners",
journal = j-TOPC,
volume = "5",
number = "2",
pages = "9:1--9:??",
month = jan,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3291527",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Automatic performance tuning (Autotuning) is an
increasingly critical tuning technique for the high
portable performance of Exascale applications. However,
constructing an autotuner from scratch remains a
challenge, even for domain experts. In this work, we
propose a performance tuning and knowledge management
suite (PAK) to help rapidly build autotuners. In order
to accommodate existing autotuning techniques, we
present an autotuning protocol that is composed of an
extractor, producer, optimizer, evaluator, and learner.
To achieve modularity and reusability, we also define
programming interfaces for each protocol component as
the fundamental infrastructure, which provides a
customizable mechanism to deploy knowledge mining in
the performance database. PAK's usability is
demonstrated by studying two important computational
kernels: stencil computation and sparse matrix-vector
multiplication (SpMV). Our proposed autotuner based on
PAK shows comparable performance and higher
productivity than traditional autotuners by writing
just a few tens of code using our autotuning
protocol.",
acknowledgement = ack-nhfb,
articleno = "9",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Anta:2018:SDP,
author = "Antonio Fern{\'a}ndez Anta and Dariusz R. Kowalski and
Miguel A. Mosteiro and Prudence W. H. Wong",
title = "Scheduling Dynamic Parallel Workload of Mobile Devices
with Access Guarantees",
journal = j-TOPC,
volume = "5",
number = "2",
pages = "10:1--10:??",
month = jan,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3291529",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We study a dynamic resource-allocation problem that
arises in various parallel computing scenarios, such as
mobile cloud computing, cloud computing systems,
Internet of Things systems, and others. Generically, we
model the architecture as client mobile devices and
static base stations. Each client ``arrives'' to the
system to upload data to base stations by radio
transmissions and then ``leaves.'' The problem, called
Station Assignment, is to assign clients to stations so
that every client uploads their data under some
restrictions, including a target subset of stations, a
maximum delay between transmissions, a volume of data
to upload, and a maximum bandwidth for each station. We
study the solvability of Station Assignment under an
adversary that controls the arrival and departure of
clients, limited to maximum rate and burstiness of such
arrivals. We show upper and lower bounds on the rate
and burstiness for various client arrival schedules and
protocol classes. To the best of our knowledge, this is
the first time that Station Assignment is studied under
adversarial arrivals and departures.",
acknowledgement = ack-nhfb,
articleno = "10",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Mirhosseini:2018:BBA,
author = "Amirhossein Mirhosseini and Mohammad Sadrosadati and
Fatemeh Aghamohammadi and Mehdi Modarressi and Hamid
Sarbazi-Azad",
title = "{BARAN}: Bimodal Adaptive Reconfigurable-Allocator
Network-on-Chip",
journal = j-TOPC,
volume = "5",
number = "3",
pages = "11:1--11:??",
month = jan,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3294049",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Virtual channels are employed to improve the
throughput under high traffic loads in
Networks-on-Chips (NoCs). However, they can impose
non-negligible overheads on performance by prolonging
clock cycle time, especially under low traffic loads
where the impact of virtual channels on performance is
trivial. In this article, we propose a novel
architecture, called BARAN, that can either improve
on-chip network performance or reduce its power
consumption (depending on the specific implementation
chosen), not both at the same time, when virtual
channels are underutilized; that is, the average number
of virtual channel allocation requests per cycle is
lower than the number of total virtual channels. We
also introduce a reconfigurable arbitration logic
within the BARAN architecture that can be configured to
have multiple latencies and, hence, multiple slack
times. The increased slack times are then used to
reduce the supply voltage of the routers or increase
their clock frequency in order to reduce power
consumption or improve the performance of the whole NoC
system. The power-centric design of BARAN reduces NoC
power consumption by 43.4\% and 40.6\% under CMP and
GPU workloads, on average, respectively, compared to a
baseline architecture while imposing negligible area
and performance overheads. The performance-centric
design of BARAN reduces the average packet latency by
45.4\% and 42.1\%, on average, under CMP and GPU
workloads, respectively, compared to the baseline
architecture while increasing power consumption by
39.7\% and 43.7\%, on average. Moreover, the
performance-centric BARAN postpones the network
saturation rate by 11.5\% under uniform random traffic
compared to the baseline architecture.",
acknowledgement = ack-nhfb,
articleno = "11",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Amer:2018:LCM,
author = "Abdelhalim Amer and Huiwei Lu and Pavan Balaji and
Milind Chabbi and Yanjie Wei and Jeff Hammond and
Satoshi Matsuoka",
title = "Lock Contention Management in Multithreaded {MPI}",
journal = j-TOPC,
volume = "5",
number = "3",
pages = "12:1--12:??",
month = jan,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3275443",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/multithreading.bib;
http://www.math.utah.edu/pub/tex/bib/pvm.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3275443",
abstract = "In this article, we investigate contention management
in lock-based thread-safe MPI libraries. Specifically,
we make two assumptions: (1) locks are the only form of
synchronization when protecting communication paths;
and (2) contention occurs, and thus serialization is
unavoidable. Our work distinguishes between lock
acquisitions with respect to work being performed
inside a critical section; productive vs. unproductive.
Waiting for message reception without doing anything
else inside a critical section is an example of
unproductive lock acquisition. We show that the
high-throughput nature of modern scalable locking
protocols translates into better communication progress
for throughput-intensive MPI communication but
negatively impacts latency-sensitive communication
because of overzealous unproductive lock acquisition.
To reduce unproductive lock acquisitions, we devised a
method that promotes threads with productive work using
a generic two-level priority locking protocol. Our
results show that using a high-throughput protocol for
productive work and a fair protocol for less productive
code paths ensures the best tradeoff for fine-grained
communication, whereas a fair protocol is sufficient
for more coarse-grained communication. Although these
efforts have been rewarding, scalability degradation
remains significant. We discuss techniques that diverge
from the pure locking model and offer the potential to
further improve scalability.",
acknowledgement = ack-nhfb,
articleno = "12",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Chen:2018:PDG,
author = "Rong Chen and Jiaxin Shi and Yanzhe Chen and Binyu
Zang and Haibing Guan and Haibo Chen",
title = "{PowerLyra}: Differentiated Graph Computation and
Partitioning on Skewed Graphs",
journal = j-TOPC,
volume = "5",
number = "3",
pages = "13:1--13:??",
month = jan,
year = "2018",
CODEN = "????",
DOI = "https://doi.org/10.1145/3298989",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Jan 23 16:12:26 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Natural graphs with skewed distributions raise unique
challenges to distributed graph computation and
partitioning. Existing graph-parallel systems usually
use a ``one-size-fits-all'' design that uniformly
processes all vertices, which either suffer from
notable load imbalance and high contention for
high-degree vertices (e.g., Pregel and GraphLab) or
incur high communication cost and memory consumption
even for low-degree vertices (e.g., PowerGraph and
GraphX). In this article, we argue that skewed
distributions in natural graphs also necessitate
differentiated processing on high-degree and low-degree
vertices. We then introduce PowerLyra, a new
distributed graph processing system that embraces the
best of both worlds of existing graph-parallel systems.
Specifically, PowerLyra uses centralized computation
for low-degree vertices to avoid frequent
communications and distributes the computation for
high-degree vertices to balance workloads. PowerLyra
further provides an efficient hybrid graph partitioning
algorithm (i.e., hybrid-cut) that combines edge-cut
(for low-degree vertices) and vertex-cut (for
high-degree vertices) with heuristics. To improve cache
locality of inter-node graph accesses, PowerLyra
further provides a locality-conscious data layout
optimization. PowerLyra is implemented based on the
latest GraphLab and can seamlessly support various
graph algorithms running in both synchronous and
asynchronous execution modes. A detailed evaluation on
three clusters using various graph-analytics and MLDM
(Machine Learning and Data Mining) applications shows
that PowerLyra outperforms PowerGraph by up to 5.53X
(from 1.24X) and 3.26X (from 1.49X) for real-world and
synthetic graphs, respectively, and is much faster than
other systems like GraphX and Giraph, yet with much
less memory consumption. A porting of hybrid-cut to
GraphX further confirms the efficiency and generality
of PowerLyra.",
acknowledgement = ack-nhfb,
articleno = "13",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Aravind:2019:GME,
author = "Alex Aravind and Wim H. Hesselink",
title = "Group Mutual Exclusion by Fetch-and-increment",
journal = j-TOPC,
volume = "5",
number = "4",
pages = "14:1--14:??",
month = mar,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3309202",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Mar 11 18:54:51 MDT 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3309202",
abstract = "The group mutual exclusion (GME) problem (also called
the room synchronization problem) arises in various
practical applications that require concurrent data
sharing. Group mutual exclusion aims to achieve
exclusive access to a shared resource (a shared room)
while facilitating concurrency among non-conflicting
requests. The problem is that threads with distinct
interests are not allowed to access the shared resource
concurrently, but multiple threads with same interest
can. In Blelloch et al. (2003), the authors presented a
simple solution to the room synchronization problem
using fetch8add (F8A) and test-and-set (T8S) atomic
operations. This algorithm has $ O(m) $ remote memory
references (RMRs) in the cache coherent (CC) model,
where $m$ is the number of forums. In Bhatt and Huang
(2010), an open problem was posed: `` Is it possible to
design a GME algorithm with constant RMR for the CC
model using fetch8add instructions? '' This question is
partially answered in this article by presenting a
group mutual exclusion algorithm using
fetch-and-increment instructions. The algorithm is
simple and scalable.",
acknowledgement = ack-nhfb,
articleno = "14",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Behzad:2019:OPH,
author = "Babak Behzad and Surendra Byna and Prabhat and Marc
Snir",
title = "Optimizing {I/O} Performance of {HPC} Applications
with Autotuning",
journal = j-TOPC,
volume = "5",
number = "4",
pages = "15:1--15:??",
month = mar,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3309205",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Mar 11 18:54:51 MDT 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3309205",
abstract = "Parallel Input output is an essential component of
modern high-performance computing (HPC). Obtaining good
I/O performance for a broad range of applications on
diverse HPC platforms is a major challenge, in part,
because of complex inter dependencies between I/O
middleware and hardware. The parallel file system and
I/O middleware layers all offer optimization parameters
that can, in theory, result in better I/O performance.
Unfortunately, the right combination of parameters is
highly dependent on the application, HPC platform,
problem size, and concurrency. Scientific application
developers do not have the time or expertise to take on
the substantial burden of identifying good parameters
for each problem configuration. They resort to using
system defaults, a choice that frequently results in
poor I/O performance. We expect this problem to be
compounded on exascale-class machines, which will
likely have a deeper software stack with hierarchically
arranged hardware resources. We present as a solution
to this problem an autotuning system for optimizing I/O
performance, I/O performance modeling, I/O tuning, and
I/O patterns. We demonstrate the value of this
framework across several HPC platforms and applications
at scale.",
acknowledgement = ack-nhfb,
articleno = "15",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Maier:2019:CHT,
author = "Tobias Maier and Peter Sanders and Roman Dementiev",
title = "Concurrent Hash Tables: Fast and General(?)!",
journal = j-TOPC,
volume = "5",
number = "4",
pages = "16:1--16:??",
month = mar,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3309206",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Mar 11 18:54:51 MDT 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/hash.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3309206",
abstract = "Concurrent hash tables are one of the most important
concurrent data structures, which are used in numerous
applications. For some applications, it is common that
hash table accesses dominate the execution time. To
efficiently solve these problems in parallel, we need
implementations that achieve speedups in highly
concurrent scenarios. Unfortunately, currently
available concurrent hashing libraries are far away
from this requirement, in particular, when adaptively
sized tables are necessary or contention on some
elements occurs. Our starting point for better
performing data structures is a fast and simple
lock-free concurrent hash table based on linear probing
that is, however, limited to word-sized key-value types
and does not support dynamic size adaptation. We
explain how to lift these limitations in a provably
scalable way and demonstrate that dynamic growing has a
performance overhead comparable to the same
generalization in sequential hash tables. We perform
extensive experiments comparing the performance of our
implementations with six of the most widely used
concurrent hash tables. Ours are considerably faster
than the best algorithms with similar restrictions and
an order of magnitude faster than the best more general
tables. In some extreme cases, the difference even
approaches four orders of magnitude. All our
implementations discussed in this publication can be
found on github [17].",
acknowledgement = ack-nhfb,
articleno = "16",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Cruz:2019:ETM,
author = "Eduardo H. M. Cruz and Matthias Diener and La{\'e}rcio
L. Pilla and Philippe O. A. Navaux",
title = "{EagerMap}: a Task Mapping Algorithm to Improve
Communication and Load Balancing in Clusters of
Multicore Systems",
journal = j-TOPC,
volume = "5",
number = "4",
pages = "17:1--17:??",
month = mar,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3309711",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Mar 11 18:54:51 MDT 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3309711",
abstract = "Communication between tasks and load imbalance have
been identified as a major challenge for the
performance and energy efficiency of parallel
applications. A common way to improve communication is
to increase its locality, that is, to reduce the
distances of data transfers, prioritizing the usage of
faster and more efficient local interconnections over
remote ones. Regarding load imbalance, cores should
execute a similar amount of work. An important problem
to be solved in this context is how to determine an
optimized mapping of tasks to cluster nodes and cores
that increases the overall locality and load balancing.
In this article, we propose the EagerMap algorithm to
determine task mappings, which is based on a greedy
heuristic to match application communication patterns
to hardware hierarchies and which can also consider the
task load. Compared to previous algorithms, EagerMap is
faster, scales better, and supports more types of
computer systems, while maintaining the same or better
quality of the determined task mapping. EagerMap is
therefore an interesting choice for task mapping on a
variety of modern parallel architectures.",
acknowledgement = ack-nhfb,
articleno = "17",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bader:2019:EEC,
author = "David A. Bader",
title = "Editorial from the {Editor-in-Chief}",
journal = j-TOPC,
volume = "6",
number = "1",
pages = "1:1--1:??",
month = jun,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3325883",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:58 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3325883",
acknowledgement = ack-nhfb,
articleno = "1",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kronbichler:2019:MMF,
author = "Martin Kronbichler and Karl Ljungkvist",
title = "Multigrid for Matrix-Free High-Order Finite Element
Computations on Graphics Processors",
journal = j-TOPC,
volume = "6",
number = "1",
pages = "2:1--2:??",
month = jun,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3322813",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:58 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3322813",
abstract = "This article presents matrix-free finite-element
techniques for efficiently solving partial differential
equations on modern many-core processors, such as
graphics cards. We develop a GPU parallelization of a
matrix-free geometric multigrid iterative solver
targeting moderate and high polynomial degrees, with
support for general curved and adaptively refined
hexahedral meshes with hanging nodes. The central
algorithmic component is the matrix-free operator
evaluation with sum factorization. We compare the
node-level performance of our implementation running on
an Nvidia Pascal P100 GPU to a highly optimized
multicore implementation running on comparable Intel
Broadwell CPUs and an Intel Xeon Phi. Our experiments
show that the GPU implementation is approximately 1.5
to 2 times faster across four different scenarios of
the Poisson equation and a variety of element degrees
in 2D and 3D. The lowest time to solution per degree of
freedom is recorded for moderate polynomial degrees
between 3 and 5. A detailed performance analysis
highlights the capabilities of the GPU architecture and
the chosen execution model with threading within the
element, particularly with respect to the evaluation of
the matrix-vector product. Atomic intrinsics are shown
to provide a fast way for avoiding the possible race
conditions in summing the elemental residuals into the
global vector associated to shared vertices, edges, and
surfaces. In addition, the solver infrastructure allows
for using mixed-precision arithmetic that performs the
multigrid V-cycle in single precision with an outer
correction in double precision, increasing throughput
by up to 83",
acknowledgement = ack-nhfb,
articleno = "2",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bonifaci:2019:GPT,
author = "Vincenzo Bonifaci and Andreas Wiese and Sanjoy K.
Baruah and Alberto Marchetti-Spaccamela and Sebastian
Stiller and Leen Stougie",
title = "A Generalized Parallel Task Model for Recurrent
Real-Time Processes",
journal = j-TOPC,
volume = "6",
number = "1",
pages = "3:1--3:??",
month = jun,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3322809",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:58 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3322809",
abstract = "A model is considered for representing recurrent
precedence-constrained tasks that are to execute on
multiprocessor platforms. A recurrent task is specified
as a directed acyclic graph (DAG), a period, and a
relative deadline. Each vertex of the DAG represents a
sequential job, while the edges of the DAG represent
precedence constraints between these jobs. All the jobs
of the DAG are released simultaneously and need to
complete execution within the specified relative
deadline of their release. Each task may release jobs
in this manner an unbounded number of times, with
successive releases occurring at least the specified
period apart. Conditional control structures are also
allowed. The scheduling problem is to determine whether
a set of such recurrent tasks can be scheduled to
always meet all deadlines upon a specified number of
identical processors. This problem is shown to be
computationally intractable, but amenable to efficient
approximate solutions. Earliest Deadline First (EDF)
and Deadline Monotonic (DM) are shown to be good
approximate global scheduling algorithms. Polynomial
and pseudo-polynomial time schedulability tests, of
differing effectiveness, are presented for determining
whether a given task set can be scheduled by EDF or DM
to always meet deadlines on a specified number of
processors.",
acknowledgement = ack-nhfb,
articleno = "3",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ltaief:2019:MPP,
author = "Hatem Ltaief and Dalal Sukkari and Aniello Esposito
and Yuji Nakatsukasa and David Keyes",
title = "Massively Parallel Polar Decomposition on
Distributed-memory Systems",
journal = j-TOPC,
volume = "6",
number = "1",
pages = "4:1--4:??",
month = jun,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3328723",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:58 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3328723",
abstract = "We present a high-performance implementation of the
Polar Decomposition (PD) on distributed-memory systems.
Building upon on the QR-based Dynamically Weighted
Halley (QDWH) algorithm, the key idea lies in finding
the best rational approximation for the scalar sign
function, which also corresponds to the polar factor
for symmetric matrices, to further accelerate the QDWH
convergence. Based on the Zolotarev rational
functions-introduced by Zolotarev (ZOLO) in 1877-this
new PD algorithm ZOLO-PD converges within two
iterations even for ill-conditioned matrices, instead
of the original six iterations needed for QDWH. ZOLO-PD
uses the property of Zolotarev functions that
optimality is maintained when two functions are
composed in an appropriate manner. The resulting
ZOLO-PD has a convergence rate up to 17, in contrast to
the cubic convergence rate for QDWH. This comes at the
price of higher arithmetic costs and memory footprint.
These extra floating-point operations can, however, be
processed in an embarrassingly parallel fashion. We
demonstrate performance using up to 102,400 cores on
two supercomputers. We demonstrate that, in the
presence of a large number of processing units, ZOLO-PD
is able to outperform QDWH by up to 2.3$ \times $
speedup, especially in situations where QDWH runs out
of work, for instance, in the strong scaling mode of
operation.",
acknowledgement = ack-nhfb,
articleno = "4",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Saha:2019:OSA,
author = "Dibakar Saha and Koushik Sinha",
title = "Optimal Schedule for All-to-All Personalized
Communication in Multiprocessor Systems",
journal = j-TOPC,
volume = "6",
number = "1",
pages = "5:1--5:??",
month = jun,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3329867",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:58 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3329867",
abstract = "In this article, we address the problem of finding an
optimal schedule for all-to-all personalized message
communication among the processors in a multiprocessor
system where every processor has a unique message for
every other processor. When there are n processors and
\lfloor n /2 \rfloor parallel databus or channels for
message communications, there exist algorithms that
require O ( n$^2$ ) time for assigning the
databus/channels to the processor-pairs to obtain a
schedule with minimum number of time slots. However, in
recent massively parallel processing systems with a
huge number of processors, the number k of available
databus/channels is usually much smaller than \lfloor n
/2",
acknowledgement = ack-nhfb,
articleno = "5",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Pumma:2019:SDL,
author = "Sarunya Pumma and Min Si and Wu-Chun Feng and Pavan
Balaji",
title = "Scalable Deep Learning via {I/O} Analysis and
Optimization",
journal = j-TOPC,
volume = "6",
number = "2",
pages = "6:1--6:??",
month = sep,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3331526",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3331526",
abstract = "Scalable deep neural network training has been gaining
prominence because of the increasing importance of deep
learning in a multitude of scientific and commercial
domains. Consequently, a number of researchers have
investigated techniques to optimize deep learning
systems. Much of the prior work has focused on runtime
and algorithmic enhancements to optimize the
computation and communication. Despite these
enhancements, however, deep learning systems still
suffer from scalability limitations, particularly with
respect to data I/O. This situation is especially true
for training models where the computation can be
effectively parallelized, leaving I/O as the major
bottleneck. In fact, our analysis shows that I/O can
take up to 90",
acknowledgement = ack-nhfb,
articleno = "6",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Aupy:2019:SSP,
author = "Guilllaume Aupy and Ana Gainaru and Valentin {Le
F{\`e}vre}",
title = "{I/O} Scheduling Strategy for Periodic Applications",
journal = j-TOPC,
volume = "6",
number = "2",
pages = "7:1--7:??",
month = sep,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3338510",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3338510",
abstract = "With the ever-growing need of data in HPC
applications, the congestion at the I/O level becomes
critical in supercomputers. Architectural enhancement
such as burst buffers and pre-fetching are added to
machines but are not sufficient to prevent congestion.
Recent online I/O scheduling strategies have been put
in place, but they add an additional congestion point
and overheads in the computation of applications. In
this work, we show how to take advantage of the
periodic nature of HPC applications to develop
efficient periodic scheduling strategies for their I/O
transfers. Our strategy computes once during the job
scheduling phase a pattern that defines the I/O
behavior for each application, after which the
applications run independently, performing their I/O at
the specified times. Our strategy limits the amount of
congestion at the I/O node level and can be easily
integrated into current job schedulers. We validate
this model through extensive simulations and
experiments on an HPC cluster by comparing it to
state-of-the-art online solutions, showing that not
only does our scheduler have the advantage of being
de-centralized and thus overcoming the overhead of
online schedulers, but also that it performs better
than the other solutions, improving the application
dilation up to 16",
acknowledgement = ack-nhfb,
articleno = "7",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Kagaris:2019:SME,
author = "Dimitri Kagaris and Sourav Dutta",
title = "Scheduling Mutual Exclusion Accesses in Equal-Length
Jobs",
journal = j-TOPC,
volume = "6",
number = "2",
pages = "8:1--8:??",
month = sep,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3342562",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3342562",
abstract = "A fundamental problem in parallel and distributed
processing is the partial serialization that is imposed
due to the need for mutually exclusive access to common
resources. In this article, we investigate the problem
of optimally scheduling (in terms of makespan) a set of
jobs, where each job consists of the same number L of
unit-duration tasks, and each task either accesses
exclusively one resource from a given set of resources
or accesses a fully shareable resource. We develop and
establish the optimality of a fast polynomial-time
algorithm to find a schedule with the shortest makespan
for any number of jobs and for any number of resources
for the case of L = 2. In the notation commonly used
for job-shop scheduling problems, this result means
that the problem J | d$_{ij}$ =1, n$_j$ =2| C$_{max}$
is polynomially solvable, adding to the polynomial
solutions known for the problems J 2 | n$_j$ {$<$}= 2 |
C$_{max}$ and J 2 | d$_{ij}$ = 1 | C$_{max}$ (whereas
other closely related versions such as J 2 | n$_j$ \leq
3 | C$_{max}$, J 2 | d$_{ij}$ ) \ldots{}",
acknowledgement = ack-nhfb,
articleno = "8",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Mollah:2019:MUG,
author = "Md Atiqul Mollah and Wenqi Wang and Peyman Faizian and
MD Shafayat Rahman and Xin Yuan and Scott Pakin and
Michael Lang",
title = "Modeling Universal Globally Adaptive Load-Balanced
Routing",
journal = j-TOPC,
volume = "6",
number = "2",
pages = "9:1--9:??",
month = sep,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3349620",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3349620",
abstract = "Universal globally adaptive load-balanced (UGAL)
routing has been proposed for various interconnection
networks and has been deployed in a number of
current-generation supercomputers. Although UGAL-based
schemes have been extensively studied, most existing
results are based on either simulation or measurement.
Without a theoretical understanding of UGAL, multiple
questions remain: For which traffic patterns is UGAL
most suited? In addition, what determines the
performance of the UGAL-based scheme on a particular
network configuration? In this work, we develop a set
of throughput models for UGALbased on linear
programming. We show that the throughput models are
valid across the torus, Dragonfly, and Slim Fly network
topologies. Finally, we identify a robust model that
can accurately and efficiently predict UGAL throughput
for a set of representative traffic patterns across
different topologies. Our models not only provide a
mechanism to predict UGAL performance on large-scale
interconnection networks but also reveal the inner
working of UGAL and further our understanding of this
type of routing.",
acknowledgement = ack-nhfb,
articleno = "9",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Bateni:2019:ISI,
author = "Mohammed Hossein Bateni and Mohammad T. Hajiaghayi and
Silvio Lattanzi",
title = "Introduction to the Special Issue for {SPAA'17}",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "10:1--10:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3363417",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3363417",
acknowledgement = ack-nhfb,
articleno = "10",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Guha:2019:DPC,
author = "Sudipto Guha and Yi Li and Qin Zhang",
title = "Distributed Partial Clustering",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "11:1--11:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3322808",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3322808",
abstract = "Recent years have witnessed an increasing popularity
of algorithm design for distributed data, largely due
to the fact that massive datasets are often collected
and stored in different locations. In the distributed
setting, communication typically dominates the query
processing time. Thus, it becomes crucial to design
communication-efficient algorithms for queries on
distributed data. Simultaneously, it has been widely
recognized that partial optimizations, where we are
allowed to disregard a small part of the data, provide
us significantly better solutions. The motivation for
disregarded points often arises from noise and other
phenomena that are pervasive in large data scenarios.
In this article, we focus on partial clustering
problems, k -center, k -median, and k -means objectives
in the distributed model, and provide algorithms with
communication sublinear of the input size. As a
consequence, we develop the first algorithms for the
partial k -median and means objectives that run in
subquadratic running time. We also initiate the study
of distributed algorithms for clustering uncertain
data, where each data point can possibly fall into
multiple locations under certain probability
distribution.",
acknowledgement = ack-nhfb,
articleno = "11",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Fraigniaud:2019:DDC,
author = "Pierre Fraigniaud and Dennis Olivetti",
title = "Distributed Detection of Cycles",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "12:1--12:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3322811",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3322811",
abstract = "Distributed property testing in networks has been
introduced by Brakerski and Patt-Shamir [6], with the
objective of detecting the presence of large dense
sub-networks in a distributed manner. Recently,
Censor-Hillel et al. [7] have revisited this notion and
formalized it in a broader context. In particular, they
have shown how to detect 3-cycles in a constant number
of rounds by a distributed algorithm. In a follow-up
work, Fraigniaud et al. [21] have shown how to detect
4-cycles in a constant number of rounds as well.
However, the techniques in these latter works were
shown not to generalize to larger cycles C$_k$ with k
{$>$}= 5. In this article, we completely settle the
problem of cycle detection by establishing the
following result: For every k {$>$}= 3, there exists a
distributed property testing algorithm for C$_k$
-freeness, performing in a constant number of rounds.
All these results hold in the classical congest model
for distributed network computing. Our algorithm is
1-sided error. Its round-complexity is O (1 \epsilon )
where \epsilon \in (0,1) is the property-testing
parameter measuring the gap between legal and illegal
instances.",
acknowledgement = ack-nhfb,
articleno = "12",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Albers:2019:ECD,
author = "Susanne Albers",
title = "On Energy Conservation in Data Centers",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "13:1--13:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3364210",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3364210",
abstract = "We formulate and study an optimization problem that
arises in the energy management of data centers and,
more generally, multiprocessor environments. Data
centers host a large number of heterogeneous servers.
Each server has an active state and several
standby/sleep states with individual power consumption
rates. The demand for computing capacity varies over
time. Idle servers may be transitioned to low-power
modes so as to rightsize the pool of active servers.
The goal is to find a state transition schedule for the
servers that minimizes the total energy consumed. On a
small scale, the same problem arises in multicore
architectures with heterogeneous processors on a chip.
One has to determine active and idle periods for the
cores so as to guarantee a certain service and minimize
the consumed energy. For this power/capacity management
problem, we develop two main results. We use the
terminology of the data center setting. First, we
investigate the scenario that each server has two
states: an active state and a sleep state. We show that
an optimal solution, minimizing energy consumption, can
be computed in polynomial time by a combinatorial
algorithm. The algorithm resorts to a single-commodity
minimum-cost flow computation. Second, we study the
general scenario that each server has an active state
and multiple standby/sleep states. We devise a",
acknowledgement = ack-nhfb,
articleno = "13",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Feldkord:2019:MSP,
author = "Bj{\"o}rn Feldkord and Friedhelm Meyer Auf Der Heide",
title = "The Mobile Server Problem",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "14:1--14:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3364204",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3364204",
abstract = "We introduce the Mobile Server problem, inspired by
current trends to move computational tasks from cloud
structures to multiple devices close to the end user.
An example of this is embedded systems in autonomous
cars that communicate to coordinate their actions. Our
model is a variant of the classical Page Migration
problem. More formally, we consider a mobile server
holding a data page. The server can move in the
Euclidean space (of arbitrary dimension). In every
round, requests for data items from the page pop up at
arbitrary points in the space. The requests are served,
each at a cost of the distance from the requesting
point and the server, and the mobile server may move,
at a cost D times the distance traveled for some
constant D. We assume a maximum distance m that the
server is allowed to move per round. We show that no
online algorithm can achieve a competitive ratio
independent of the length of the input sequence in this
setting. Hence, we augment the maximum movement
distance of the online algorithms to \ldots{}",
acknowledgement = ack-nhfb,
articleno = "14",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Azar:2019:TBC,
author = "Yossi Azar and Danny Vainstein",
title = "Tight Bounds for Clairvoyant Dynamic Bin Packing",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "15:1--15:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3364214",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3364214",
abstract = "In this article, we focus on the Clairvoyant Dynamic
Bin Packing (DBP) problem, which extends the Classical
Online Bin Packing problem in that items arrive and
depart over time and the departure time of an item is
known upon its arrival. The problem naturally arises
when handling cloud-based networks. We focus
specifically on the MinUsageTime objective function,
which aims to minimize the overall usage time of all
bins that are opened during the packing process.
Earlier work has shown a O (log \mu / log log \mu)
upper bound on the algorithm's competitiveness, where
\mu is defined as the ratio between the maximal and
minimal durations of all items. We improve the upper
bound by giving a \ldots{}",
acknowledgement = ack-nhfb,
articleno = "15",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Cooper:2019:NCT,
author = "Colin Cooper and Tomasz Radzik and Nicolas Rivera",
title = "New Cover Time Bounds for the Coalescing-Branching
Random Walk on Graphs",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "16:1--16:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3364206",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3364206",
abstract = "We present new bounds on the cover time of the
coalescing-branching random walk process COBRA. The
COBRA process, introduced in Dutta et al. [9], can be
viewed as spreading a single item of information
throughout an undirected graph in synchronised rounds.
In each round, each vertex that has received the
information in the previous round (possibly
simultaneously from more than one neighbour and
possibly not for the first time), ``pushes'' the
information to k randomly selected neighbours. The
COBRA process is typically studied for integer
branching rates k {$>$}= 2 (with the case k =1
corresponding to a random walk). The aim of the process
is to propagate the information quickly, but with a
limited number of transmissions per vertex per round.
The COBRA cover time is the expected number of rounds
until all vertices have received the information at
least once. Our main results are bounds of O ( m + (
d$_{max}$ )$^2$ log n ) and O ( m log n ) on the COBRA
cover time for arbitrary connected graphs with n
vertices, m edges and maximum graph degree d$_{max}$,
and bounds of \ldots{}",
acknowledgement = ack-nhfb,
articleno = "16",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Sun:2019:DGC,
author = "He Sun and Luca Zanetti",
title = "Distributed Graph Clustering and Sparsification",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "17:1--17:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3364208",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3364208",
abstract = "Graph clustering is a fundamental computational
problem with a number of applications in algorithm
design, machine learning, data mining, and analysis of
social networks. Over the past decades, researchers
have proposed a number of algorithmic design methods
for graph clustering. Most of these methods, however,
are based on complicated spectral techniques or convex
optimisation and cannot be directly applied for
clustering many networks that occur in practice, whose
information is often collected on different sites.
Designing a simple and distributed clustering algorithm
is of great interest and has comprehensive applications
for processing big datasets. In this article, we
present a simple and distributed algorithm for graph
clustering: For a wide class of graphs that are
characterised by a strong cluster-structure, our
algorithm finishes in a poly-logarithmic number of
rounds and recovers a partition of the graph close to
optimal. One of the main procedures behind our
algorithm is a sampling scheme that, given a dense
graph as input, produces a sparse subgraph that
provably preserves the cluster-structure of the input.
Compared with previous sparsification algorithms that
require Laplacian solvers or involve combinatorial
constructions, this procedure is easy to implement in a
distributed setting and runs fast in practice.",
acknowledgement = ack-nhfb,
articleno = "17",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Khan:2019:NOP,
author = "Shahbaz Khan",
title = "Near Optimal Parallel Algorithms for Dynamic {DFS} in
Undirected Graphs",
journal = j-TOPC,
volume = "6",
number = "3",
pages = "18:1--18:??",
month = oct,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3364212",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3364212",
abstract = "Depth first search (DFS) tree is a fundamental data
structure for solving various graph problems. The
classical algorithm [54] for building a DFS tree
requires O ( m + n ) time for a given undirected graph
G having n vertices and m edges. Recently, Baswana et
al. [5] presented a simple algorithm for updating the
DFS tree of an undirected graph after an edge/vertex
update in {\tilde O}( n )$^1$ time. However, their
algorithm is strictly sequential. \ldots{}",
acknowledgement = ack-nhfb,
articleno = "18",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Rauchwerger:2019:ISI,
author = "Lawrence Rauchwerger and Jaejin Lee and Armando
Solar-Lezama and Guy Steele",
title = "Introduction to the Special Issue on {PPoPP} 2017
(Part 1)",
journal = j-TOPC,
volume = "6",
number = "4",
pages = "19:1--19:??",
month = dec,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3373151",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Dec 27 16:13:12 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
acknowledgement = ack-nhfb,
articleno = "19e",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Schardl:2019:TER,
author = "Tao B. Schardl and William S. Moses and Charles E.
Leiserson",
title = "{Tapir}: Embedding Recursive Fork-join Parallelism
into {LLVM}'s Intermediate Representation",
journal = j-TOPC,
volume = "6",
number = "4",
pages = "19:1--19:??",
month = dec,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3365655",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Dec 27 16:13:12 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/pvm.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3365655",
abstract = "Tapir (pronounced TAY-per) is a compiler intermediate
representation (IR) that embeds recursive fork-join
parallelism, as supported by task-parallel programming
platforms such as Cilk and OpenMP, into a mainstream
compiler's IR. Mainstream compilers typically treat
parallel linguistic constructs as syntactic sugar for
function calls into a parallel runtime. These calls
prevent the compiler from performing optimizations on
and across parallel control constructs. Remedying this
situation has generally been thought to require an
extensive reworking of compiler analyses and code
transformations to handle parallel semantics. Tapir
leverages the ``serial-projection property,'' which is
commonly satisfied by task-parallel programs, to handle
the semantics of these programs without an extensive
rework of the compiler. For recursive fork-join
programs that satisfy the serial-projection property,
Tapir enables effective compiler optimization of
parallel programs with only minor changes to existing
compiler analyses and code transformations. Tapir uses
the serial-projection property to order logically
parallel fine-grained tasks in the program's
control-flow graph. This ordered representation of
parallel tasks allows the compiler to optimize parallel
codes effectively with only minor modifications. For
example, to implement Tapir/LLVM, a prototype of Tapir
in the LLVM compiler, we added or modified less than
3,000 lines of LLVM's half-million-line core middle-end
functionality. These changes sufficed to enable LLVM's
existing compiler optimizations for serial
code-including loop-invariant-code motion,
common-subexpression elimination, and tail-recursion
elimination-to work with parallel control constructs
such as parallel loops and Cilk's Cilk\_Spawn keyword.
Tapir also supports parallel optimizations, such as
loop scheduling, which restructure the parallel control
flow of the program. By making use of existing LLVM
optimizations and new parallel optimizations,
Tapir/LLVM can optimize recursive fork-join programs
more effectively than traditional compilation methods.
On a suite of 35 Cilk application benchmarks,
Tapir/LLVM produces more efficient executables for 30
benchmarks, with faster 18-core running times for 26 of
them, compared to a nearly identical compiler that
compiles parallel linguistic constructs the traditional
way.",
acknowledgement = ack-nhfb,
articleno = "19",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Utterback:2019:POR,
author = "Robert Utterback and Kunal Agrawal and I-Ting Angelina
Lee and Milind Kulkarni",
title = "Processor-Oblivious Record and Replay",
journal = j-TOPC,
volume = "6",
number = "4",
pages = "20:1--20:??",
month = dec,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3365659",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Dec 27 16:13:12 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/multithreading.bib;
http://www.math.utah.edu/pub/tex/bib/pvm.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/ft_gateway.cfm?id=3365659",
abstract = "Record-and-replay systems are useful tools for
debugging non-deterministic parallel programs by first
recording an execution and then replaying that
execution to produce the same access pattern. Existing
record-and-replay systems generally target thread-based
execution models, and record the behaviors and
interleavings of individual threads. Dynamic
multithreaded languages and libraries, such as the Cilk
family, OpenMP, TBB, and the like, do not have a notion
of threads. Instead, these languages provide a
processor-oblivious model of programming, where
programs expose task parallelism using high-level
constructs such as spawn/sync without regard to the
number of threads/cores available to run the program.
Thread-based record-and-replay would violate the
processor-oblivious nature of these programs, as they
incorporate the number of threads into the recorded
information, constraining the replayed execution to the
same number of threads. In this article, we present a
processor-oblivious record-and-replay scheme for
dynamic multithreaded languages where record and replay
can use different number of processors and both are
scheduled using work stealing. We provide theoretical
guarantees for our record and replay scheme-namely that
record is optimal for programs with one lock and replay
is near-optimal for all cases. In addition, we
implemented this scheme in the Cilk Plus runtime system
and our evaluation indicates that
processor-obliviousness does not cause substantial
overheads.",
acknowledgement = ack-nhfb,
articleno = "20",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Yeh:2019:PGR,
author = "Tsung Tai Yeh and Amit Sabne and Putt Sakdhnagool and
Rudolf Eigenmann and Timothy G. Rogers",
title = "{Pagoda}: a {GPU} Runtime System for Narrow Tasks",
journal = j-TOPC,
volume = "6",
number = "4",
pages = "21:1--21:??",
month = nov,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3365657",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/multithreading.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "Massively multithreaded GPUs achieve high throughput
by running thousands of threads in parallel. To fully
utilize the their hardware, contemporary workloads
spawn work to the GPU in bulk by launching large tasks,
where each task is a kernel that contains thousands of
threads that occupy the entire GPU. GPUs face severe
underutilization and their performance benefits vanish
if the tasks are narrow, i.e., they contain less than
512 threads. Latency-sensitive applications in network,
signal, and image processing that generate a large
number of tasks with relatively small inputs are
examples of such limited parallelism. This article
presents Pagoda, a runtime system that virtualizes GPU
resources, using an OS-like daemon kernel called
MasterKernel. Tasks are spawned from the CPU onto
Pagoda as they become available, and are scheduled by
the MasterKernel at the warp granularity. This level of
control enables the GPU to keep scheduling and
executing tasks as long as free warps are found,
dramatically reducing underutilization. Experimental
results on real hardware demonstrate that Pagoda
achieves a geometric mean speedup of 5.52X over
PThreads running on a 20-core CPU, 1.76X over
CUDA-HyperQ, and 1.44X over GeMTC, the state-of-the-art
runtime GPU task scheduling system.",
acknowledgement = ack-nhfb,
articleno = "21",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Steele:2019:UBP,
author = "Guy L. {Steele Jr.} and Jean-Baptiste Tristan",
title = "Using Butterfly-patterned Partial Sums to Draw from
Discrete Distributions",
journal = j-TOPC,
volume = "6",
number = "4",
pages = "22:1--22:??",
month = nov,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3365662",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/prng.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "We describe a SIMD technique for drawing values from
multiple discrete distributions, such as sampling from
the random variables of a mixture model, that avoids
computing a complete table of partial sums of the
relative probabilities. A table of alternate
(``butterfly-patterned'') form is faster to compute,
making better use of coalesced memory accesses; from
this table, complete partial sums are computed on the
fly during a binary search. Measurements using Cuda 7.5
on an NVIDIA Titan Black GPU show that this technique
makes an entire machine-learning application that uses
a Latent Dirichlet Allocation topic model with 1,024
topics about 13\% faster (when using single-precision
floating-point data) or about 35\% faster (when using
double-precision floating-point data) than doing a
straightforward matrix transposition after using
coalesced accesses.",
acknowledgement = ack-nhfb,
articleno = "22",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Vandierendonck:2019:HDI,
author = "Hans Vandierendonck and Dimitrios S. Nikolopoulos",
title = "Hyperqueues: Design and Implementation of
Deterministic Concurrent Queues",
journal = j-TOPC,
volume = "6",
number = "4",
pages = "23:1--23:??",
month = nov,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3365660",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Wed Nov 20 07:59:59 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The hyperqueue is a programming abstraction for queues
that results in deterministic and scale-free parallel
programs. Hyperqueues extend the concept of Cilk++
hyperobjects to provide thread-local views on a shared
data structure. While hyperobjects are organized around
private local views, hyperqueues provide a shared view
on a queue data structure. Hereby, hyperqueues
guarantee determinism for programs using concurrent
queues. We define the programming API and semantics of
two instances of the hyperqueue concept. These
hyperqueues differ in their API and the degree of
concurrency that is extracted. We describe the
implementation of the hyperqueues in a work-stealing
scheduler and demonstrate scalable performance on
pipeline-parallel benchmarks from PARSEC and
StreamIt.",
acknowledgement = ack-nhfb,
articleno = "23",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Ren:2019:ESP,
author = "Bin Ren and Shruthi Balakrishna and Youngjoon Jo and
Sriram Krishnamoorthy and Kunal Agrawal and Milind
Kulkarni",
title = "Extracting {SIMD} Parallelism from Recursive
Task-Parallel Programs",
journal = j-TOPC,
volume = "6",
number = "4",
pages = "24:1--24:??",
month = dec,
year = "2019",
CODEN = "????",
DOI = "https://doi.org/10.1145/3365663",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Fri Dec 27 16:13:12 MST 2019",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
abstract = "The pursuit of computational efficiency has led to the
proliferation of throughput-oriented hardware, from
GPUs to increasingly wide vector units on commodity
processors and accelerators. This hardware is designed
to execute data-parallel computations in a vectorized
manner efficiently. However, many algorithms are more
naturally expressed as divide-and-conquer, recursive,
task-parallel computations. In the absence of data
parallelism, it seems that such algorithms are not well
suited to throughput-oriented architectures. This
article presents a set of novel code transformations
that expose the data parallelism latent in recursive,
task-parallel programs. These transformations
facilitate straightforward vectorization of
task-parallel programs on commodity hardware. We also
present scheduling policies that maintain high
utilization of vector resources while limiting space
usage. Across several task-parallel benchmarks, we
demonstrate both efficient vector resource utilization
and substantial speedup on chips using Intel's SSE4.2
vector units, as well as accelerators using Intel's
AVX512 units. We then show through rigorous sampling
that, in practice, our vectorization techniques are
effective for a much larger class of programs.",
acknowledgement = ack-nhfb,
articleno = "24",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "http://dl.acm.org/citation.cfm?id=2632163",
}
@Article{Tumeo:2020:ITS,
author = "Antonino Tumeo and Fabrizio Petrini and John Feo and
Mahantesh Halappanavar",
title = "Introduction to the {TOPC} Special Issue on
Innovations in Systems for Irregular Applications,{Part
1}",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "1:1--1:2",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3383318",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3383318",
acknowledgement = ack-nhfb,
articleno = "1",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Anzt:2020:LBS,
author = "Hartwig Anzt and Terry Cojean and Chen Yen-Chen and
Jack Dongarra and Goran Flegar and Pratik Nayak and
Stanimire Tomov and Yuhsiang M. Tsai and Weichung
Wang",
title = "Load-balancing Sparse Matrix Vector Product Kernels on
{GPUs}",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "2:1--2:26",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3380930",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3380930",
abstract = "Efficient processing of Irregular Matrices on Single
Instruction, Multiple Data (SIMD)-type architectures is
a persistent challenge. Resolving it requires
innovations in the development of data formats,
computational techniques, and implementations that
strike a balance between thread divergence, which is
inherent for Irregular Matrices, and padding, which
alleviates the performance-detrimental thread
divergence but introduces artificial overheads. To this
end, in this article, we address the challenge of
designing high performance sparse matrix-vector product
(SpMV) kernels designed for Nvidia Graphics Processing
Units (GPUs). We present a compressed sparse row (CSR)
format suitable for unbalanced matrices. We also
provide a load-balancing kernel for the coordinate
(COO) matrix format and extend it to a hybrid algorithm
that stores part of the matrix in SIMD-friendly Ellpack
format (ELL) format. The ratio between the ELL- and the
COO-part is determined using a theoretical analysis of
the nonzeros-per-row distribution. For the over 2,800
test matrices available in the Suite Sparse matrix
collection, we compare the performance against SpMV
kernels provided by NVIDIA's cuSPARSE library and a
heavily-tuned sliced ELL (SELL-P) kernel that prevents
unnecessary padding by considering the irregular
matrices as a combination of matrix blocks stored in
ELL format.",
acknowledgement = ack-nhfb,
articleno = "2",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Hadade:2020:SPU,
author = "Ioan Hadade and Timothy M. Jones and Feng Wang and
Luca di Mare",
title = "Software Prefetching for Unstructured Mesh
Applications",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "3:1--3:23",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3380932",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3380932",
abstract = "This article demonstrates the utility and
implementation of software prefetching in an
unstructured finite volume computational fluid dynamics
code of representative size and complexity to an
industrial application and across a number of modern
\ldots{}",
acknowledgement = ack-nhfb,
articleno = "3",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Grutzmacher:2020:APC,
author = "Thomas Gr{\"u}tzmacher and Terry Cojean and Goran
Flegar and Hartwig Anzt and Enrique S.
Quintana-Ort{\'\i}",
title = "Acceleration of {PageRank} with Customized Precision
Based on Mantissa Segmentation",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "4:1--4:19",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3380934",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/fparith.bib;
http://www.math.utah.edu/pub/tex/bib/pagerank.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3380934",
abstract = "We describe the application of a
communication-reduction technique for the PageRank
algorithm that dynamically adapts the precision of the
data access to the numerical requirements of the
algorithm as the iteration converges. Our
variable-precision strategy, using a customized
precision format based on mantissa segmentation (CPMS),
abandons the IEEE 754 single- and double-precision
number representation formats employed in the standard
implementation of PageRank, and instead handles the
data in memory using a customized floating-point
format. The customized format enables fast data access
in different accuracy, prevents overflow/underflow by
preserving the IEEE 754 double-precision exponent, and
efficiently avoids data duplication, since all bits of
the original IEEE 754 double-precision mantissa are
preserved in memory, but re-organized for efficient
reduced precision access. With this approach, the
truncated values (omitting significand bits), as well
as the original IEEE double-precision values, can be
retrieved without duplicating the data in different
formats.\par
Our numerical experiments on an NVIDIA V100 GPU (Volta
architecture) and a server equipped with two Intel Xeon
Platinum 8168 CPUs (48 cores in total) expose that,
compared with a standard IEEE double-precision
implementation, the CPMS-based PageRank completes about
10\% faster if high-accuracy output is needed, and
about 30\% faster if reduced output accuracy is
acceptable.",
acknowledgement = ack-nhfb,
articleno = "4",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Das:2020:SMP,
author = "Apurba Das and Seyed-Vahid Sanei-Mehri and Srikanta
Tirthapura",
title = "Shared-memory Parallel Maximal Clique Enumeration from
Static and Dynamic Graphs",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "5:1--5:28",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3380936",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3380936",
abstract = "Maximal Clique Enumeration (MCE) is a fundamental
graph mining problem and is useful as a primitive in
identifying dense structures in a graph. Due to the
high computational cost of MCE, parallel methods are
imperative for dealing with large graphs. We \ldots{}",
acknowledgement = ack-nhfb,
articleno = "5",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Hamilton:2020:ASC,
author = "Kathleen E. Hamilton and Catherine D. Schuman and
Steven R. Young and Ryan S. Bennink and Neena Imam and
Travis S. Humble",
title = "Accelerating Scientific Computing in the
Post-{Moore}'s Era",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "6:1--6:31",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3380940",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3380940",
abstract = "Novel uses of graphical processing units for
accelerated computation revolutionized the field of
high-performance scientific computing by providing
specialized workflows tailored to algorithmic
requirements. As the era of Moore's law draws to a
close, \ldots{}",
acknowledgement = ack-nhfb,
articleno = "6",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Lakhotia:2020:GSC,
author = "Kartik Lakhotia and Rajgopal Kannan and Sourav Pati
and Viktor Prasanna",
title = "{GPOP}: a Scalable Cache- and Memory-efficient
Framework for Graph Processing over Parts",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "7:1--7:24",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3380942",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3380942",
abstract = "The past decade has seen the development of many
shared-memory graph processing frameworks intended to
reduce the effort of developing high-performance
parallel applications. However, many of these
frameworks, based on Vertex-centric or Edge-centric
\ldots{}",
acknowledgement = ack-nhfb,
articleno = "7",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Anderson:2020:RRO,
author = "Jeff Anderson and Engin Kayraklioglu and Shuai Sun and
Joseph Crandall and Yousra Alkabani and Vikram Narayana
and Volker Sorger and Tarek El-Ghazawi",
title = "{ROC}: a Reconfigurable Optical Computer for
Simulating Physical Processes",
journal = j-TOPC,
volume = "7",
number = "1",
pages = "8:1--8:29",
month = apr,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3380944",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Apr 6 08:56:55 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3380944",
abstract = "Due to the end of Moore's law and Dennard scaling, we
are entering a new era of processors. Computing systems
are increasingly facing power and performance
challenges due to both device- and circuit-related
challenges with resistive and capacitive \ldots{}",
acknowledgement = ack-nhfb,
articleno = "8",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Monemi:2020:EDW,
author = "Alireza Monemi and Farshad Khunjush and Maurizio
Palesi and Hamid Sarbazi-Azad",
title = "An Enhanced Dynamic Weighted Incremental Technique for
{QoS} Support in {NoC}",
journal = j-TOPC,
volume = "7",
number = "2",
pages = "9:1--9:31",
month = may,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3391442",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Jun 1 09:19:25 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3391442",
abstract = "Providing Quality-of-Service (QoS) in many-core
network-on-chip (NoC) platforms is critical due to the
high level of resource sharing in such systems. This
article presents a hard-built Equality-of-Service (EoS)
and Differential-Service (DS) as subsets \ldots{}",
acknowledgement = ack-nhfb,
articleno = "9",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Natarajan:2020:FLL,
author = "Aravind Natarajan and Arunmoezhi Ramachandran and
Neeraj Mittal",
title = "{FEAST}: a Lightweight Lock-free Concurrent Binary
Search Tree",
journal = j-TOPC,
volume = "7",
number = "2",
pages = "10:1--10:64",
month = may,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3391438",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Jun 1 09:19:25 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3391438",
abstract = "We present a lock-free algorithm for concurrent
manipulation of a binary search tree (BST) in an
asynchronous shared memory system that supports search,
insert, and delete operations. In addition to read and
write instructions, our algorithm uses \ldots{}",
acknowledgement = ack-nhfb,
articleno = "10",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Salah:2020:TSE,
author = "Ahmad Salah and Kenli Li and Qing Liao and Mervat
Hashem and Zhiyong Li and Anthony T. Chronopoulos and
Albert Y. Zomaya",
title = "A Time-space Efficient Algorithm for Parallel $k$-way
In-place Merging based on Sequence Partitioning and
Perfect Shuffle",
journal = j-TOPC,
volume = "7",
number = "2",
pages = "11:1--11:23",
month = may,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3391443",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Jun 1 09:19:25 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3391443",
abstract = "The huge data volumes, big data, and the emergence of
new parallel architectures lead to revisiting classic
computer science topics. The motivation of the proposed
work for revisiting the parallel k -way in-place
merging is primarily related to the \ldots{}",
acknowledgement = ack-nhfb,
articleno = "11",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Duan:2020:CSR,
author = "Shaohua Duan and Pradeep Subedi and Philip Davis and
Keita Teranishi and Hemanth Kolla and Marc Gamell and
Manish Parashar",
title = "{CoREC}: Scalable and Resilient In-memory Data Staging
for In-situ Workflows",
journal = j-TOPC,
volume = "7",
number = "2",
pages = "12:1--12:29",
month = may,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3391448",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Jun 1 09:19:25 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3391448",
abstract = "The dramatic increase in the scale of current and
planned high-end HPC systems is leading new challenges,
such as the growing costs of data movement and IO, and
the reduced mean time between failures (MTBF) of system
components. In-situ workflows, i.e., \ldots{}",
acknowledgement = ack-nhfb,
articleno = "12",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Alam:2020:GMS,
author = "Maksudul Alam and Maleq Khan and Kalyan S. Perumalla
and Madhav Marathe",
title = "Generating Massive Scale-free Networks: Novel Parallel
Algorithms using the Preferential Attachment Model",
journal = j-TOPC,
volume = "7",
number = "2",
pages = "13:1--13:35",
month = may,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3391446",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Mon Jun 1 09:19:25 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3391446",
abstract = "Recently, there has been substantial interest in the
study of various random networks as mathematical models
of complex systems. As real-life complex systems grow
larger, the ability to generate progressively large
random networks becomes all the more \ldots{}",
acknowledgement = ack-nhfb,
articleno = "13",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Lee:2020:ISI,
author = "Jaejin Lee and Lawrence Rauchwerger and Armando
Solar-Lezama and Guy Steele",
title = "Introduction to the Special Issue on {PPoPP 2017}
(Part 2)",
journal = j-TOPC,
volume = "7",
number = "3",
pages = "14:1--14:2",
month = aug,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3407185",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu Aug 6 08:56:07 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3407185",
acknowledgement = ack-nhfb,
articleno = "14",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Jiang:2020:CSM,
author = "Peng Jiang and Yang Xia and Gagan Agrawal",
title = "Combining {SIMD} and Many\slash Multi-core Parallelism
for Finite-state Machines with Enumerative
Speculation",
journal = j-TOPC,
volume = "7",
number = "3",
pages = "15:1--15:26",
month = aug,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3399714",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu Aug 6 08:56:07 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3399714",
abstract = "Finite-state Machine (FSM) is the key kernel behind
many popular applications, including regular expression
matching, text tokenization, and Huffman decoding.
Parallelizing FSMs is extremely difficult because of
the strong dependencies and unpredictable \ldots{}",
acknowledgement = ack-nhfb,
articleno = "15",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Basin:2020:KKV,
author = "Dmitry Basin and Edward Bortnikov and Anastasia
Braginsky and Guy Golan-Gueta and Eshcar Hillel and
Idit Keidar and Moshe Sulamy",
title = "{KiWi}: a Key--value Map for Scalable Real-time
Analytics",
journal = j-TOPC,
volume = "7",
number = "3",
pages = "16:1--16:28",
month = aug,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3399718",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu Aug 6 08:56:07 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/java2020.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3399718",
abstract = "We present KiWi, the first atomic KV-map to
efficiently support simultaneous large scans and
real-time access. The key to achieving this is treating
scans as first class citizens and organizing the data
structure around them. KiWi provides wait-free scans,
whereas its put operations are lightweight and
lock-free. It optimizes memory management jointly with
data structure access. We implement KiWi and compare it
to state-of-the-art solutions. Compared to other
KV-maps providing atomic scans, KiWi performs either
long scans or concurrent puts an order of magnitude
faster. Its scans are twice as fast as non-atomic ones
implemented via iterators in the Java skiplist.",
acknowledgement = ack-nhfb,
articleno = "16",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Chabbi:2020:EAL,
author = "Milind Chabbi and Abdelhalim Amer and Xu Liu",
title = "Efficient Abortable-locking Protocol for Multi-level
{NUMA} Systems: Design and Correctness",
journal = j-TOPC,
volume = "7",
number = "3",
pages = "17:1--17:32",
month = aug,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3399728",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu Aug 6 08:56:07 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3399728",
abstract = "The popularity of Non-Uniform Memory Access (NUMA)
architectures has led to numerous locality-preserving
hierarchical lock designs, such as HCLH, HMCS, and
cohort locks. Locality-preserving locks trade fairness
for higher throughput. Hence, some \ldots{}",
acknowledgement = ack-nhfb,
articleno = "17",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Ben-Nun:2020:GAM,
author = "Tal Ben-Nun and Michael Sutton and Sreepathi Pai and
Keshav Pingali",
title = "{Groute}: Asynchronous Multi-{GPU} Programming Model
with Applications to Large-scale Graph Processing",
journal = j-TOPC,
volume = "7",
number = "3",
pages = "18:1--18:27",
month = aug,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3399730",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu Aug 6 08:56:07 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3399730",
abstract = "Nodes with multiple GPUs are becoming the platform of
choice for high-performance computing. However, most
applications are written using bulk-synchronous
programming models, which may not be optimal for
irregular algorithms that benefit from low-. \ldots{}",
acknowledgement = ack-nhfb,
articleno = "18",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Alappat:2020:RAC,
author = "Christie Alappat and Achim Basermann and Alan R.
Bishop and Holger Fehske and Georg Hager and Olaf
Schenk and Jonas Thies and Gerhard Wellein",
title = "A Recursive Algebraic Coloring Technique for
Hardware-efficient Symmetric Sparse Matrix--vector
Multiplication",
journal = j-TOPC,
volume = "7",
number = "3",
pages = "19:1--19:37",
month = aug,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3399732",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu Aug 6 08:56:07 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3399732",
abstract = "The symmetric sparse matrix-vector multiplication
(SymmSpMV) is an important building block for many
numerical linear algebra kernel operations or graph
traversal applications. Parallelizing SymmSpMV on
today's multicore platforms with up to 100 cores is
difficult due to the need to manage conflicting updates
on the result vector. Coloring approaches can be used
to solve this problem without data duplication, but
existing coloring algorithms do not take load balancing
and deep memory hierarchies into account, hampering
scalability and full-chip performance. In this work, we
propose the recursive algebraic coloring engine (RACE),
a novel coloring algorithm and open-source library
implementation that eliminates the shortcomings of
previous coloring methods in terms of hardware
efficiency and parallelization overhead. We describe
the level construction, distance-$k$ coloring, and load
balancing steps in RACE, use it to parallelize
SymmSpMV, and compare its performance on 31 sparse
matrices with other state-of-the-art coloring
techniques and Intel MKL on two modern multicore
processors. RACE outperforms all other approaches
substantially. By means of a parameterized roofline
model, we analyze the SymmSpMV performance in detail
and discuss outliers. While we focus on SymmSpMV in
this article, our algorithm and software are applicable
to any sparse matrix operation with data dependencies
that can be resolved by distance-$k$ coloring.",
acknowledgement = ack-nhfb,
articleno = "19",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}
@Article{Davydov:2020:ADS,
author = "Denis Davydov and Martin Kronbichler",
title = "Algorithms and Data Structures for Matrix-Free Finite
Element Operators with {MPI}-Parallel Sparse
Multi-Vectors",
journal = j-TOPC,
volume = "7",
number = "3",
pages = "20:1--20:30",
month = aug,
year = "2020",
CODEN = "????",
DOI = "https://doi.org/10.1145/3399736",
ISSN = "2329-4949 (print), 2329-4957 (electronic)",
ISSN-L = "2329-4949",
bibdate = "Thu Aug 6 08:56:07 MDT 2020",
bibsource = "http://www.math.utah.edu/pub/tex/bib/pvm.bib;
http://www.math.utah.edu/pub/tex/bib/topc.bib",
URL = "https://dl.acm.org/doi/abs/10.1145/3399736",
abstract = "Traditional solution approaches for problems in
quantum mechanics scale as $ O(M^3) $, where $M$ is the
number of electrons. Various methods have been proposed
to address this issue and obtain a linear scaling $
O(M)$. One promising formulation is the direct
minimization of energy. Such methods take advantage of
physical localization of the solution, allowing users
to seek it in terms of non-orthogonal orbitals with
local support.\par
This work proposes a numerically efficient
implementation of sparse parallel vectors within the
open-source finite element library deal.II. The main
algorithmic ingredient is the matrix-free evaluation of
the Hamiltonian operator by cell-wise quadrature. Based
on an a-priori chosen support for each vector, we
develop algorithms and data structures to perform (i)
matrix-free sparse matrix multivector products (SpMM),
(ii) the projection of an operator onto a sparse
sub-space (inner products), and (iii)
post-multiplication of a sparse multivector with a
square matrix. The node-level performance is analyzed
using a roofline model. Our matrix-free implementation
of finite element operators with sparse multivectors
achieves a performance of 157 GFlop/s on an Intel
Cascade Lake processor with 20 cores. Strong and weak
scaling results are reported for a representative
benchmark problem using quadratic and quartic finite
element bases.",
acknowledgement = ack-nhfb,
articleno = "20",
fjournal = "ACM Transactions on Parallel Computing",
journal-URL = "https://dl.acm.org/loi/topc",
}