%%% -*-BibTeX-*- %%% ==================================================================== %%% BibTeX-file{ %%% author = "Nelson H. F. Beebe", %%% version = "1.17", %%% date = "04 November 2016", %%% time = "16:53:38 MDT", %%% filename = "jea.bib", %%% address = "University of Utah %%% Department of Mathematics, 110 LCB %%% 155 S 1400 E RM 233 %%% Salt Lake City, UT 84112-0090 %%% USA", %%% telephone = "+1 801 581 5254", %%% FAX = "+1 801 581 4148", %%% checksum = "18869 10945 60151 566334", %%% email = "beebe at math.utah.edu, beebe at acm.org, %%% beebe at computer.org (Internet)", %%% codetable = "ISO/ASCII", %%% keywords = "ACM Journal of Experimental Algorithmics; %%% bibliography; BibTeX; JEA", %%% license = "public domain", %%% supported = "yes", %%% docstring = "This is a COMPLETE bibliography of the %%% ACM Journal of Experimental Algorithmics %%% (CODEN none, ISSN 1084-6654). The journal %%% appears once a year, and has no separate %%% issue numbers. %%% %%% At version 1.17, the COMPLETE year coverage %%% looked like this: %%% %%% 1996 ( 4) 2003 ( 6) 2010 ( 12) %%% 1997 ( 5) 2004 ( 6) 2011 ( 18) %%% 1998 ( 9) 2005 ( 15) 2012 ( 2) %%% 1999 ( 8) 2006 ( 17) 2013 ( 14) %%% 2000 ( 17) 2007 ( 0) 2014 ( 0) %%% 2001 ( 10) 2008 ( 29) 2015 ( 25) %%% 2002 ( 12) 2009 ( 28) 2016 ( 18) %%% %%% Article: 255 %%% %%% Total entries: 255 %%% %%% The author will be grateful for reports of %%% any errors or omissions in this file; they %%% will be corrected in future editions. %%% %%% Articles and letters or corrections that %%% comment on them are cross-referenced in both %%% directions, so that citation of one of them %%% will automatically include the others. %%% %%% The ACM maintains Web pages for this journal at %%% %%% http://portal.acm.org/browse_dl.cfm?idx=J430 %%% %%% That data has been automatically converted %%% to BibTeX form, corrected for spelling and %%% page number errors, and merged into this %%% file. %%% %%% ACM copyrights explicitly permit abstracting %%% with credit, so article abstracts, keywords, %%% and subject classifications have been %%% included in this bibliography wherever %%% available. Article reviews have been %%% omitted, until their copyright status has %%% been clarified. %%% %%% Spelling has been verified with the UNIX %%% spell and GNU ispell programs using the %%% exception dictionary stored in the companion %%% file with extension .sok. %%% %%% BibTeX citation tags are uniformly chosen %%% as name:year:abbrev, where name is the %%% family name of the first author or editor, %%% year is a 4-digit number, and abbrev is a %%% 3-letter condensation of important title %%% words. Citation tags were automatically %%% generated by the biblabel software %%% developed for the BibNet Project. %%% %%% In this bibliography, entries are sorted in %%% publication order, with the help of %%% ``bibsort -byvolume''. The bibsort utility, %%% and several related programs for %%% bibliography maintenance, is available on %%% ftp.math.utah.edu in /pub/tex/bib, and at %%% other Internet sites which mirror it, %%% including the Comprehensive TeX Archive %%% Network (CTAN); the command `finger %%% ctan<at>tug.org' will produce a list of %%% CTAN hosts. %%% %%% The checksum field above contains a CRC-16 %%% checksum as the first value, followed by the %%% equivalent of the standard UNIX wc (word %%% count) utility output of lines, words, and %%% characters. This is produced by Robert %%% Solovay's checksum utility.", %%% } %%% ====================================================================

@Preamble{ "\input path.sty" # "\hyphenation{ }" # "\ifx \undefined \mathbb \def \mathbb #1{{\bf #1}}\fi" }

%%% ==================================================================== %%% Acknowledgement abbreviations:

@String{ack-nhfb= "Nelson H. F. Beebe, University of Utah, Department of Mathematics, 110 LCB, 155 S 1400 E RM 233, Salt Lake City, UT 84112-0090, USA, Tel: +1 801 581 5254, FAX: +1 801 581 4148, e-mail: \path|beebe@math.utah.edu|, \path|beebe@acm.org|, \path|beebe@computer.org| (Internet), URL: \path|http://www.math.utah.edu/~beebe/|"}

%%% ==================================================================== %%% Journal abbreviations:

@String{j-ACM-J-EXP-ALGORITHMICS= "ACM Journal of Experimental Algorithmics"}

%%% ==================================================================== %%% Bibliography entries, sorted in publication order.

@Article{Knuth:1996:II, author = "Donald E. Knuth", title = "Irredundant intervals", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "1", pages = "1:1--1:??", month = "????", year = "1996", CODEN = "????", DOI = "http://doi.acm.org/10.1145/235141.235146", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:01:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "This expository note presents simplifications of a theorem due to Gy{\H{o}}ri and an algorithm due to Franzblau and Kleitman: Given a family $F$ of $m$ intervals on a linearly ordered set of n elements, we can construct in $O(m + n)^2$ steps an irredundant subfamily having maximum cardinality, as well as a generating family having minimum cardinality. The algorithm is of special interest because it solves a problem analogous to finding a maximum independent set, but on a class of objects that is more general than a matroid. This note is also a complete, runnable computer program, which can be used for experiments in conjunction with the public-domain software of The Stanford GraphBase.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gittleman:1996:PSS, author = "Arthur Gittleman", title = "Predicting string search speed", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "1", pages = "2:1--2:??", month = "????", year = "1996", CODEN = "????", DOI = "http://doi.acm.org/10.1145/235141.235147", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:01:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "String search is fundamental in many text processing applications. Sunday recently gave several algorithms to find the first occurrence of a pattern string as a substring of a text, providing experimental data from searches in a text of about 200K characters to support his claim that his algorithms are faster than the standard Boyer--Moore algorithm. We present a methodology for the average-case analysis of the performance of string search algorithms---for such algorithms, a worst-case analysis does not yield much useful information, since the performance of the algorithm is directly affected by such characteristics as the size of the character set, the character frequencies, and the structure of the text. Knuth described a finite automaton which can be used to save information about character comparisons. Baeza-Yates, Gonnet, and Regnier gave a probabilistic analysis of the worst- and average-case behavior of a string search algorithm based upon such an automaton. We construct Knuth automata to model Sunday's algorithms and use the methods of Baeza-Yates et al. to obtain an average-case analysis which confirms Sunday's experimental data.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Bader:1996:PPA, author = "David A. Bader and David R. Helman and Joseph J{\'a}J{\'a}", title = "Practical parallel algorithms for personalized communication and integer sorting", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "1", pages = "3:1--3:??", month = "????", year = "1996", CODEN = "????", DOI = "http://doi.acm.org/10.1145/235141.235148", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:01:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A fundamental challenge for parallel computing is to obtain high-level, architecture independent, algorithms which efficiently execute on general-purpose parallel machines. With the emergence of message passing standards such as MPI, it has become easier to design efficient and portable parallel algorithms by making use of these communication primitives. While existing primitives allow an assortment of collective communication routines, they do not handle an important communication event when most or all processors have non-uniformly sized personalized messages to exchange with each other. We focus in this paper on the h-relation personalized communication whose efficient implementation will allow high performance implementations of a large class of algorithms. While most previous h-relation algorithms use randomization, this paper presents a new deterministic approach for h-relation personalized communication with asymptotically optimal complexity for h>p$^2$. As an application, we present an efficient algorithm for stable integer sorting. The algorithms presented in this paper have been coded in Split-C and run on a variety of platforms, including the Thinking Machines CM-5, IBM SP-1 and SP-2, Cray Research T3D, Meiko Scientific CS-2, and the Intel Paragon. Our experimental results are consistent with the theoretical analysis and illustrate the scalability and efficiency of our algorithms across different platforms. In fact, they seem to outperform all similar algorithms known to the authors on these platforms.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{LaMarca:1996:ICP, author = "Anthony LaMarca and Richard Ladner", title = "The influence of caches on the performance of heaps", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "1", pages = "4:1--4:??", month = "????", year = "1996", CODEN = "????", DOI = "http://doi.acm.org/10.1145/235141.235145", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:01:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "As memory access times grow larger relative to processor cycle times, the cache performance of algorithms has an increasingly large impact on overall performance. Unfortunately, most commonly used algorithms were not designed with cache performance in mind. This paper investigates the cache performance of implicit heaps. We present optimizations which significantly reduce the cache misses that heaps incur and improve their overall performance. We present an analytical model called collective analysis that allows cache performance to be predicted as a function of both cache configuration and algorithm configuration. As part of our investigation, we perform an approximate analysis of the cache performance of both traditional heaps and our improved heaps in our model. In addition empirical data is given for five architectures to show the impact our optimizations have on overall performance. We also revisit a priority queue study originally performed by Jones [25]. Due to the increases in cache miss penalties, the relative performance results we obtain on today's machines differ greatly from the machines of only ten years ago. We compare the performance of implicit heaps, skew heaps and splay trees and discuss the difference between our results and Jones's.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Buchsbaum:1997:AAS, author = "Adam L. Buchsbaum and Raffaele Giancarlo", title = "Algorithmic aspects in speech recognition: an introduction", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "2", pages = "1:1--1:??", month = "????", year = "1997", CODEN = "????", DOI = "http://doi.acm.org/10.1145/264216.264219", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:14 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Speech recognition is an area with a considerable literature, but there is little discussion of the topic within the computer science algorithms literature. Many computer scientists, however, are interested in the computational problems of speech recognition. This paper presents the field of speech recognition and describes some of its major open problems from an algorithmic viewpoint. Our goal is to stimulate the interest of algorithm designers and experimenters to investigate the algorithmic problems of effective automatic speech recognition.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "automata theory; graph searching", } @Article{Battiti:1997:RSH, author = "Roberto Battiti and Marco Protasi", title = "Reactive search, a history-sensitive heuristic for {MAX}-{SAT}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "2", pages = "2:1--2:??", month = "????", year = "1997", CODEN = "????", DOI = "http://doi.acm.org/10.1145/264216.264220", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:14 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The Reactive Search (RS) method proposes the integration of a simple history-sensitive (machine learning) scheme into local search for the on-line determination of free parameters. In this paper a new RS algorithm is proposed for the approximated solution of the Maximum Satisfiability problem: a component based on local search with temporary prohibitions (Tabu Search) is complemented with a reactive scheme that determines the appropriate value of the prohibition parameter by monitoring the Hamming distance along the search trajectory. The proposed algorithm (H-RTS) can therefore be characterized as a dynamic version of Tabu Search. In addition, the non-oblivious functions recently introduced in the framework of approximation algorithms are used to discover a better local optimum in the initial part of the search. The algorithm is developed in two phases. First the bias-diversification properties of individual candidate components are analyzed by extensive empirical evaluation, then a reactive scheme is added to the winning component, based on Tabu Search. The final tests on a benchmark of random MAX-3-SAT and MAX-4-SAT problems demonstrate the superiority of H-RTS with respect to alternative heuristics.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Smith:1997:EHF, author = "Bradley J. Smith and Gregory L. Heileman and Chaouki Abdallah", title = "The exponential hash function", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "2", pages = "3:1--3:??", month = "????", year = "1997", CODEN = "????", DOI = "http://doi.acm.org/10.1145/264216.264221", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:14 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper an efficient open address hash function called exponential hashing is developed. The motivation for this hash function resulted from our ongoing efforts to apply dynamical systems theory to the study of hashing; however, the analysis conducted in this paper is primarily based on traditional number theory. Proofs of optimal table parameter choices are provided for a number of hash functions. We also demonstrate experimentally that exponential hashing essentially matches the performance of a widely-used optimal double hash function for uniform data distributions, and performs significantly better for nonuniform data distributions. We show that exponential hashing exhibits a higher integer Lyapunov exponent and entropy than double hashing for initial data probes, which offers one explanation for its improved performance on nonuniform data distributions.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "ADT; chaos; dynamic dictionary; dynamical systems theory; exponential hashing; Lyapunov exponent; number theory", } @Article{Purchase:1997:ESB, author = "H. C. Purchase and R. F. Cohen and M. I. James", title = "An experimental study of the basis for graph drawing algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "2", pages = "4:1--4:??", month = "????", year = "1997", CODEN = "????", DOI = "http://doi.acm.org/10.1145/264216.264222", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:14 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Designers of graph drawing algorithms and systems claim to illuminate application data by producing layouts that optimise measurable aesthetic qualities. Examples of these aesthetics include symmetry (where possible, a symmetrical view of the graph should be displayed), minimise arc crossing(the number of arc crossings in the display should be minimised), and minimise bends (the total number of bends in polyline arcs should be minimised).The aim of this paper is to describe our work to validate these claims by performing empirical studies of human understanding of graphs drawn using various layout aesthetics. This work is important since it helps indicate to algorithm and system designers what are the aesthetic qualities most important to aid understanding, and consequently to build more effective systems.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "human-computer interaction", } @Article{Alberts:1997:ESD, author = "David Alberts and Giuseppe Cattaneo and Giuseppe F. Italiano", title = "An empirical study of dynamic graph algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "2", pages = "5:1--5:??", month = "????", year = "1997", CODEN = "????", DOI = "http://doi.acm.org/10.1145/264216.264223", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:14 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The contributions of this paper are both of theoretical and of experimental nature. From the experimental point of view, we conduct an empirical study on some dynamic connectivity algorithms which where developed recently. In particular, the following implementations were tested and compared with simple algorithms: simple sparsification by Eppstein et al. and the recent randomized algorithm by Henzinger and King. In our experiments, we considered both random and non-random inputs. Moreover, we present a simplified variant of the algorithm by Henzinger and King, which for random inputs was always faster than the original implementation. For non-random inputs, simple sparsification was the fastest algorithm for small sequences of updates; for medium and large sequences of updates, the original algorithm by Henzinger and King was faster. From the theoretical point of view, we analyze the average case running time of simple sparsification and prove that for dynamic random graphs its logarithmic overhead vanishes.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Hemaspaandra:1998:PBA, author = "Lane A. Hemaspaandra and Kulathur S. Rajasethupathy and Prasanna Sethupathy and Marius Zimand", title = "Power balance and apportionment algorithms for the {United States Congress}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "1:1--1:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297106", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We measure the performance, in the task of apportioning the Congress of the United States, of an algorithm combining a heuristic-driven (simulated annealing) search with an exact-computation dynamic programming evaluation of the apportionments visited in the search. We compare this with the actual algorithm currently used in the United States to apportion Congress, and with a number of other algorithms that have been proposed. We conclude that on every set of census data in this country's history, the heuristic-driven apportionment provably yields far fairer apportionments than those of any of the other algorithm considered, including the algorithm currently used by the United States for Congressional apportionment.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "apportionment algorithms; power indices; simulated annealing", } @Article{Cho:1998:WBL, author = "Seonghun Cho and Sartaj Sahni", title = "Weight-biased leftist trees and modified skip lists", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "2:1--2:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297111", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We propose the weight biased leftist tree as an alternative to traditional leftist trees [CRAN72] for the representation of mergeable priority queues. A modified version of skip lists [PUGH90] that uses fixed size nodes is also proposed. Experimental results show our modified skip list structure is faster than the original skip list structure for the representation of dictionaries. Experimental results comparing weight biased leftist trees and competing priority queue structures are presented.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dictionary; leftist trees; priority queues; skip lists", } @Article{Yan:1998:LBE, author = "Yong Yan and Xiaodong Zhang", title = "Lock bypassing: an efficient algorithm for concurrently accessing priority heaps", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "3:1--3:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297116", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The heap representation of priority queues is one of the most widely used data structures in the design of parallel algorithms. Efficiently exploiting the parallelism of a priority heap has significant influence on the efficiency of a wide range of applications and parallel algorithms. In this paper, we propose an aggressive priority heap operating algorithm, called the lock bypassing algorithm (LB) on shared memory systems. This algorithm minimizes interference of concurrent enqueue and dequeue operations on priority heaps while keeping the strict priority property: a dequeue always returns the minimum of a heap. The unique idea that distinguishes the LB algorithm from previous concurrent algorithms on priority heaps is the use of locking-on-demand and lock-bypassing techniques to minimize locking granularity and to maximize parallelism. The LB algorithm allows an enqueue operation to bypass the locks along its insertion path until it reaches a possible place where it can perform the insertion. Meanwhile a dequeue operation also makes its locking range and locking period as small as possible by carefully tuning its execution procedure. The LB algorithm is shown to be correct in terms of deadlock freedom and heap consistency. The performance of the LB algorithm was evaluated analytically and experimentally in comparison with previous algorithms. Analytical results show that the LB algorithm reduces by half the number of locks waited for in the worst case by previous algorithms. The experimental results show that the LB algorithm outperforms previously designed algorithms by up to a factor of 2 in hold time.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "aggressive locking; parallel algorithm; performance evaluation; priority heap; shared-memory system", } @Article{Helman:1998:NDP, author = "David R. Helman and Joseph J{\'a}J{\'a} and David A. Bader", title = "A new deterministic parallel sorting algorithm with an experimental evaluation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "4:1--4:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297128", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We introduce a new deterministic parallel sorting algorithm for distributed memory machines based on the regular sampling approach. The algorithm uses only two rounds of regular all-to-all personalized communication in a scheme that yields very good load balancing with virtually no overhead. Moreover, unlike previous variations, our algorithm efficiently handles the presence of duplicate values without the overhead of tagging each element with a unique identifier. This algorithm was implemented in SPLIT-C and run on a variety of platforms, including the Thinking Machines CM-5, the IBM SP-2-WN, and the Cray Research T3D. We ran our code using widely different benchmarks to examine the dependence of our algorithm on the input distribution. Our experimental results illustrate the efficiency and scalability of our algorithm across different platforms. In fact, the performance compares closely to that of our random sample sort algorithm, which seems to outperform all similar algorithms known to the authors on these platforms. Together, their performance is nearly invariant over the set of input distributions, unlike previous efficient algorithms. However, unlike our randomized sorting algorithm, the performance and memory requirements of our regular sorting algorithm can be deterministically guaranteed.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "generalized sorting; integer sorting; parallel algorithms; parallel performance; sorting by regular sampling", } @Article{Frigioni:1998:EAD, author = "Daniele Frigioni and Mario Ioffreda and Umberto Nanni and Giulio Pasqualone", title = "Experimental analysis of dynamic algorithms for the single", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "5:1--5:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297147", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper we propose the first experimental study of the fully dynamic single-source shortest-paths problem on directed graphs with positive real edge weights. In particular, we perform an experimental analysis of three different algorithms: Dijkstra's algorithm, and the two output bounded algorithms proposed by Ramalingam and Reps in [30] and by Frigioni, Marchetti-Spaccamela and Nanni in [18], respectively. The main goal of this paper is to provide a first experimental evidence for: (a) the effectiveness of dynamic algorithms for shortest paths with respect to a traditional static approach to this problem; (b) the validity of the theoretical model of output boundedness to analyze dynamic graph algorithms. Beside random generated graphs, useful to capture the 'asymptotic' behavior of the algorithms, we also developed experiments by considering a widely used graph from the real world, i.e., the Internet graph.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dynamic algorithms; experimental analysis of algorithms; shortest paths", } @Article{Magun:1998:GMA, author = "Jakob Magun", title = "Greeding matching algorithms, an experimental study", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "6:1--6:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297131", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We conduct an experimental study of several greedy algorithms for finding large matchings in graphs. Further we propose a new graph reduction, called $k$-Block Reduction, and present two novel algorithms using extra heuristics in the matching step and $k$-Block Reduction for $k = 3$. Greedy matching algorithms can be used for finding a good approximation of the maximum matching in a graph $G$ if no exact solution is required, or as a fast preprocessing step to some other matching algorithm. The studied greedy algorithms run in $O(m)$. They are easy to implement and their correctness and their running time are simple to prove. Our experiments show that a good greedy algorithm looses on average at most one edge on random graphs from $G(n,p)$ with up to 10,000 vertices. Furthermore the experiments show for which edge densities in random graphs the maximum matching problem is difficult to solve.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Andersson:1998:IR, author = "Arne Andersson and Stefan Nilsson", title = "Implementing radixsort", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "7:1--7:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297136", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present and evaluate several optimization and implementation techniques for string sorting. In particular, we study a recently published radix sorting algorithm, Forward radixsort, that has a provably good worst-case behavior. Our experimental results indicate that radix sorting is considerably faster (often more than twice as fast) than comparison-based sorting methods. This is true even for small input sequences. We also show that it is possible to implement a radixsort with good worst-case running time without sacrificing average-case performance. Our implementations are competitive with the best previously published string sorting programs.", acknowledgement = ack-nhfb, articleno = "7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "adaptive radixsort; algorithms; forward radixsort; radix sorting; sorting; string sorting", } @Article{Cherkassky:1998:APC, author = "Boris V. Cherkassky and Andrew V. Goldberg and Paul Martin and Joao C. Setubal and Jorge Stolfi", title = "Augment or push: a computational study of bipartite matching and unit-capacity flow algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "8:1--8:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297140", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We conduct a computational study of unit capacity flow and bipartite matching algorithms. Our goal is to determine which variant of the push-relabel method is most efficient in practice and to compare push-relabel algorithms with augmenting path algorithms. We have implemented and compared three push-relabel algorithms, three augmenting-path algorithms (one of which is new), and one augment-relabel algorithm. The depth-first search augmenting path algorithm was thought to be a good choice for the bipartite matching problem, but our study shows that it is not robust (meaning that it is not consistently fast on all or most inputs). For the problems we study, our implementations of the FIFO and lowest-level selection push-relabel algorithms have the most robust asymptotic rate of growth and work best overall. Augmenting path algorithms, although not as robust, on some problem classes are faster by a moderate constant factor. Our study includes several new problem families and input graphs with as many as $5 \times 105$ vertices.", acknowledgement = ack-nhfb, articleno = "8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Radzik:1998:IDT, author = "Tomasz Radzik", title = "Implementation of dynamic trees with in-subtree operations", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "3", pages = "9:1--9:??", month = "????", year = "1998", CODEN = "????", DOI = "http://doi.acm.org/10.1145/297096.297144", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We describe an implementation of dynamic trees with 'in-subtree' operations. Our implementation follows Sleator and Tarjan's framework of dynamic-tree implementations based on splay trees. We consider the following two examples of 'in-subtree' operations. (a) For a given node v, find a node with the minimum key in the subtree rooted at v. (b) For a given node v, find a random node with key X in the subtree rooted at v (value X is fixed throughout the whole computation). The first operation may provide support for edge deletions in the dynamic minimum spanning tree problem. The second one may be useful in local search methods for degree-constrained minimum spanning tree problems. We conducted experiments with our dynamic-tree implementation within these two contexts, and the results suggest that this implementation may lead to considerably faster codes than straightforward approaches do.", acknowledgement = ack-nhfb, articleno = "9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; design; dynamic minimum spanning tree; dynamic trees; experimentation; performance; splay trees", } @Article{Burke:1999:MAS, author = "E. K. Burke and A. J. Smith", title = "A memetic algorithm to schedule planned maintenance for the national grid", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "1:1--1:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347801", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The combination of local search operators, problem specific information and a genetic algorithm has provided very good results in certain scheduling problems, particularly in timetabling and maintenance scheduling problems. The resulting algorithm from this hybrid approach has been termed a Memetic Algorithm. This paper investigates the use of such an algorithm for the scheduling of transmission line maintenance for a known problem that has been addressed in the literature using a combination of a genetic algorithm and greedy optimisers. This problem is concerned with the scheduling of maintenance for an electricity transmission network where every transmission line must be maintained once within a specified time period. The objective is to avoid situations where sections of the network are disconnected, and to minimise the overloading of lines which are in service. In this paper we look at scheduling maintenance for the South Wales region of the national transmission network. We present and discuss, in some detail, a memetic algorithm that incorporates local search operators including tabu search and simulated annealing. A comparison is made both with the results from previous work, and against a selection of optimising techniques. The approach presented in this paper shows a significant improvement over previously published results on previously tackled problems. We also present results on another problem which has not been tackled in the literature but which is closer to the real world maintenance scheduling problems faced by such companies as The National Grid Company plc using the South Wales region.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "heuristics; hill climbing; maintenance scheduling; memetic algorithms; simulated annealing; tabu search", } @Article{Kim:1999:NSP, author = "Sun Kim", title = "A new string-pattern matching algorithm using partitioning and hashing efficiently", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "2:1--2:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347803", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper, we present a new string-pattern matching algorithm that partitions the text into segments of the input pattern length and searches for pattern occurrences using a simple hashing scheme. Unlike the well known Boyer--Moore style algorithm, our algorithm does not compute variable shift length, thus providing a conceptually simpler way to search for patterns. Empirical evaluation shows that our algorithm runs significantly faster than Sunday's and Horspool's extensions of the Boyer--Moore algorithm. The notion of the non-occurrence heuristic used in our algorithm, together with a text partitioning scheme, leads to a simplified scheme for searching for pattern occurrences, thus yielding better run time performance.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Eiron:1999:MMC, author = "N. Eiron and M. Rodeh and I. Steinwarts", title = "Matrix multiplication: a case study of enhanced data cache utilization", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "3:1--3:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347806", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Modern machines present two challenges to algorithm engineers and compiler writers: They have superscalar, super-pipelined structure, and they have elaborate memory subsystems specifically designed to reduce latency and increase bandwidth. Matrix multiplication is a classical benchmark for experimenting with techniques used to exploit machine architecture and to overcome the limitations of contemporary memory subsystems. This research aims at advancing the state of the art of algorithm engineering by balancing instruction level parallelism, two levels of data tiling, copying to provably avoid any cache conflicts, and prefetching in parallel to computational operations, in order to fully exploit the memory bandwidth. Measurements on IBM's RS/6000 43P workstation show that the resultant matrix multiplication algorithm outperforms IBM's ESSL by 6.8-31.8\%, is less sensitive to the size of the input data, and scales better. In this paper we introduce a cache aware algorithm for matrix multiplication. We also suggest generic guidelines that may be applied to compute intensive algorithm to efficiently utilize the data cache. We believe that some of our concepts may be embodied in compilers.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; BLAS; blocking; cache; matrix multiplication; performance; prefetching", } @Article{Erlebach:1999:EIO, author = "T. Erlebach and K. Jansen", title = "Efficient implementation of an optimal greedy algorithm for wavelength assignment in directed tree networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "4:1--4:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347808", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In all-optical networks with wavelength-division multiplexing several connections can share a physical link if the signals are transmitted on different wavelengths. As the number of available wavelengths is limited in practice, it is important to find wavelength assignments minimizing the number of different wavelengths used. This path coloring problem is NP-hard, and the best known polynomial-time approximation algorithm for directed tree networks achieves approximation ratio $5 / 3$, which is optimal in the class of greedy algorithms for this problem. It is shown how the algorithm can be modified in order to improve its running-time to $O({\rm Tec}(N,L))$ for sets of paths with maximum load $L$ in trees with $N$ nodes, where ${\rm Tec}(n, k)$ is the time for edge-coloring a $k$-regular bipartite graph with n nodes. An implementation of this efficient version of the algorithm in C++ using the LEDA class library is described, and experimental results regarding the running-times and the number of wavelengths used are reported. An additional heuristic that reduces the number of wavelengths used in the average case while maintaining the worst-case bound of $5 L / 3$ is described.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; bipartite edge coloring; directed tree networks; experimentation; path coloring", } @Article{Huson:1999:HTR, author = "D. Huson and S. Nettles and K. Rice and T. Warnow and S. Yooseph", title = "Hybrid tree reconstruction methods", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "5:1--5:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347812", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A major computational problem in biology is the reconstruction of evolutionary trees for species sets, and accuracy is measured by comparing the topologies of the reconstructed tree and the model tree. One of the major debates in the field is whether large evolutionary trees can be even approximately accurately reconstructed from biomolecular sequences of realistically bounded lengths (up to about 2000 nucleotides) using standard techniques (polynomial-time distance methods, and heuristics for NP-hard optimization problems). Using both analytical and experimental techniques, we show that on large trees, the two most popular methods in systematic biology, Neighbor-Joining and Maximum Parsimony heuristics, as well as two promising methods introduced by theoretical computer scientists, are all likely to have significant errors in the topology reconstruction of the model tree. We also present a new general technique for combining outputs of different methods (thus producing hybrid methods), and show experimentally how one such hybrid method has better performance than its constituent parts.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Jacob:1999:CSR, author = "R. Jacob and M. Marathe and K. Nagel", title = "A computational study of routing algorithms for realistic transportation networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "6:1--6:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347814", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We carry out an experimental analysis of a number of shortest-path (routing) algorithms investigated in the context of the TRANSIMS (TRansportation ANalysis and SIMulation System) project. The main focus of the paper is to study how various heuristic as well as exact solutions and associated data structures affect the computational performance of the software developed for realistic transportation networks. For this purpose we have used a road network representing, with high degree of resolution, the Dallas Fort-Worth urban area. We discuss and experimentally analyze various one-to-one shortest-path algorithms. These include classical exact algorithms studied in the literature as well as heuristic solutions that are designed to take into account the geometric structure of the input instances. Computational results are provided to compare empirically the efficiency of various algorithms. Our studies indicate that a modified Dijkstra's algorithm is computationally fast and an excellent candidate for use in various transportation planning applications as well as ITS related technologies.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "design and analysis of algorithms; experimental analysis; network design; shortest-paths algorithms; transportation planning", } @Article{Muller-Hannemann:1999:IWM, author = "M. M{\"u}ller-Hannemann and A. Schwartz", title = "Implementing weighted $b$-matching algorithms: towards a flexible software design", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "7:1--7:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347815", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present a case study on the design of an implementation of a fundamental combinatorial optimization problem: weighted b-matching. Although this problem is well-understood in theory and efficient algorithms are known, only little experience with implementations is available. This study was motivated by the practical need for an efficient b-matching solver as a subroutine in our approach to a mesh refinement problem in computer-aided design (CAD).The intent of this paper is to demonstrate the importance of flexibility and adaptability in the design of complex algorithms, but also to discuss how such goals can be achieved for matching algorithms by the use of design patterns. Starting from the basis of the famous blossom algorithm we explain how to exploit in different ways the flexibility of our software design which allows an incremental improvement of efficiency by exchanging subalgorithms and data structures. In a comparison with a code by Miller and Pekny we also demonstrate that our implementation is even without fine-tuning very competitive. Our code is significantly faster, with improvement factors ranging between 15 and 466 on TSPLIB instances.", acknowledgement = ack-nhfb, articleno = "7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; b-matching; blossom algorithm; design patterns; experimentation; object-oriented design; software design", } @Article{Schwerdt:1999:CWT, author = "J. Schwerdt and M. Smid and J. Majhi and R. Janardan", title = "Computing the width of a three-dimensional point set: an experimental study", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "4", pages = "8:1--8:??", month = "????", year = "1999", CODEN = "????", DOI = "http://doi.acm.org/10.1145/347792.347816", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:02:52 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We describe a robust, exact, and efficient implementation of an algorithm that computes the width of a three-dimensional point set. The algorithm is based on efficient solutions to problems that are at the heart of computational geometry: three-dimensional convex hulls, point location in planar graphs, and computing intersections between line segments. The latter two problems have to be solved for planar graphs and segments on the unit sphere, rather than in the two-dimensional plane. The implementation is based on LEDA, and the geometric objects are represented using exact rational arithmetic.", acknowledgement = ack-nhfb, articleno = "8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "computational geometry; implementation; layered manufacturing; spherical geometry", } @Article{Eppstein:2000:FHC, author = "David Eppstein", title = "Fast hierarchical clustering and other applications of dynamic closest pairs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "1:1--1:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.351829", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We develop data structures for dynamic closest pair problems with arbitrary distance functions, that do not necessarily come from any geometric structure on the objects. Based on a technique previously used by the author for Euclidean closest pairs, we show how to insert and delete objects from an n-object set, maintaining the closest pair, in $O(n \log^2 n)$ time per update and $O(n)$ space. With quadratic space, we can instead use a quadtree-like structure to achieve an optimal time bound, $O(n)$ per update. We apply these data structures to hierarchical clustering, greedy matching, and TSP heuristics, and discuss other potential applications in machine learning, Gr{\"o}bner bases, and local improvement algorithms for partition and placement problems. Experiments show our new methods to be faster in practice than previously used heuristics.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "conga line data structure; matching; nearest-neighbor heuristic; quadtree; TSP", } @Article{Chong:2000:CBD, author = "Kyn-Rak Chong and Sartaj Sahni", title = "Correspondence-based data structures for double-ended priority queues", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "2:1--2:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.351828", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We describe three general methods--total, dual, and leaf correspondence--that may be used to derive efficient double-ended priority queues from single-ended priority queues. These methods are illustrated by developing double-ended priority queues based on the classical heap. Experimental results indicate that the leaf-correspondence method generally leads to a faster double-ended priority queue than either total or dual correspondence. On randomly generated test sets, however, the splay tree outperforms the tested correspondence-based double-ended priority queues.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "correspondence-based data structures; double-ended priority queues; heaps; leftist trees; runtime efficiency; splay trees", } @Article{Xiao:2000:IMP, author = "Li Xiao and Xiaodong Zhang and Stefan A. Kubricht", title = "Improving memory performance of sorting algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "3:1--3:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384245", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Memory hierarchy considerations during sorting algorithm design and implementation play an important role in significantly improving execution performance. Existing algorithms mainly attempt to reduce capacity misses on direct-mapped caches. To reduce other types of cache misses that occur in the more common set-associative caches and the TLB, we restructure the mergesort and quicksort algorithms further by integrating tiling, padding, and buffering techniques and by repartitioning the data set. Our study shows that substantial performance improvements can be obtained using our new methods.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "caches; memory performance; mergesort; quicksort; TLB", } @Article{Navarro:2000:FFS, author = "Gonzalo Navarro and Mathieu Raffinot", title = "Fast and flexible string matching by combining bit-parallelism and suffix automata", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "4:1--4:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384246", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The most important features of a string matching algorithm are its efficiency and its flexibility. Efficiency has traditionally received more attention, while flexibility in the search pattern is becoming a more and more important issue. Most classical string matching algorithms are aimed at quickly finding an exact pattern in a text, being Knuth--Morris--Pratt (KMP) and the Boyer--Moore (BM) family the most famous ones. A recent development uses deterministic 'suffix automata' to design new optimal string matching algorithms, e.g. BDM and TurboBDM. Flexibility has been addressed quite separately by the use of 'bit-parallelism', which simulates automata in their nondeterministic form by using bits and exploiting the intrinsic parallelism inside the computer word, e.g. the Shift-Or algorithm. Those algorithms are extended to handle classes of characters and errors in the pattern and/or in the text, their drawback being their inability to skip text characters. In this paper we merge bit-parallelism and suffix automata, so that a nondeterministic suffix automaton is simulated using bit-parallelism. The resulting algorithm, called BNDM, obtains the best from both worlds. It is much simpler to implement than BDM and nearly as simple as Shift-Or. It inherits from Shift-Or the ability to handle flexible patterns and from BDM the ability to skip characters. BNDM is 30\%-40\% faster than BDM and up to 7 times faster than Shift-Or. When compared to the fastest existing algorithms on exact patterns (which belong to the BM family), BNDM is from 20\% slower to 3 times faster, depending on the alphabet size. With respect to flexible pattern searching, BNDM is by far the fastest technique to deal with classes of characters and is competitive to search allowing errors. In particular, BNDM seems very adequate for computational biology applications, since it is the fastest algorithm to search on DNA sequences and flexible searching is an important problem in that area. As a theoretical development related to flexible pattern matching, we introduce a new automaton to recognize suffixes of patterns with classes of characters. To the best of our knowledge, this automaton has not been studied before.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Caldwell:2000:DIM, author = "Andrew E. Caldwell and Andrew B. Kahng and Igor L. Markov", title = "Design and implementation of move-based heuristics for {VLSI} hypergraph partitioning", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "5:1--5:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384247", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We summarize the techniques of implementing move-based hypergraph partitioning heuristics and evaluating their performance in the context of VLSI design applications. Our first contribution is a detailed software architecture, consisting of seven reusable components, that allows flexible, efficient and accurate assessment of the practical implications of new move-based algorithms and partitioning formulations. Our second contribution is an assessment of the modern context for hypergraph partitioning research for VLSI design applications. In particular, we discuss the current level of sophistication in implementation know-how and experimental evaluation, and we note how requirements for real-world partitioners --- if used as motivation for research --- should affect the evaluation of prospective contributions. Two 'implicit decisions' in the implementation of the Fiduccia-Mattheyses heuristic are used to illustrate the difficulty of achieving meaningful experimental evaluation of new algorithmic ideas.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; balanced min-cut hypergraph partitioning; experimentation; measurement; performance; VLSI CAD", } @Article{Levine:2000:FRC, author = "Matthew S. Levine", title = "Finding the right cutting planes for the {TSP}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "6:1--6:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384248", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Given an instance of the Traveling Salesman Problem (TSP), a reasonable way to get a lower bound on the optimal answer is to solve a linear programming relaxation of an integer programming formulation of the problem. These linear programs typically have an exponential number of constraints, but in theory they can be solved efficiently with the ellipsoid method as long as we have an algorithm that can take a solution and either declare it feasible or find a violated constraint. In practice, it is often the case that many constraints are violated, which raises the question of how to choose among them so as to improve performance. For the simplest TSP formulation it is possible to efficiently find all the violated constraints, which gives us a good chance to try to answer this question empirically. Looking at random two dimensional Euclidean instances and the large instances from TSPLIB, we ran experiments to evaluate several strategies for picking among the violated constraints. We found some information about which constraints to prefer, which resulted in modest gains, but were unable to get large improvements in performance.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; combinatorial optimization; cutting plane; experimentation; minimum cut; performance; traveling salesman problem", } @Article{Sanders:2000:FPQ, author = "Peter Sanders", title = "Fast priority queues for cached memory", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "7:1--7:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384249", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The cache hierarchy prevalent in todays high performance processors has to be taken into account in order to design algorithms that perform well in practice. This paper advocates the adaption of external memory algorithms to this purpose. This idea and the practical issues involved are exemplified by engineering a fast priority queue suited to external memory and cached memory that is based on k-way merging. It improves previous external memory algorithms by constant factors crucial for transferring it to cached memory. Running in the cache hierarchy of a workstation the algorithm is at least two times faster than an optimized implementation of binary heaps and 4-ary heaps for large inputs.", acknowledgement = ack-nhfb, articleno = "7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "cache; cache efficiency; data structure; external memory; heap; implementation; multi way merging; priority queue; secondary storage", } @Article{Muller-Hannemann:2000:IWM, author = "Matthias M{\"u}ller-Hannemann and Alexander Schwartz", title = "Implementing weighted $b$-matching algorithms: insights from a computational study", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "8:1--8:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384250", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present an experimental study of an implementation of weighted perfect b-matching based on the primal-dual blossom algorithm. Although this problem is well-understood in theory and efficient algorithms are known, only little experience with implementations is available. In this paper several algorithmic variants are compared on synthetic and application problem data of very sparse graphs. This study was motivated by the practical need for an efficient b-matching solver for the latter application, namely as a subroutine in our approach to a mesh refinement problem in computer-aided design (CAD).Linear regression and operation counting is used to analyze code variants. The experiments confirm that a fractional jump-start speeds up the algorithm, they indicate that a variant based on pairing heaps is slightly superior to a k-heap variant, and that scaling of large b-values is not necessary, whereas a delayed blossom shrinking heuristic significantly improves running times only for graphs with average degree two. The fastest variant of our implementation appears to be highly superior to a code by Miller and Pekny (1995).", acknowledgement = ack-nhfb, articleno = "8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "b-matching; blossom algorithm; operation counting", } @Article{Shibuya:2000:CSP, author = "Tetsuo Shibuya", title = "Computing the $n \times m$ shortest path efficiently", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "9:1--9:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384251", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Computation of all the shortest paths between multiple sources and multiple destinations on various networks is required in many problems, such as the traveling salesperson problem (TSP) and the vehicle routing problem (VRP). This paper proposes new algorithms that compute the set of shortest paths efficiently by using the A* algorithm. The efficiency and properties of these algorithms are examined by using the results of experiments on an actual road network.", acknowledgement = ack-nhfb, articleno = "9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "*{$<$} m shortest paths; algorithm; algorithms; A{$<$} experimentation; GIS; n \times /sup{$>$} sup>", } @Article{Vishkin:2000:ELR, author = "Dascal Vishkin and Uzi Vishkin", title = "Experiments with list ranking for explicit multi-threaded {(XMT)} instruction parallelism", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "10:1--10:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384252", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Algorithms for the problem of list ranking are empirically studied with respect to the Explicit Multi-Threaded (XMT) platform for instruction-level parallelism (ILP). The main goal of this study is to understand the differences between XMT and more traditional parallel computing implementation platforms/models as they pertain to the well studied list ranking problem. The main two findings are: (i) good speedups for much smaller inputs are possible and (ii) in part, the first finding is based on a new variant of a 1984 algorithm, called the No-Cut algorithm. The paper incorporates analytic (non-asymptotic) performance analysis into experimental performance analysis for relatively small inputs. This provides an interesting example where experimental research and theoretical analysis complement one another. Explicit Multi-Threading (XMT) is a fine-grained computation framework introduced in our SPAA'98 paper. Building on some key ideas of parallel computing, XMT covers the spectrum from algorithms through architecture to implementation; the main implementation related innovation in XMT was through the incorporation of low-overhead hardware and software mechanisms (for more effective fine-grained parallelism). The reader is referred to that paper for detail on these mechanisms. The XMT platform aims at faster single-task completion time by way of ILP.", acknowledgement = ack-nhfb, articleno = "10", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Werneck:2000:FMC, author = "Renato Werneck and Jo{\~a}o Setubal and Arlindo da Conceic{\~a}o", title = "Finding minimum congestion spanning trees", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "11:1--11:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384253", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Given a weighted graph $G = (V, E)$, a positive integer $k$, and a penalty function $w_p$, we want to find $k$ spanning trees on $G$, not necessarily disjoint, of minimum total weight, such that the weight of each edge is subject to a penalty given by $w_p$ if it belongs to more than one tree. The objective function to be minimized is $\sum_{e \in E} W_e(i_e)$, where $i_e$ is the number of times edge $e$ appears in the solution and $W_e(i_e) = i_e w_p(e, i_e)$ is the aggregate cost of using edge $e$ $i_e$ times. For the case when $W_e$ is weakly convex, which should have wide application in congestion problems, we present a polynomial time algorithm; the algorithm's complexity is quadratic in $k$. We also present two heuristics with complexity linear in $k$. In an experimental study we show that these heuristics are much faster than the exact algorithm also in practice. These experiments present a diverse combination of input families (four), varying $k$ (up to 1000), and penalty functions (two). In most inputs tested the solutions given by the heuristics were within 1\% of optimal or much better, especially for large $k$. The worst quality observed was 3.2\% of optimal.", acknowledgement = ack-nhfb, articleno = "11", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Schulz:2000:DAL, author = "Frank Schulz and Dorothea Wagner and Karsten Weihe", title = "{Dijkstra}'s algorithm on-line: an empirical case study from public railroad transport", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "12:1--12:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384254", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Traffic information systems are among the most prominent real-world applications of Dijkstra's algorithm for shortest paths. We consider the scenario of a central information server in the realm of public railroad transport on wide-area networks. Such a system has to process a large number of on-line queries for optimal travel connections in real time. In practice, this problem is usually solved by heuristic variations of Dijkstra's algorithm, which do not guarantee an optimal result. We report results from a pilot study, in which we focused on the travel time as the only optimization criterion. In this study, various speed-up techniques for Dijkstra's algorithm were analysed empirically. This analysis was based on the timetable data of all German trains and on a 'snapshot' of half a million customer queries.", acknowledgement = ack-nhfb, articleno = "12", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Flato:2000:DIP, author = "Eyal Flato and Dan Halperin and Iddo Hanniel and Oren Nechushtan and Eti Ezra", title = "The design and implementation of planar maps in {CGAL}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "13:1--13:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384255", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Planar maps are fundamental structures in computational geometry. They are used to represent the subdivision of the plane into regions and have numerous applications. We describe the planar map package of CGAL--a Computational Geometry Algorithms Library. We discuss its modular design and implementation. In particular we introduce the two main classes of the design--planar maps and topological maps that enable the convenient separation between geometry and topology. The modular design is implemented using a generic programming approach. By switching a template parameter--the geometric traits class, one can use the same code for planar maps of different objects such as line segments or circular arcs. More flexibility is achieved by choosing a point location algorithm out of three implemented algorithms or plugging in an algorithm implemented by the user. The user of the planar maps package can benefit both from its flexibility and robustness. We present several examples of geometric traits classes and point location algorithms which demonstrate the possibility to adapt the general package to specific needs.", acknowledgement = ack-nhfb, articleno = "13", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Rahman:2000:ACE, author = "Naila Rahman and Rajeev Raman", title = "Analysing cache effects in distribution sorting", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "14:1--14:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384256", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study cache effects in distribution sorting algorithms for sorting keys drawn independently at random from a uniform distribution (`uniform keys'). We note that the performance of a recently-published distribution sorting algorithm, Flashsort1, which sorts $n$ uniform floating-point keys in $O(n)$ expected time, does not scale well with the input size due to poor cache utilisation. We present an approximate analysis for distribution sorting uniform keys which, as validated by simulation results, predicts the expected cache misses of Flashsort1 quite well. Using this analysis, we design a multiple-pass variant of Flashsort1 which outperforms Flashsort1 and comparison-based algorithms on uniform floating-point keys for moderate to large values of $n$. Using experimental results we also show that the integer distribution sorting algorithm MSB radix sort performs well on both uniform integer and uniform floating-point keys.", acknowledgement = ack-nhfb, articleno = "14", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "cache; efficient sorting algorithms; external-memory algorithms; memory hierarchy", } @Article{Bojesen:2000:PEC, author = "Jesper Bojesen and Jyrki Katajainen and Maz Spork", title = "Performance engineering case study: heap construction", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "15:1--15:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384257", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The behaviour of three methods for constructing a binary heap on a computer with a hierarchical memory is studied. The methods considered are the original one proposed by Williams [1964], in which elements are repeatedly inserted into a single heap; the improvement by Floyd [1964], in which small heaps are repeatedly merged to bigger heaps; and a recent method proposed, e.g., by Fadel et al. [1999] in which a heap is built layerwise. Both the worst-case number of instructions and that of cache misses are analysed. It is well-known that Floyd's method has the best instruction count. Let N denote the size of the heap to be constructed, B the number of elements that fit into a cache line, and let c and d be some positive constants. Our analysis shows that, under reasonable assumptions, repeated insertion and layerwise construction both incur at most cN/B cache misses, whereas repeated merging, as programmed by Floyd, can incur more than (dN log2 B)/B cache misses. However, for our memory-tuned versions of repeated insertion and repeated merging the number of cache misses incurred is close to the optimal bound N/B. In addition to these theoretical findings, we communicate many practical experiences which we hope to be valuable for others doing experimental algorithmic work.", acknowledgement = ack-nhfb, articleno = "15", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; binary heaps; code tuning; experimentation; memory tuning; performance; theory", } @Article{Boghossian:2000:RSP, author = "N. P. Boghossian and O. Kohlbacher and H. P. Lenhof", title = "Rapid software prototyping in molecular modeling using the biochemical algorithms library {(BALL)}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "16:1--16:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384258", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In the next century, virtual laboratories will play a key role in biotechnology. Computer experiments will not only replace some of the time-consuming and expensive real-world experiments, but they will also provide insights that cannot be obtained using 'wet' experiments. The field that deals with the modeling of atoms, molecules, and their reactions is called Molecular Modeling. The advent of Life Sciences gave rise to numerous new developments in this area. However, the implementation of new simulation tools is extremely time-consuming. This is mainly due to the large amount of supporting code that is required in addition to the code necessary to implement the new idea. The only way to reduce the development time is to reuse reliable code, preferably using object-oriented approaches. We have designed and implemented BALL, the first object-oriented application framework for rapid prototyping in Molecular Modeling. By the use of the composite design pattern and polymorphism we were able to model the multitude of complex biochemical concepts in a well-structured and comprehensible class hierarchy, the BALL kernel classes. The isomorphism between the biochemical structures and the kernel classes leads to an intuitive interface. Since BALL was designed for rapid software prototyping, ease of use, extensibility, and robustness were our principal design goals. Besides the kernel classes, BALL provides fundamental components for import/export of data in various file formats, Molecular Mechanics simulations, three-dimensional visualization, and more complex ones like a numerical solver for the Poisson--Boltzmann equation.", acknowledgement = ack-nhfb, articleno = "16", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "libraries; molecular modeling; rapid software prototyping", } @Article{Brengel:2000:ESP, author = "Klaus Brengel and Andreas Crauser and Paolo Ferragina and Ulrich Meyer", title = "An experimental study of priority queues in external memory", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "5", pages = "17:1--17:??", month = "????", year = "2000", CODEN = "????", DOI = "http://doi.acm.org/10.1145/351827.384259", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:09 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper we compare the performance of eight different priority queue implementations: four of them are explicitly designed to work in an external-memory setting, the others are standard internal-memory queues available in the LEDA library [Mehlhorn and N{\"a}her 1999]. Two of the external-memory priority queues are obtained by engineering known internal-memory priority queues with the aim of achieving effective performance on external storage devices (i.e., Radix heaps [Ahuja et al. 1990] and array heaps [Thorup 1996]). Our experimental framework includes some simple tests, like random sequences of insert or delete-minimum operations, as well as more advanced tests consisting of intermixed sequences of update operations and 'application driven' update sequences originated by simulations of Dijkstra's algorithm on large graph instances. Our variegate spectrum of experimental results gives a good picture of the features of these priority queues, thus being helpful to anyone interested in the use of such data structures on very large data sets.", acknowledgement = ack-nhfb, articleno = "17", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Arge:2001:EAP, author = "Lars Arge and Laura Toma and Jeffrey Scott Vitter", title = "{I/O}-Efficient Algorithms for Problems on Grid-Based Terrains", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "1:1--1:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945395", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The potential and use of Geographic Information Systems is rapidly increasing due to the increasing availability of massive amounts of geospatial data from projects like NASA's Mission to Planet Earth. However, the use of these massive datasets also exposes scalability problems with existing GIS algorithms. These scalability problems are mainly due to the fact that most GIS algorithms have been designed to minimize internal computation time, while I/O communication often is the bottleneck when processing massive amounts of data. In this paper, we consider I/O-efficient algorithms for problems on grid-based terrains. Detailed grid-based terrain data is rapidly becoming available for much of the Earth's surface. We describe [EQUATION] I/O algorithms for several problems on [EQUATION] grids for which only $O(N)$ algorithms were previously known. Here $M$ denotes the size of the main memory and $B$ the size of a disk block. We demonstrate the practical merits of our work by comparing the empirical performance of our new algorithm for the {\em flow accumulation\/} problem with that of the previously best known algorithm. Flow accumulation, which models flow of water through a terrain, is one of the most basic hydrologic attributes of a terrain. We present the results of an extensive set of experiments on real-life terrain datasets of different sizes and characteristics. Our experiments show that while our new algorithm scales nicely with dataset size, the previously known algorithm 'breaks down' once the size of the dataset becomes bigger than the available main memory. For example, while our algorithm computes the flow accumulation for the Appalachian Mountains in about three hours, the previously known algorithm takes several weeks.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Demestrescu:2001:BCM, author = "Camil Demestrescu and Irene Finocchi", title = "Breaking cycles for minimizing crossings", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "2:1--2:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945396", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the one-sided crossing minimization problem (CP): given a bipartite graph $G$ and a permutation $x_0$ of the vertices on a layer, find a permutation $x_1$ of the vertices on the other layer which minimizes the number of edge crossings in any straightline drawing of $G$ where vertices are placed on two parallel lines and sorted according to $x_0$ and $x_1$. Solving CP represents a fundamental step in the construction of aesthetically pleasing layouts of hierarchies and directed graphs, but unfortunately this problem has been proved to be NP-complete.\par In this paper we address the strong relation between CP and the problem of computing minimum feedback arc sets in directed graphs and we devise a new approximation algorithm for CP, called PM, that exploits this dependency. We experimentally and visually compare the performance of PM with the performance of well-known algorithms and of recent attractive strategies. Experiments are carried out on different families of randomly generated graphs, on pathological instances, and on real test sets. Performance indicators include both number of edge crossings and running time, as well as structural measures of the problem instances. We found CP to be a very interesting and rich problem from a combinatorial point of view. Our results clearly separate the behavior of the algorithms, proving the effectiveness of PM on most test sets and showing tradeoffs between quality of the solutions and running time. However, if the visual complexity of the drawings is considered, we found no clear winner. This confirms the importance of optimizing also other aesthetic criteria such as symmetry, edge length, and angular resolution.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "bipartite graphs; crossing minimization; experimental algorithms", } @Article{Gabow:2001:NFB, author = "Harold Gabow and Tadayoshi Kohno", title = "A Network-Flow-Based Scheduler: Design, Performance History, and Experimental Analysis", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "3:1--3:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945397", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We describe a program that schedules physician attending teams at Denver Health Medical Center. The program uses network flow techniques to prune an exponentially sized search space. We describe the program design, its performance history at the hospital, and experiments on a simplified version of the program.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "graphs; matroids; scheduling", } @Article{Iyer:2001:ESP, author = "Raj Iyer and David Karger and Hariharan Rahul and Mikkel Thorup", title = "An Experimental Study of Polylogarithmic, Fully Dynamic, Connectivity Algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "4:1--4:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945398", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present an experimental study of different variants of the amortized $O(\log^n)$-time fully-dynamic connectivity algorithm of Holm, de Lichtenberg, and Thorup (STOC'98). The experiments build upon experiments provided by Alberts, Cattaneo, and Italiano (SODA'96) on the randomized amortized $O(\log^3 n)$ fully-dynamic connectivity algorithm of Henzinger and King (STOC'95). Our experiments shed light upon similarities and differences between the two algorithms. We also present a slightly modified version of the Henzinger--King algorithm that runs in $O(\log^2 n)$ time, which resulted from our experiments.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Liberatore:2001:CSB, author = "Vincenzo Liberatore", title = "Caching and Scheduling for Broadcast Disk Systems", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "5:1--5:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945399", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Unicast connections lead to performance and scalability problems when a large client population attempts to access the same data. Broadcast push and broadcast disk technology address the problem by broadcasting data items from a server to a large number of clients. Broadcast disk performance depends mainly on caching strategies at the client site and on how the broadcast is scheduled at the server site. An on-line broadcast disk paging strategy makes caching decisions without knowing future page requests or access probabilities. This paper gives new implementations of existing on-line algorithms and reports on extensive empirical investigations. The gray algorithm [Khanna and Liberatore 2000] always outperformed other on-line strategies on both synthetic and Web traces. Moreover, caching limited the skewness of broadcast schedules, and led to favor efficient caching algorithms over refined scheduling strategies when the cache was large.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "broadcast disk; caching; scheduling", } @Article{Narasimhan:2001:GMS, author = "Giri Narasimhan and Martin Zachariasen", title = "Geometric Minimum Spanning Trees via Well-Separated Pair Decompositions", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "6:1--6:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945400", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Let $S$ be a set of $n$ points in $\Re^d$. We present an algorithm that uses the well-separated pair decomposition and computes the minimum spanning tree of $S$ under any $L_p$ or polyhedral metric. A theoretical analysis shows that it has an expected running time of $O(n \log n)$ for uniform point distributions; this is verified experimentally. Extensive experimental results show that this approach is practical. Under a variety of input distributions, the resulting implementation is robust and performs well for points in higher dimensional space.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Rahman:2001:ARS, author = "Naila Rahman and Rajeev Raman", title = "Adapting Radix Sort to the Memory Hierarchy", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "7:1--7:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945401", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We demonstrate the importance of reducing misses in the translation-lookaside buffer (TLB) for obtaining good performance on modern computer architectures. We focus on least-significant bit first (LSB) radix sort, standard implementations of which make many TLB misses. We give three techniques which simultaneously reduce cache and TLB misses for LSB radix sort: reducing working set size, explicit block transfer and pre-sorting. We note that: \item All the techniques above yield algorithms whose implementations outperform optimised cache-tuned implementations of LSB radix sort and comparison-based sorting algorithms. The fastest running times are obtained by the pre-sorting approach and these are over twice as fast as optimised cache-tuned implementations of LSB radix sort and quicksort. Even the simplest optimisation, using the TLB size to guide the choice of radix in standard implementations of LSB radix sort, gives good improvements over cache-tuned algorithms. \item One of the pre-sorting algorithms and explicit block transfer make few cache and TLB misses in the worst case. This is not true of standard implementations of LSB radix sort. We also apply these techniques to the problem of permuting an array of integers, and obtain gains of over 30\% relative to the naive algorithm by using explicit block transfer.", acknowledgement = ack-nhfb, articleno = "7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "cache; efficient sorting algorithms; external-memory algorithms; locality of reference; memory hierarchy; radix sort; translation-lookaside buffer (TLB)", } @Article{Stallmann:2001:HES, author = "Matthias Stallmann and Franc Brglez and Debabrata Ghosh", title = "Heuristics, Experimental Subjects, and Treatment Evaluation in Bigraph Crossing Minimization", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "8:1--8:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945402", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The bigraph crossing problem, embedding the two node sets of a bipartite graph along two parallel lines so that edge crossings are minimized, has applications to circuit layout and graph drawing. Experimental results for several previously known and two new heuristics suggest continued exploration of the problem, particularly sparse instances. We emphasize careful design of experimental subject classes and present novel views of the results. All source code, data, and scripts are available on-line", acknowledgement = ack-nhfb, articleno = "8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "crossing number; design of experiments; graph drawing; graph embedding; graph equivalence classes; layout", } @Article{Frigioni:2001:ESD, author = "Daniele Frigioni and Tobias Miller and Christos Zaroliagis", title = "An Experimental Study of Dynamic Algorithms for Transitive Closure", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "9:1--9:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945403", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We perform an extensive experimental study of several dynamic algorithms for transitive closure. In particular, we implemented algorithms given by Italiano, Yellin, Cicerone et al., and two recent randomized algorithms by Henzinger and King. We propose a fine-tuned version of Italiano's algorithms as well as a new variant of them, both of which were always faster than any of the other implementations of the dynamic algorithms. We also considered simple-minded algorithms that were easy to implement and likely to be fast in practice. Wetested and compared the above implementations on random inputs, on non-random inputs that are worst-case inputs for the dynamic algorithms, and on an input motivated by a real-world graph.", acknowledgement = ack-nhfb, articleno = "9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dynamic algorithm; experimentation; transitive closure", } @Article{Matias:2001:EFP, author = "Yossi Matias and Nasir Rajpoot and Cenk Sahinalp", title = "The Effect of Flexible Parsing for Dynamic Dictionary-Based Data Compression", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "6", pages = "10:1--10:??", month = "????", year = "2001", CODEN = "????", DOI = "http://doi.acm.org/10.1145/945394.945404", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:03:55 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We report on the performance evaluation of greedy parsing with a single step lookahead (which we call flexible Parsing or {\em FP\/}) as an alternative to the commonly used greedy parsing (with no-lookaheads) scheme. Greedy parsing is the basis of most popular compression programs including UNIX {\tt compress} and {\tt gzip}, however it usually results in far from optimal parsing\slash compression with regard to the dictionary construction scheme in use. Flexible parsing, however, is optimal [MS99], i.e. partitions any given input to the smallest number of phrases possible, for dictionary construction schemes which satisfy the prefix property throughout their execution.\par We focus on the application of {\em FP\/} in the context of the LZW variant of the Lempel--Ziv'78 dictionary construction method [Wel84, ZL78], which is of considerable practical interest. We implement two compression algorithms which use (1) {\em FP\/} with LZW dictionary (LZW-{\em FP\/}), and (2) {\em FP\/} with an alternative flexible dictionary (FPA as introduced in [Hor95]). Our implementations are based on novel on-line data structures enabling us to use linear time and space. We test our implementations on a collection of input sequences which includes textual files, DNA sequences, medical images, and pseudorandom binary files, and compare our results with two of the most popular compression programs UNIX {\tt compress} and {\tt gzip}. Our results demonstrate that flexible parsing is especially useful for non-textual data, on which it improves over the compression rates of {\tt compress} and {\tt gzip} by up to 20\% and 35\%, respectively.", acknowledgement = ack-nhfb, articleno = "10", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Backes:2002:HLB, author = "Werner Backes and Susanne Wetzel", title = "Heuristics on lattice basis reduction in practice", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "1--1", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944619", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper we provide a survey on LLL lattice basis reduction in practice. We introduce several new heuristics as to speed up known lattice basis reduction methods and improve the quality of the computed reduced lattice basis in practice. We analyze substantial experimental data and to our knowledge, we are the first to present general heuristics for determining which variant of the reduction algorithm, for varied parameter choices, yields the most efficient reduction strategy for reducing a particular problem instance.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dynamic approximation; general reduction heuristics; lattice basis reduction; modular and iterative heuristics", } @Article{Iwama:2002:PLS, author = "Kazuo Iwama and Daisuke Kawai and Shuichi Miyazaki and Yasuo Okabe and Jun Umemoto", title = "Parallelizing local search for {CNF} satisfiability using vectorization and {PVM}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "2--2", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944620", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The purpose of this paper is to speed up the local search algorithm for the CNF Satisfiability problem. Our basic strategy is to run some 10$^5$ independent search paths simultaneously using PVM on a vector supercomputer VPP800, which consists of 40 vector processors. Using the above parallelization and vectorization together with some improvement of data structure, we obtained 600-times speedup in terms of the number of flips the local search can make per second, compared to the original GSAT by Selman and Kautz. We ran our parallel GSAT for benchmark instances and compared the running time with those of existing SAT programs. We could observe an apparent benefit of parallelization: Especially, we were able to solve two instances that have never been solved before this paper. We also tested parallel local search for the SAT encoding of the class scheduling problem. Again we were able to get almost the best answer in reasonable time.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; CNF Satisfiability; distributed computing; experimentation; local search algorithms; parallelization; PVM; vector supercomputer; vectorization", } @Article{Albers:2002:ESO, author = "Susanne Albers and Bianca Schr{\"o}der", title = "An experimental study of online scheduling algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "3--3", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944621", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present the first comprehensive experimental study of online algorithms for Graham's scheduling problem. Graham's scheduling problem is a fundamental problem in scheduling theory where a sequence of jobs has to be scheduled on $m$ identical parallel machines so as to minimize the makespan. Graham gave an elegant algorithm that is $(2 - 1 / m)$-competitive. Recently a number of new online algorithms were developed that achieve competitive ratios around 1.9. Since competitive analysis can only capture the worst case behavior of an algorithm a question often asked is: Are these new algorithms geared only towards a pathological case or do they perform better in practice, too?We address this question by analyzing the algorithms on various job sequences. In our actual tests, we analyzed the algorithms (1) on real world jobs and (2) on jobs generated by probability distributions. It turns out that the performance of the algorithms depends heavily on the characteristics of the respective work load. On job sequences that are generated by standard probability distributions, Graham's strategy is clearly the best. However, on the real world jobs the new algorithms often outperform Graham's strategy. Our experimental study confirms theoretical results in the sense that there are also job sequences in practice on which the new online algorithms perform better. Our study can help to inform practitioners about the new scheduling strategies as an alternative to Graham's algorithm.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; experimentation; online algorithms; performance; scheduling", } @Article{Mehlhorn:2002:IWM, author = "Kurt Mehlhorn and Guido Sch{\"a}fer", title = "Implementation of {$O(nm \log n)$} weighted matchings in general graphs: the power of data structures", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "4--4", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944622", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We describe the implementation of an algorithm which solves the weighted matching problem in general graphs with $n$ vertices and $m$ edges in time $O(nm \log n)$. Our algorithm is a variant of the algorithm of Galil, Micali and Gabow [Galil et al. 1986] and extensively uses sophisticated data structures, in particular {\em concatenable priority queues}, so as to reduce the time needed to perform dual adjustments and to find tight edges in Edmonds' blossom-shrinking algorithm. We compare our implementation to the experimentally fastest implementation, named {\em Blossom IV}, due to Cook and Rohe [Cook and Rohe 1997]. Blossom IV requires only very simple data structures and has an asymptotic running time of $O(n^2 m)$. Our experiments show that our new implementation is superior to Blossom IV. A closer inspection reveals that the running time of Edmonds' blossom-shrinking algorithm in practice heavily depends on the time spent to perform dual adjustments and to find tight edges. Therefore, optimizing these operations, as is done in our implementation, indeed speeds-up the practical performance of implementations of Edmonds' algorithm.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Edelkamp:2002:IHQ, author = "Stefan Edelkamp and Patrick Stiegeler", title = "Implementing {{\em HEAPSORT\/}} with $(n \log n - 0.9 n)$ and {{\em QUICKSORT}\/} with $(n \log n + 0.2 n)$ comparisons", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "5--5", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944623", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "With refinements to the {\em WEAK-HEAPSORT\/} algorithm we establish the general and practical relevant sequential sorting algorithm {\em INDEX-WEAK-HEAPSORT\/} with exactly $n \lceil \log n \rceil - 2^{\lceil \log n \rceil} + 1 \leq n \log n 0.9 n$ comparisons and at most $n \log n + 0.1 n$ transpositions on any given input. It comprises an integer array of size $n$ and is best used to generate an index for the data set. With {\em RELAXED-WEAK-HEAPSORT\/} and {\em GREEDY-WEAK-HEAPSORT\/} we discuss modifications for a smaller set of pending element transpositions. If extra space to create an index is not available, with {\em QUICK-WEAK-HEAPSORT\/} we propose an efficient {\em QUICKSORT\/} variant with $n \log n + 0.2 n + o(n)$ comparisons on the average. Furthermore, we present data showing that {\em WEAK-HEAPSORT, INDEX-WEAK-HEAPSORT\/} and {\em QUICK-WEAK-HEAPSORT\/} compete with other performant {\em QUICKSORT\/} and {\em HEAPSORT\/} variants.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Erlebach:2002:IAA, author = "Thomas Erlebach and Klaus Jansen", title = "Implementation of approximation algorithms for weighted and unweighted edge-disjoint paths in bidirected trees", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "6--6", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944624", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Given a set of weighted directed paths in a bidirected tree, the maximum weight edge-disjoint paths problem (MWEDP) is to select a subset of the given paths such that the selected paths are edge-disjoint and the total weight of the selected paths is maximized. MWEDP is {\em NP\/}-hard for bidirected trees of unbounded degree, even if all weights are the same (the unweighted case). Three different approximation algorithms are implemented: a known combinatorial $(5/3 + \epsilon)$-approximation algorithm $A_1$ for the unweighted case, a new combinatorial 2-approximation algorithm $A_2$ for the weighted case, and a known $(5 / 3 + \epsilon)$-approximation algorithm $A_3$ for the weighted case that is based on linear programming. For algorithm $A_1$, it is shown how efficient data structures can be used to obtain a worst-case running-time of $O(m + n + 4^{1/\epsilon} \sqrt n c m)$ for instances consisting of $m$ paths in a tree with $n$ nodes. Experimental results regarding the running-times and the quality of the solutions obtained by the three approximation algorithms are reported. Where possible, the approximate solutions are compared to the optimal solutions, which were computed by running CPLEX on an integer linear programming formulation of MWEDP.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; combinatorial optimization; experimentation; linear programming", } @Article{Lassous:2002:PLR, author = "Isabelle Gu{\'e}rin Lassous and Jens Gustedt", title = "Portable list ranking: an experimental study", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "7--7", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944625", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present and analyze two portable algorithms for the List Ranking Problem in the Coarse Grained Multicomputer model (CGM). We report on implementations of these algorithms and experiments that were done with these on a variety of parallel and distributed architectures, ranging from PC clusters to a mainframe parallel machine. With these experiments, we validate the chosen CGM model, and also show the possible gains and limits of such algorithms.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Vahrenhold:2002:PPL, author = "Jan Vahrenhold and Klaus H. Hinrichs", title = "Planar point location for large data sets: to seek or not to seek", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "8--8", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944626", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present an algorithm for external memory planar point location that is both effective and easy to implement. The base algorithm is an external memory variant of the bucket method by Edahiro, Kokubo and Asano that is combined with Lee and Yang's batched internal memory algorithm for planar point location. Although our algorithm is not optimal in terms of its worst-case behavior, we show its efficiency for both batched and single-shot queries by experiments with real-world data. The experiments show that the algorithm benefits from the mainly sequential disk access pattern and significantly outperforms the fastest algorithm for internal memory. Due to its simple concept, the algorithm can take advantage of multiple disks and processors in a rather straightforward yet efficient way.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Wickremesinghe:2002:ESU, author = "Rajiv Wickremesinghe and Lars Arge and Jeffrey S. Chase and Jeffrey Scott Vitter", title = "Efficient sorting using registers and caches", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "9--9", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944627", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Modern computer systems have increasingly complex memory systems. Common machine models for algorithm analysis do not reflect many of the features of these systems, e.g., large register sets, lockup-free caches, cache hierarchies, associativity, cache line fetching, and streaming behavior. Inadequate models lead to poor algorithmic choices and an incomplete understanding of algorithm behavior on real machines. A key step toward developing better models is to quantify the performance effects of features not reflected in the models. This paper explores the effect of memory system features on sorting performance. We introduce a new cache-conscious sorting algorithm, R-MERGE, which achieves better performance in practice over algorithms that are superior in the theoretical models. R-MERGE is designed to minimize memory stall cycles rather than cache misses by considering features common to many system designs.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Herrmann:2002:FCN, author = "Francine Herrmann and Alain Hertz", title = "Finding the chromatic number by means of critical graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "10--10", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944628", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We propose a new exact algorithm for finding the chromatic number of a graph $G$. The algorithm attempts to determine the smallest possible induced subgraph $G'$ of $G$ which has the same chromatic number as $G$. Such a subgraph is said critical since all proper induced sub-graph of $G'$ have a chromatic number strictly smaller than $G'$. The proposed method is particularly helpful when a $k$-coloring of a non-critical graph is known, and it has to be proved that no $(k - 1)$-coloring of $G$ exists. Computational experiments on random graphs and on DIMACS benchmark problems demonstrate that the new proposed algorithm can solve larger problem than previous known exact methods.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; experimentation; performance", } @Article{Fekete:2002:SHP, author = "S{\'a}ndor P. Fekete and Henk Meijer and Andr{\'e} Rohe and Walter Tietze", title = "Solving a 'Hard' problem to approximate an 'Easy' one: heuristics for maximum matchings and maximum traveling salesman problems", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "11--11", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944629", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider geometric instances of the Maximum Weighted Matching Problem (MWMP) and the Maximum Traveling Salesman Problem (MTSP) with up to 3,000,000 vertices. Making use of a geometric duality relationship between MWMP, MTSP, and the Fermat--Weber-Problem (FWP), we develop a heuristic approach that yields in near-linear time solutions as well as upper bounds. Using various computational tools, we get solutions within considerably less than 1\% of the optimum. An interesting feature of our approach is that, even though an FWP is hard to compute in theory and Edmonds' algorithm for maximum weighted matching yields a polynomial solution for the MWMP, the practical behavior is just the opposite, and we can solve the FWP with high accuracy in order to find a good heuristic solution for the MWMP.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "approximation; Fermat--Weber problem; geometric optimization; geometric problems; heuristics; maximum traveling salesman problem (MTSP); maximum weighted matching; near-linear algorithms", } @Article{Neri:2002:RCL, author = "Filippo Neri", title = "Relational concept learning by cooperative evolution", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "7", pages = "12--12", month = "????", year = "2002", CODEN = "????", DOI = "http://doi.acm.org/10.1145/944618.944630", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Concept learning is a computationally demanding task that involves searching large hypothesis spaces containing candidate descriptions. Stochastic search combined with parallel processing provide a promising approach to successfully deal with such computationally intensive tasks. Learning systems based on distributed genetic algorithms (GA) were able to find concept descriptions as accurate as the ones found by state-of-the-art learning systems based on alternative approaches. However, genetic algorithms' exploitation has the drawback of being computationally demanding. We show how a suitable architectural choice, named cooperative evolution, allows to solve complex applications in an acceptable user waiting time and with a reasonable computational cost by using GA-based learning systems because of the effective exploitation of distributed computation. A variety of experimental settings is analyzed and an explanation for the empirical observations is proposed.", acknowledgement = ack-nhfb, fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "distributed genetic algorithm; first order logic concept learning; relational concept learning", } @Article{Kumar:2003:AME, author = "Piyush Kumar and Joseph S. B. Mitchell and E. Alper Yildirim", title = "Approximate minimum enclosing balls in high dimensions using core-sets", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "8", pages = "1.1:1--1.1:??", month = "????", year = "2003", CODEN = "????", DOI = "http://doi.acm.org/10.1145/996546.996548", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:56 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study the minimum enclosing ball (MEB) problem for sets of points or balls in high dimensions. Using techniques of second-order cone programming and 'core-sets', we have developed $(1 + \epsilon)$-approximation algorithms that perform well in practice, especially for very high dimensions, in addition to having provable guarantees. We prove the existence of core-sets of size $O(1/\epsilon)$, improving the previous bound of $O(1/\epsilon^2)$, and we study empirically how the core-set size grows with dimension. We show that our algorithm, which is simple to implement, results in fast computation of nearly optimal solutions for point sets in much higher dimension than previously computable using exact techniques.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "approximation algorithms; minimum enclosing ball; second-order cone programming", } @Article{Arge:2003:EPL, author = "Lars Arge and Andrew Danner and Sha-Mayn Teh", title = "{I/O}-efficient point location using persistent {B}-trees", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "8", pages = "1.2:1--1.2:??", month = "????", year = "2003", CODEN = "????", DOI = "http://doi.acm.org/10.1145/996546.996549", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:56 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present an external planar point location data structure that is I/O-efficient both in theory and practice. The developed structure uses linear space and answers a query in optimal $O(\log B N)$ I/Os, where $B$ is the disk block size. It is based on a persistent B-tree, and all previously developed such structures assume a total order on the elements in the structure. As a theoretical result of independent interest, we show how to remove this assumption. Most previous theoretical I/O-efficient planar point location structures are relatively complicated and have not been implemented. Based on a bucket approach, Vahrenhold and Hinrichs therefore developed a simple and practical, but theoretically non-optimal, heuristic structure. We present an extensive experimental evaluation that shows that, on a range of real-world Geographic Information Systems (GIS) data, our structure uses a similar number of I/Os as the structure of Vahrenhold and Hinrichs to answer a query. On a synthetically generated worst-case dataset our structure uses significantly fewer I/Os.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Buchsbaum:2003:FPM, author = "Adam L. Buchsbaum and Glenn S. Fowler and Balachannder Kirishnamurthy and Kiem-Phong Vo and Jia Wang", title = "Fast prefix matching of bounded strings", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "8", pages = "1.3:1--1.3:??", month = "????", year = "2003", CODEN = "????", DOI = "http://doi.acm.org/10.1145/996546.996550", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:56 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Longest Prefix Matching (LPM) is the problem of finding which string from a given set is the longest prefix of another, given string. LPM is a core problem in many applications, including IP routing, network data clustering, and telephone network management. These applications typically require very fast matching of bounded strings, i.e., strings that are short and based on small alphabets. We note a simple correspondence between bounded strings and natural numbers that maps prefixes to nested intervals so that computing the longest prefix matching a string is equivalent to finding the shortest interval containing its corresponding integer value. We then present {\em retries}, a fast and compact data structure for LPM on general alphabets. Performance results show that retries often outperform previously published data structures for IP look-up. By extending LPM to general alphabets, retries admit new applications that could not exploit prior LPM solutions designed for IP look-ups.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "IP routing; prefix matching; table look-up; tries", } @Article{Breimer:2003:LAL, author = "Eric A. Breimer and Mark K. Goldberg and Darren T. Lim", title = "A learning algorithm for the longest common subsequence problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "8", pages = "2.1:1--2.1:??", month = "????", year = "2003", CODEN = "????", DOI = "http://doi.acm.org/10.1145/996546.996552", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:56 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present an experimental study of a learning algorithm for the longest common subsequence problem, {\em LCS}. Given an arbitrary input domain, the algorithm learns an {\em LCS\/}-procedure tailored to that domain. The learning is done with the help of an oracle, which can be any {\em LCS\/}-algorithm. After solving a limited number of training inputs using an oracle, the learning algorithm outputs a new {\em LCS\/}-procedure. Our experiments demonstrate that, by allowing a slight loss of optimality, learning yields a procedure which is significantly faster than the oracle. The oracle used for the experiments is the {\em np\/}-procedure by Wu {\em et al.}, a modification of Myers' classical {\em LCS\/}-algorithm. We show how to scale up the results of learning on small inputs to inputs of arbitrary lengths. For the domain of two random 2-symbol inputs of length $n$, learning yields a program with 0.999 expected accuracy, which runs in $O(n^{1.41})$-time, in contrast with $O(n^2 \log n)$ running time of the fastest theoretical algorithm that produces optimal solutions. For the domain of random 2-symbol inputs of length 100,000, the program runs 10.5 times faster than the {\em np\/}-procedure, producing 0.999- accurate outputs. The scaled version of the evolved algorithm applied to random inputs of length 1 million runs approximately 30 times faster than the {\em np\/}-procedure while constructing 0.999- accurate solutions. We apply the evolved algorithm to DNA sequences of various lengths by training on random 4-symbol sequences of up to length 10,000. The evolved algorithm, scaled up to the lengths of up to 1.8 million, produces solutions with the 0.998-accuracy in a fraction of the time used by the {\em np}.", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Venkataraman:2003:BAP, author = "Gayathri Venkataraman and Sartaj Sahni and Srabani Mukhopadhyaya", title = "A blocked all-pairs shortest-paths algorithm", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "8", pages = "2.2:1--2.2:??", month = "????", year = "2003", CODEN = "????", DOI = "http://doi.acm.org/10.1145/996546.996553", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:56 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We propose a blocked version of Floyd's all-pairs shortest-paths algorithm. The blocked algorithm makes better utilization of cache than does Floyd's original algorithm. Experiments indicate that the blocked algorithm delivers a speedup (relative to the unblocked Floyd's algorithm) between 1.6 and 1.9 on a Sun Ultra Enterprise 4000/5000 for graphs that have between 480 and 3200 vertices. The measured speedup on an SGI O2 for graphs with between 240 and 1200 vertices is between 1.6 and 2.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "all pairs shortest paths; blocking; cache; speedup", } @Article{Petit:2003:EML, author = "Jordi Petit", title = "Experiments on the minimum linear arrangement problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "8", pages = "2.3:1--2.3:??", month = "????", year = "2003", CODEN = "????", DOI = "http://doi.acm.org/10.1145/996546.996554", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:04:56 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "This paper deals with the Minimum Linear Arrangement problem from an experimental point of view. Using a testsuite of sparse graphs, we experimentally compare several algorithms to obtain upper and lower bounds for this problem. The algorithms considered include Successive Augmentation heuristics, Local Search heuristics and Spectral Sequencing. The testsuite is based on two random models and 'real life' graphs. As a consequence of this study, two main conclusions can be drawn: On one hand, the best approximations are usually obtained using Simulated Annealing, which involves a large amount of computation time. Solutions found with Spectral Sequencing are close to the ones found with Simulated Annealing and can be obtained in significantly less time. On the other hand, we notice that there exists a big gap between the best obtained upper bounds and the best obtained lower bounds. These two facts together show that, in practice, finding lower and upper bounds for the Minimum Linear Arrangement problem is hard.", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Brandes:2004:GNC, author = "Ulrik Brandes and Frank Schulz and Dorothea Wagner and Thomas Willhalm", title = "Generating node coordinates for shortest-path computations in transportation networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "9", pages = "1.1:1--1.1:??", month = "????", year = "2004", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1005813.1005815", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:22 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Speed-up techniques that exploit given node coordinates have proven useful for shortest-path computations in transportation networks and geographic information systems. To facilitate the use of such techniques when coordinates are missing from some, or even all, of the nodes in a network we generate artificial coordinates using methods from graph drawing. Experiments on a large set of German train timetables indicate that the speed-up achieved with coordinates from our drawings is close to that achieved with the true coordinates---and in some special cases even better.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "graph drawing; shortest paths; transportation networks; travel planning", } @Article{Niewiadomski:2004:PSD, author = "Robert Niewiadomski and Jos{\'e} Nelson Amaral and Robert C. Holte", title = "A performance study of data layout techniques for improving data locality in refinement-based pathfinding", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "9", pages = "1.2:1--1.2:??", month = "????", year = "2004", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1005813.1041511", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:22 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The widening gap between processor speed and memory latency increases the importance of crafting data structures and algorithms to exploit temporal and spatial locality. Refinement-based pathfinding algorithms, such as Classic Refinement (CR), find quality paths in very large sparse graphs where traditional search techniques fail to generate paths in acceptable time. In this paper, we present a performance evaluation study of three simple data structure transformations aimed at improving the data reference locality of CR. These transformations are robust to changes in computer architecture and the degree of compiler optimization. We test our alternative designs on four contemporary architectures, using two compilers for each machine. In our experiments, the application of these techniques results in performance improvements of up to 67\% with consistent improvements above 15\%. Analysis reveals that these improvements stem from improved data reference locality at the page level and to a lesser extent at the cache line level.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "cache-conscious algorithms; classical refinement; pathfinding", } @Article{Marathe:2004:ESS, author = "Madhav V. Marathe and Alessandro Panconesi and Larry D. {Risinger, Jr.}", title = "An experimental study of a simple, distributed edge-coloring algorithm", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "9", pages = "1.3:1--1.3:??", month = "????", year = "2004", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1005813.1041515", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:22 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We conduct an experimental analysis of a distributed randomized algorithm for edge coloring simple undirected graphs. The algorithm is extremely simple yet, according to the probabilistic analysis, it computes nearly optimal colorings very quickly [Grable and Panconesi 1997]. We test the algorithm on a number of random as well as nonrandom graph families. The test cases were chosen based on two objectives: (i) to provide insights into the worst-case behavior (in terms of time and quality) of the algorithm and (ii) to test the performance of the algorithm with instances that are likely to arise in practice. Our main results include the following:(1) The empirical results obtained compare very well with the recent empirical results reported by other researchers [Durand et al. 1994, 1998; Jain and Werth 1995].(2) The empirical results confirm the bounds on the running time and the solution quality as claimed in the theoretical paper. Our results show that for certain classes of graphs the algorithm is likely to perform much better than the analysis suggests.(3) The results demonstrate that the algorithm might be well suited (from a theoretical as well as practical standpoint) for edge coloring graphs quickly and efficiently in a distributed setting. Based on our empirical study, we propose a simple modification of the original algorithm with substantially improved performance in practice.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "distributed algorithms; edge coloring; experimental analysis of algorithms; high performance computing; randomized algorithms; scheduling", } @Article{Fredriksson:2004:AOS, author = "Kimmo Fredriksson and Gonzalo Navarro", title = "Average-optimal single and multiple approximate string matching", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "9", pages = "1.4:1--1.4:??", month = "????", year = "2004", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1005813.1041513", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:22 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present a new algorithm for multiple approximate string matching. It is based on reading backwards enough l-grams from text windows so as to prove that no occurrence can contain the part of the window read, and then shifting the window. We show analytically that our algorithm is optimal on average. Hence our first contribution is to fill an important gap in the area, since no average-optimal algorithm existed for multiple approximate string matching. We consider several variants and practical improvements to our algorithm, and show experimentally that they are resistant to the number of patterns and the fastest for low difference ratios, displacing the long-standing best algorithms. Hence our second contribution is to give a practical algorithm for this problem, by far better than any existing alternative in many cases of interest. On real-life texts, our algorithm is especially interesting for computational biology applications. In particular, we show that our algorithm can be successfully used to search for one pattern, where many more competing algorithms exist. Our algorithm is also average-optimal in this case, being the second after that of Chang and Marr. However, our algorithm permits higher difference ratios than Chang and Marr, and this is our third contribution. In practice, our algorithm is competitive in this scenario too, being the fastest for low difference ratios and moderate alphabet sizes. This is our fourth contribution, which also answers affirmatively the question of whether a practical average-optimal approximate string-matching algorithm existed.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; approximate string matching; biological sequences; multiple string matching; optimality", } @Article{Sinha:2004:CCS, author = "Ranjan Sinha and Justin Zobel", title = "Cache-conscious sorting of large sets of strings with dynamic tries", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "9", pages = "1.5:1--1.5:??", month = "????", year = "2004", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1005813.1041517", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:22 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Ongoing changes in computer architecture are affecting the efficiency of string-sorting algorithms. The size of main memory in typical computers continues to grow but memory accesses require increasing numbers of instruction cycles, which is a problem for the most efficient of the existing string-sorting algorithms as they do not utilize cache well for large data sets. We propose a new sorting algorithm for strings, burstsort, based on dynamic construction of a compact trie in which strings are kept in buckets. It is simple, fast, and efficient. We experimentally explore key implementation options and compare burstsort to existing string-sorting algorithms on large and small sets of strings with a range of characteristics. These experiments show that, for large sets of strings, burstsort is almost twice as fast as any previous algorithm, primarily due to a lower rate of cache miss.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Goh:2004:TAP, author = "Rick Siow Mong Goh and Ian Li-Jin Thng", title = "Twol-amalgamated priority queues", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "9", pages = "1.6:1--1.6:??", month = "????", year = "2004", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1005813.1057625", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:22 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Priority queues are essential function blocks in numerous applications such as discrete event simulations. This paper describes and exemplifies the ease of obtaining high performance priority queues using a two-tier list-based structure. This new implementation, called the {\em Twol\/} structure, is amalgamated with three priority queues, namely, the Henriksen's queue, splay tree and skew heap, to enhance the efficiency of these {\em basal\/} priority queue structures. Using a model that combines traditional average case and amortized complexity analysis, Twol-amalgamated priority queues that maintain $N$ active events are theoretically proven to offer $O(1)$ {\em expected amortized complexity\/} under reasonable assumptions. They are also demonstrated empirically to offer stable near $O(1)$ performance for widely varying priority increment distributions and for queue sizes ranging from 10 to 10 million. Extensive empirical results show that the Twol-amalgamated priority queues consistently outperform those basal structures (i.e., without the Twol structure) with an average speedup of about three to five times on widely different hardware architectures. These results provide testimony that the Twol-amalgamated priority queues are suitable for implementation in sizable application scenarios such as, but not limited to, large-scale discrete event simulation.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithm analysis; calendar queue; discrete event simulation; future event list; Henriksen's; pending event set; priority queue; simulator; skew heap; splay tree; tree", } @Article{Ioannidis:2005:ADS, author = "Ioannis Ioannidis and Ananth Grama and Mikhail Atallah", title = "Adaptive data structures for {IP} lookups", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "1.1:1--1.1:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1064548", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The problem of efficient data structures for IP lookups has been well studied in the literature. Techniques such as LC tries and extensible hashing are commonly used. In this paper, we address the problem of generalizing LC tries, based on traces of past lookups, to provide performance guarantees for memory suboptimal structures. As a specific example, if a memory-optimal (LC) trie takes 6 MB and the total memory at the router is 8 MB, how should the trie be modified to make best use of the 2 MB of excess memory? We present a greedy algorithm for this problem and prove that, if for the optimal data structure there are $b$ fewer memory accesses on average for each lookup compared with the original trie, the solution produced by the greedy algorithm will have at least $9 \times b /11$ fewer memory accesses on average (compared to the original trie). An efficient implementation of this algorithm presents significant additional challenges. We describe an implementation with a time complexity of $O(\xi(d) n \log n)$ and a space complexity of $O(n)$, where $n$ is the number of nodes of the trie and $d$ its depth. The depth of a trie is fixed for a given version of the Internet protocol and is typically $O(\log n)$. In this case, $\xi(d) = O(\log^2 n)$. We also demonstrate experimentally the performance and scalability of the algorithm on actual routing data.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "IP lookups; level compression", } @Article{Lesh:2005:NHI, author = "N. Lesh and J. Marks and A. McMahon and M. Mitzenmacher", title = "New heuristic and interactive approaches to {$2$D} rectangular strip packing", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "1.2:1--1.2:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1083322", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper, we consider the two-dimensional rectangular strip packing problem. A standard simple heuristic, Bottom-Left-Decreasing (BLD), has been shown to perform quite well in practice. We introduce and demonstrate the effectiveness of BLD*, a stochastic search variation of BLD. While BLD places the rectangles in decreasing order of height, width, area, and perimeter, BLD* successively tries random orderings, chosen from a distribution determined by their Kendall-tau distance from one of these fixed orderings. Our experiments on benchmark problems show that BLD* produces significantly better packings than BLD after only 1 min of computation. Furthermore, we also show that BLD* outperforms recently reported metaheuristics. Furthermore, we observe that people seem able to reason about packing problems extremely well. We incorporate our new algorithms in an interactive system that combines the advantages of computer speed and human reasoning. Using the interactive system, we are able to quickly produce significantly better solutions than BLD* by itself.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "2D rectangular strip packing; cutting stock/trim; interactive methods", } @Article{Wagner:2005:GCE, author = "Dorothea Wagner and Thomas Willhalm and Christos Zaroliagis", title = "Geometric containers for efficient shortest-path computation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "1.3:1--1.3:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1103378", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A fundamental approach in finding efficiently best routes or optimal itineraries in traffic information systems is to reduce the search space (part of graph visited) of the most commonly used shortest path routine (Dijkstra's algorithm) on a suitably defined graph. We investigate reduction of the search space while simultaneously retaining data structures, created during a preprocessing phase, of size linear (i.e., optimal) to the size of the graph. We show that the search space of Dijkstra's algorithm can be significantly reduced by extracting geometric information from a given layout of the graph and by encapsulating precomputed shortest-path information in resulted geometric objects (containers). We present an extensive experimental study comparing the impact of different types of geometric containers using test data from real-world traffic networks. We also present new algorithms as well as an empirical study for the dynamic case of this problem, where edge weights are subject to change and the geometric containers have to be updated and show that our new methods are two to three times faster than recomputing everything from scratch. Finally, in an appendix, we discuss the software framework that we developed to realize the implementations of all of our variants of Dijkstra's algorithm. Such a framework is not trivial to achieve as our goal was to maintain a common code base that is, at the same time, small, efficient, and flexible, as we wanted to enhance and combine several variants in any possible way.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "data structures and algorithms; Dijkstra's algorithm; geometric container; graph algorithms; shortest path; traffic network", } @Article{Lopez-Ortiz:2005:FSS, author = "Alejandro L{\'o}pez-Ortiz and Mehdi Mirzazadeh and Mohammad Ali Safari and Hossein Sheikhattar", title = "Fast string sorting using order-preserving compression", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "1.4:1--1.4:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180611", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We give experimental evidence for the benefits of order-preserving compression in sorting algorithms. While, in general, any algorithm might benefit from compressed data because of reduced paging requirements, we identified two natural candidates that would further benefit from order-preserving compression, namely string-oriented sorting algorithms and word-RAM algorithms for keys of bounded length. The word-RAM model has some of the fastest known sorting algorithms in practice. These algorithms are designed for keys of bounded length, usually 32 or 64 bits, which limits their direct applicability for strings. One possibility is to use an order-preserving compression scheme, so that a bounded-key-length algorithm can be applied. For the case of standard algorithms, we took what is considered to be the among the fastest nonword RAM string sorting algorithms, Fast MKQSort, and measured its performance on compressed data. The Fast MKQSort algorithm of Bentley and Sedgewick is optimized to handle text strings. Our experiments show that order-compression techniques results in savings of approximately 15\% over the same algorithm on noncompressed data. For the word-RAM, we modified Andersson's sorting algorithm to handle variable-length keys. The resulting algorithm is faster than the standard Unix sort by a factor of 1.5 $X$. Last, we used an order-preserving scheme that is within a constant additive term of the optimal Hu--Tucker, but requires linear time rather than $O(m \log m)$, where $m = |\Sigma|$ is the size of the alphabet.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "order-preserving compression; sorting; unit-cost RAM; word-RAM", } @Article{Ribeiro:2005:P, author = "Celso C. Ribeiro and Simone L. Martins", title = "Preface", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.1:1--2.1:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180620", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Sinha:2005:URS, author = "Ranjan Sinha and Justin Zobel", title = "Using random sampling to build approximate tries for efficient string sorting", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.10:1--2.10:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180622", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Algorithms for sorting large datasets can be made more efficient with careful use of memory hierarchies and reduction in the number of costly memory accesses. In earlier work, we introduced burstsort, a new string-sorting algorithm that on large sets of strings is almost twice as fast as previous algorithms, primarily because it is more cache efficient. Burstsort dynamically builds a small trie that is used to rapidly allocate each string to a bucket. In this paper, we introduce new variants of our algorithm: SR-burstsort, DR-burstsort, and DRL-burstsort. These algorithms use a random sample of the strings to construct an approximation to the trie prior to sorting. Our experimental results with sets of over 30 million strings show that the new variants reduce, by up to 37\%, cache misses further than did the original burstsort, while simultaneously reducing instruction counts by up to 24\%. In pathological cases, even further savings can be obtained.", acknowledgement = ack-nhfb, articleno = "2.10", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "cache-aware; cache-conscious; data structure; in-memory; sorting; string", } @Article{Bracht:2005:GAA, author = "Evandro C. Bracht and Luis and A. A. Meira and F. K. Miyazawa", title = "A greedy approximation algorithm for the uniform metric labeling problem analyzed by a primal-dual technique", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.11:1--2.11:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180623", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the uniform metric labeling problem. This NP-hard problem considers how to assign objects to labels respecting assignment and separation costs. The known approximation algorithms are based on solutions of large linear programs and are impractical for moderate- and large-size instances. We present an 8log $n$-approximation algorithm that can be applied to large-size instances. The algorithm is greedy and is analyzed by a primal-dual technique. We implemented the presented algorithm and two known approximation algorithms and compared them at randomized instances. The gain of time was considerable with small error ratios. We also show that the analysis is tight, up to a constant factor.", acknowledgement = ack-nhfb, articleno = "2.11", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "approximation algorithms; graph labeling", } @Article{deSouza:2005:DMP, author = "Cid C. de Souza and Andre M. Lima and Guido Araujo and Nahri B. Moreano", title = "The datapath merging problem in reconfigurable systems: {Complexity}, dual bounds and heuristic evaluation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.2:1--2.2:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180613", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper, we investigate the data path merging problem (DPM) in reconfigurable systems. DPM is modeled as a graph optimization problem and is shown to be {\em NP\/}-hard. An Integer Programming (IP) formulation of the problem is presented and some valid inequalities for the convex hull of integer solutions are introduced. These inequalities form the basis of a branch-and-cut algorithm that we implemented. This algorithm was used to compute lower bounds for a set of DPM instances, allowing us to assess the performance of two heuristics proposed earlier in the literature for the problem. Moreover, the branch-and-cut algorithm also was proved to be a valuable tool to solve small-sized DPM instances to optimality.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "data path merging; heuristics; lower bounds; reconfigurable systems", } @Article{Du:2005:IAA, author = "Jingde Du and Stavros G. Kolliopoulos", title = "Implementing approximation algorithms for the single-source unsplittable flow problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.3:1--2.3:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180614", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In the {\em single-source unsplittable flow\/} problem, commodities must be routed simultaneously from a common source vertex to certain sinks in a given graph with edge capacities. The demand of each commodity must be routed along a single path so that the total flow through any edge is at most, its capacity. This problem was introduced by Kleinberg [1996a] and generalizes several NP-complete problems. A cost value per unit of flow may also be defined for every edge. In this paper, we implement the 2-approximation algorithm of Dinitz et al. [1999] for congestion, which is the best known, and the (3, 1)-approximation algorithm of Skutella [2002] for congestion and cost, which is the best known bicriteria approximation. We experimentally study the quality of approximation achieved by the algorithms and the effect of heuristics on their performance. We also compare these algorithms against the previous best ones by Kolliopoulos and Stein [1999].", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "approximation algorithms; network flow; unsplittable flow", } @Article{Duch:2005:IPM, author = "Amalia Duch and Conrado Mart{\'\i}nez", title = "Improving the performance of multidimensional search using fingers", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.4:1--2.4:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180615", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We propose two variants of $K$-d trees where {\em fingers\/} are used to improve the performance of orthogonal range search and nearest neighbor queries when they exhibit locality of reference. The experiments show that the second alternative yields significant savings. Although it yields more modest improvements, the first variant does it with much less memory requirements and great simplicity, which makes it more attractive on practical grounds.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "experimental algorithmics; Finger search; K-d trees; locality; multidimensional data structures; nearest-neighbors searching; orthogonal range searching", } @Article{Holzer:2005:CST, author = "Martin Holzer and Frank Schulz and Dorothea Wagner and Thomas Willhalm", title = "Combining speed-up techniques for shortest-path computations", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.5:1--2.5:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180616", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In practice, computing a shortest path from one node to another in a directed graph is a very common task. This problem is classically solved by Dijkstra's algorithm. Many techniques are known to speed up this algorithm heuristically, while optimality of the solution can still be guaranteed. In most studies, such techniques are considered individually. The focus of our work is {\em combination\/} of speed-up techniques for Dijkstra's algorithm. We consider all possible combinations of four known techniques, namely, {\em goal-directed search}, {\em bidirectional search}, {\em multilevel approach}, and {\em shortest-path containers}, and show how these can be implemented. In an extensive experimental study, we compare the performance of the various combinations and analyze how the techniques harmonize when jointly applied. Several real-world graphs from road maps and public transport and three types of generated random graphs are taken into account.", acknowledgement = ack-nhfb, articleno = "2.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "combination; Dijkstra's algorithm; shortest path; speed-up", } @Article{Hyyro:2005:IBP, author = "Heikki Hyyr{\"o} and Kimmo Fredriksson and Gonzalo Navarro", title = "Increased bit-parallelism for approximate and multiple string matching", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.6:1--2.6:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180617", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Bit-parallelism permits executing several operations simultaneously over a set of bits or numbers stored in a single computer word. This technique permits searching for the approximate occurrences of a pattern of length $m$ in a text of length $n$ in time $O(\lceil m / w \rceil n)$, where $w$ is the number of bits in the computer word. Although this is asymptotically the optimal bit-parallel speedup over the basic $O(mn)$ time algorithm, it wastes bit-parallelism's power in the common case where $m$ is much smaller than $w$, since $w - m$ bits in the computer words are unused. In this paper, we explore different ways to increase the bit-parallelism when the search pattern is short. First, we show how multiple patterns can be packed into a single computer word so as to search for all them simultaneously. Instead of spending $O(rn)$ time to search for $r$ patterns of length $m \leq w / 2$, we need $O(\lceil rm / w \rceil n)$ time. Second, we show how the mechanism permits boosting the search for a single pattern of length $m \leq w / 2$, which can be searched for in $O(\lceil n / \lfloor w / m \rfloor \rceil)$ bit-parallel steps instead of $O(n)$. Third, we show how to extend these algorithms so that the time bounds essentially depend on $k$ instead of $m$, where $k$ is the maximum number of differences permitted. Finally, we show how the ideas can be applied to other problems such as multiple exact string matching and one-against-all computation of edit distance and longest common subsequences. Our experimental results show that the new algorithms work well in practice, obtaining significant speedups over the best existing alternatives, especially on short patterns and moderate number of differences allowed. This work fills an important gap in the field, where little work has focused on very short patterns.", acknowledgement = ack-nhfb, articleno = "2.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "approximate string matching; bit-parallelism; multiple string matching", } @Article{Nikolov:2005:SEH, author = "Nikola S. Nikolov and Alexandre Tarassov and J{\"u}rgen Branke", title = "In search for efficient heuristics for minimum-width graph layering with consideration of dummy nodes", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.7:1--2.7:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180618", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We propose two fast heuristics for solving the NP-hard problem of graph layering with the minimum width and consideration of dummy nodes. Our heuristics can be used at the layer-assignment phase of the Sugiyama method for drawing of directed graphs. We evaluate our heuristics by comparing them to the widely used fast-layering algorithms in an extensive computational study with nearly 6000 input graphs. We also demonstrate how the well-known longest-path and Coffman--Graham algorithms can be used for finding narrow layerings with acceptable aesthetic properties.", acknowledgement = ack-nhfb, articleno = "2.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dummy vertices; hierarchical graph drawing; layer assignment; layered graphs; layering", } @Article{Pemmaraju:2005:AIC, author = "Sriram V. Pemmaraju and Sriram Penumatcha and Rajiv Raman", title = "Approximating interval coloring and max-coloring in chordal graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.8:1--2.8:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180619", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider two coloring problems: interval coloring and max-coloring for chordal graphs. Given a graph $G = (V, E)$ and positive-integral vertex weights $w: V \rightarrow N$, the {\em interval-coloring\/} problem seeks to find an assignment of a real interval $I(u)$ to each vertex $u \in V$, such that two constraints are satisfied: (i) for every vertex $u \in V$, $|I(u)| = w(u)$ and (ii) for every pair of adjacent vertices $u$ and $v$, $I(u) \cap I(v) = \emptyset$. The goal is to minimize the {\em span\/} $|\cup_{v \in V} I(v)|$. The {\em max-coloring problem\/} seeks to find a proper vertex coloring of $G$ whose color classes $C_1$, $C_2$, \ldots{}, $C_k$, minimize the sum of the weights of the heaviest vertices in the color classes, that is, $\sum^k_i = 1 \hbox{max}_{v \epsilon C i w (v)}$. Both problems arise in efficient memory allocation for programs. The interval-coloring problem models the compile-time memory allocation problem and has a rich history dating back at least to the 1970s. The max-coloring problem arises in minimizing the total buffer size needed by a dedicated memory manager for programs. In another application, this problem models scheduling of conflicting jobs in batches to minimize the {\em makespan}. Both problems are NP-complete even for interval graphs, although there are constant-factor approximation algorithms for both problems on interval graphs. In this paper, we consider these problems for {\em chordal graphs}, a subclass of perfect graphs. These graphs naturally generalize interval graphs and can be defined as the class of graphs that have no induced cycle of length $> 3$. Recently, a 4-approximation algorithm (which we call GeomFit) has been presented for the max-coloring problem on perfect graphs (Pemmaraju and Raman 2005). This algorithm can be used to obtain an interval coloring as well, but without the constant-factor approximation guarantee. In fact, there is no known constant-factor approximation algorithm for the interval-coloring problem on perfect graphs. We study the performance of GeomFit and several simple $O(\log(n))$-factor approximation algorithms for both problems. We experimentally evaluate and compare four simple heuristics: first-fit, best-fit, GeomFit, and a heuristic based on partitioning the graph into vertex sets of similar weight. Both for max-coloring and for interval coloring, GeomFit deviates from OPT by about 1.5\%, on average. The performance of first-fit comes close second, deviating from OPT by less than 6\%, on average, for both problems. Best-fit comes third and graph-partitioning heuristic comes a distant last. Our basic data comes from about 10,000 runs of each of the heuristics for each of the two problems on randomly generated chordal graphs of various sizes, sparsity, and structure.", acknowledgement = ack-nhfb, articleno = "2.8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "chordal graphs; dynamic storage allocation; graph coloring; perfect graphs", } @Article{Santos:2005:TSH, author = "Haroldo G. Santos and Luiz S. Ochi and Marcone J. F. Souza", title = "A {Tabu} search heuristic with efficient diversification strategies for the class\slash teacher timetabling problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "10", pages = "2.9:1--2.9:??", month = "????", year = "2005", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1064546.1180621", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:05:40 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The Class/Teacher Timetabling Problem (CTTP) deals with the weekly scheduling of encounters between teachers and classes of an educational institution. Since CTTP is a NP-hard problem for nearly all of its variants, the use of heuristic methods for its resolution is justified. This paper presents an efficient Tabu Search (TS) heuristic with two different memory based diversification strategies for CTTP. Results obtained through an application of the method to a set of real world problems show that it produces better solutions than a previously proposed TS found in the literature and faster times are observed in the production of good quality solutions.", acknowledgement = ack-nhfb, articleno = "2.9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "metaheuristics; tabu search; timetabling", } @Article{Salmela:2006:MSM, author = "Leena Salmela and Jorma Tarhio and Jari Kyt{\"o}joki", title = "Multipattern string matching with $q$-grams", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "1.1:1--1.1:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1187438", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present three algorithms for exact string matching of multiple patterns. Our algorithms are filtering methods, which apply $q$-grams and bit parallelism. We ran extensive experiments with them and compared them with various versions of earlier algorithms, e.g., different trie implementations of the Aho--Corasick algorithm. All of our algorithms appeared to be substantially faster than earlier solutions for sets of 1,000--10,000 patterns and the good performance of two of them continues to 100,000 patterns. The gain is because of the improved filtering efficiency caused by $q$-grams.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "content scanning; intrusion detection; multiple string matching", } @Article{Sinha:2006:CES, author = "Ranjan Sinha and Justin Zobel and David Ring", title = "Cache-efficient string sorting using copying", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "1.2:1--1.2:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1187439", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Burstsort is a cache-oriented sorting technique that uses a dynamic trie to efficiently divide large sets of string keys into related subsets small enough to sort in cache. In our original burstsort, string keys sharing a common prefix were managed via a bucket of pointers represented as a list or array; this approach was found to be up to twice as fast as the previous best string sorts, mostly because of a sharp reduction in out-of-cache references. In this paper, we introduce C-burstsort, which copies the unexamined tail of each key to the bucket and discards the original key to improve data locality. On both Intel and PowerPC architectures, and on a wide range of string types, we show that sorting is typically twice as fast as our original burstsort and four to five times faster than multikey quicksort and previous radixsorts. A variant that copies both suffixes and record pointers to buckets, CP-burstsort, uses more memory, but provides stable sorting. In current computers, where performance is limited by memory access latencies, these new algorithms can dramatically reduce the time needed for internal sorting of large numbers of strings.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; cache; experimental algorithms; sorting; string management; tries", } @Article{Penner:2006:CFI, author = "Michael Penner and Viktor K. Prasanna", title = "Cache-Friendly implementations of transitive closure", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "1.3:1--1.3:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1210586", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The topic of cache performance has been well studied in recent years. Compiler optimizations exist and optimizations have been done for many problems. Much of this work has focused on dense linear algebra problems. At first glance, the Floyd--Warshall algorithm appears to fall into this category. In this paper, we begin by applying two standard cache-friendly optimizations to the Floyd--Warshall algorithm and show limited performance improvements. We then discuss the unidirectional space time representation (USTR). We show analytically that the USTR can be used to reduce the amount of processor-memory traffic by a factor of $O(\sqrt C)$, where $C$ is the cache size, for a large class of algorithms. Since the USTR leads to a tiled implementation, we develop a tile size selection heuristic to intelligently narrow the search space for the tile size that minimizes total execution time. Using the USTR, we develop a cache-friendly implementation of the Floyd--Warshall algorithm. We show experimentally that this implementation minimizes the level-1 and level-2 cache misses and TLB misses and, therefore, exhibits the best overall performance. Using this implementation, we show a $2 \times$ improvement in performance over the best compiler optimized implementation on three different architectures. Finally, we show analytically that our implementation of the Floyd--Warshall algorithm is asymptotically optimal with respect to processor-memory traffic. We show experimental results for the Pentium III, Alpha, and MIPS R12000 machines using problem sizes between 1024 and 2048 vertices. We demonstrate improved cache performance using the Simplescalar simulator.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "data structures; Floyd--Warshall algorithm; systolic array algorithms", } @Article{Goshi:2006:ADM, author = "Justin Goshi and Richard E. Ladner", title = "Algorithms for dynamic multicast key distribution", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "1.4:1--1.4:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1210587", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study the problem of multicast key distribution for group security. Secure group communication systems typically rely on a group key, which is a secret shared among the members of the group. This key is used to provide privacy by encrypting all group communications. Because groups can be large and highly dynamic, it becomes necessary to change the group key in a scalable and secure fashion when members join and leave the group. We present a series of algorithms for solving this problem based on key trees. The algorithms attempt to minimize the worst-case communication cost of updates by maintaining balanced key tree structures. We focus on the trade-off between the communication cost because of the structure of the tree and that due to the overhead of restructuring the tree to maintain its balanced structure. The algorithms are analyzed for worst-case tree structure bounds and evaluated empirically via simulations.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dynamic key distribution; experimental algorithms; multicast", } @Article{Aleksandrov:2006:PPG, author = "Lyudmil Aleksandrov and Hristo Djidjev and Hua Guo and Anil Maheshwari", title = "Partitioning planar graphs with costs and weights", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "1.5:1--1.5:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1210588", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A graph separator is a set of vertices or edges whose removal divides an input graph into components of bounded size. This paper describes new algorithms for computing separators in planar graphs as well as techniques that can be used to speed up the implementation of graph partitioning algorithms and improve the partition quality. In particular, we consider planar graphs with costs and weights on the vertices, where weights are used to estimate the sizes of the partitions and costs are used to estimate the size of the separator. We show that in these graphs one can always find a small cost separator (consisting of vertices or edges) that partitions the graph into components of bounded weight. We describe implementations of the partitioning algorithms and discuss results of our experiments.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "graph algorithms; graph partitioning algorithms; graph separators; implementation", } @Article{Ilinkin:2006:HEC, author = "Ivayio Ilinkin and Ravi Janardan and Michiel Smid and Eric Johnson and Paul Castillo and J{\"o}rg Schwerdt", title = "Heuristics for estimating contact area of supports in layered manufacturing", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "1.6:1--1.6:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1210589", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Layered manufacturing is a technology that allows physical prototypes of three-dimensional(3D) models to be built directly from their digital representation, as a stack of two-dimensional(2D) layers. A key design problem here is the choice of a suitable direction in which the digital model should be oriented and built so as to minimize the area of contact between the prototype and temporary support structures that are generated during the build. Devising an efficient algorithm for computing such a direction has remained a difficult problem for quite some time. In this paper, a suite of efficient and practical heuristics is presented for estimating the minimum contact area. Also given is a technique for evaluating the quality of the estimate provided by any heuristic, which does not require knowledge of the (unknown and hard-to-compute) optimal solution; instead, it provides an indirect upper bound on the quality of the estimate via two relatively easy-to-compute quantities. The algorithms are based on various techniques from computational geometry, such as ray-shooting, convex hulls, boolean operations on polygons, and spherical arrangements, and have been implemented and tested. Experimental results on a wide range of real-world models show that the heuristics perform quite well in practice.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithm implementation and testing; computational geometry", } @Article{Pearce:2006:DTS, author = "David J. Pearce and Paul H. J. Kelly", title = "A dynamic topological sort algorithm for directed acyclic graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "1.7:1--1.7:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1210590", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the problem of maintaining the topological order of a directed acyclic graph (DAG) in the presence of edge insertions and deletions. We present a new algorithm and, although this has inferior time complexity compared with the best previously known result, we find that its simplicity leads to better performance in practice. In addition, we provide an empirical comparison against the three main alternatives over a large number of random DAGs. The results show our algorithm is the best for sparse digraphs and only a constant factor slower than the best on dense digraphs.", acknowledgement = ack-nhfb, articleno = "1.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dynamic graph algorithms; topological sort", } @Article{Flammini:2006:RAF, author = "Michele Flammini and Alfredo Navarra and Stephane Perennes", title = "The ``real'' approximation factor of the {MST} heuristic for the minimum energy broadcasting", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.10:1--2.10:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216587", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "This paper deals with one of the most studied problems in the last few years in the field of wireless communication in ad-hoc networks. The problem consists of reducing the total energy consumption of wireless radio stations distributed over a given area of interest in order to perform the basic pattern of communication by a broadcast. Recently, a tight 6-approximation of the minimum spanning tree heuristic has been proven. While such a bound is theoretically optimal if compared to the known lower bound of 6, there is an obvious gap with practical experimental results. By extensive experiments, proposing a new technique to generate input instances and supported by theoretical results, we show how the approximation ratio can be actually considered close to 4 for a ``real-world'' set of instances. We consider, in fact, instances more representative of common practices. Those are usually composed by considerable number of nodes uniformly and randomly distributed inside the area of interest.", acknowledgement = ack-nhfb, articleno = "2.10", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "ad-hoc networks; broadcast; energy saving; spanning tree", } @Article{Nikoletseas:2006:JSS, author = "Sotiris Nikoletseas", title = "{JEA Special Section}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.1:1--2.1:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216578", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Fahle:2006:FBB, author = "Torsten Fahle and Karsten Tiemann", title = "A faster branch-and-bound algorithm for the test-cover problem based on set-covering techniques", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.2:1--2.2:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216579", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The test-cover problem asks for the minimal number of tests needed to uniquely identify a disease, infection, etc. A collection of branch-and-bound algorithms was proposed by De Bontridder et al. [2002]. Based on their work, we introduce several improvements that are compatible with all techniques described in De Bontridder et al. [2002] and the more general setting of {\em weighted\/} test-cover problems. We present a faster data structure, cost-based variable fixing, and adapt well-known set-covering techniques, including Lagrangian relaxation and upper-bound heuristics. The resulting algorithm solves benchmark instances up to 10 times faster than the former approach and up to 100 times faster than a general MIP solver.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "branch-and-bound; Lagrangian relaxation; set-cover problem; test-cover problem; variable fixing", } @Article{Leone:2006:FPN, author = "Pierre Leone and Jose Rolim and Paul Albuquerque and Christian Mazza", title = "A framework for probabilistic numerical evaluation of sensor networks: a case study of a localization protocol", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.3:1--2.3:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216580", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper we show how to use stochastic estimation methods to investigate topological properties of sensor networks as well as the behavior of dynamical processes on these networks. The framework is particularly important to study problems for which no theoretical results are known, or cannot be directly applied in practice, for instance, when only asymptotic results are available. We also interpret Russo's formula in the context of sensor networks and thus obtain practical information on their reliability. As a case study, we analyze a localization protocol for wireless sensor networks and validate our approach by numerical experiments. Finally, we mention three applications of our approach: estimating the number of pivotal sensors in a real network, minimizing the number of such sensors for robustness purposes during the network design and estimating the distance between successive localized positions for mobile sensor networks.", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "localization process and reliability; sensor networks; stochastic recursive estimation", } @Article{Festa:2006:GPR, author = "Paola Festa and Panos M. Pardalos and Leonidas S. Pitsoulis and Mauricio G. C. Resende", title = "{GRASP} with path relinking for the weighted {MAXSAT} problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.4:1--2.4:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216581", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A GRASP with path relinking for finding good-quality solutions of the weighted maximum satisfiability problem (MAX-SAT) is described in this paper. GRASP, or Greedy Randomized Adaptive Search Procedure, is a randomized multistart metaheuristic, where, at each iteration, locally optimal solutions are constructed, each independent of the others. Previous experimental results indicate its effectiveness for solving weighted MAX-SAT instances. Path relinking is a procedure used to intensify the search around good-quality isolated solutions that have been produced by the GRASP heuristic. Experimental comparison of the pure GRASP (without path relinking) and the GRASP with path relinking illustrates the effectiveness of path relinking in decreasing the average time needed to find a good-quality solution for the weighted maximum satisfiability problem.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; experimentation; GRASP; heuristics; path relinking; performance; time-to-target plots", } @Article{Mehlhorn:2006:IMC, author = "Kurt Mehlhorn and Dimitrios Michail", title = "Implementing minimum cycle basis algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.5:1--2.5:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216582", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this paper, we consider the problem of computing a minimum cycle basis of an undirected graph $G$ = ($V$, $E$) with $n$ vertices and $m$ edges. We describe an efficient implementation of an $O(m^3 + mn^2 \log n)$ algorithm. For sparse graphs, this is the currently best-known algorithm. This algorithm's running time can be partitioned into two parts with time $O(m^3)$ and $O(m^2 n + mn^2 \log n)$, respectively. Our experimental findings imply that for random graphs the true bottleneck of a sophisticated implementation is the $O(m^2 n + mn^2 \log n)$ part. A straightforward implementation would require $\Omega(n m)$ shortest-path computations. Thus, we develop several heuristics in order to get a practical algorithm. Our experiments show that in random graphs our techniques result in a significant speed-up. Based on our experimental observations, we combine the two fundamentally different approaches to compute a minimum cycle basis to obtain a new hybrid algorithm with running time $O(m^2 n^2)$. The hybrid algorithm is very efficient, in practice, for random dense unweighted graphs. Finally, we compare these two algorithms with a number of previous implementations for finding a minimum cycle basis of an undirected graph.", acknowledgement = ack-nhfb, articleno = "2.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "cycle basis; graph algorithms", } @Article{Heinrich-Litan:2006:RCR, author = "Laura Heinrich-Litan and Marco E. L{\"u}bbecke", title = "Rectangle covers revisited computationally", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.6:1--2.6:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216583", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the problem of covering an orthogonal polygon with a minimum number of axis-parallel rectangles from a computational point of view. We propose an integer program which is the first general approach to obtain provably optimal solutions to this well-studied NP-hard problem. It applies to common variants like covering only the corners or the boundary of the polygon and also to the weighted case. In experiments, it turns out that the linear programming relaxation is extremely tight and rounding a fractional solution is an immediate high-quality heuristic. We obtain excellent experimental results for polygons originating from VLSI design, fax data sheets, black and white images, and for random instances. Making use of the dual linear program, we propose a stronger lower bound on the optimum, namely, the cardinality of a fractional stable set. Furthermore, we outline ideas how to make use of this bound in primal--dual-based algorithms. We give partial results, which make us believe that our proposals have a strong potential to settle the main open problem in the area: To find a constant factor approximation algorithm for the rectangle cover problem.", acknowledgement = ack-nhfb, articleno = "2.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "integer programming; linear programming", } @Article{Panagopoulou:2006:APN, author = "Panagiota N. Panagopoulou and Paul G. Spirakis", title = "Algorithms for pure {Nash} equilibria in weighted congestion games", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.7:1--2.7:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216584", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In large-scale or evolving networks, such as the Internet, there is no authority possible to enforce a centralized traffic management. In such situations, game theory, and especially the concepts of Nash equilibria and congestion games [Rosenthal 1973] are a suitable framework for analyzing the equilibrium effects of selfish routes selection to network delays. We focus here on {\em single-commodity\/} networks where selfish users select paths to route their loads (represented by arbitrary integer {\em weights\/}). We assume that individual link delays are equal to the total load of the link. We then focus on the algorithm suggested in Fotakis et al. [2005], i.e., a potential-based method for finding {\em pure\/} Nash equilibria in such networks. A superficial analysis of this algorithm gives an upper bound on its time, which is polynomial in $n$ (the number of users) and the sum of their weights $W$. This bound can be exponential in $n$ when some weights are exponential. We provide strong experimental evidence that this algorithm actually converges to a pure Nash equilibrium in {\em polynomial time}. More specifically, our experimental findings suggest that the running time is a polynomial function of $n$ and $\log W$. In addition, we propose an initial allocation of users to paths that dramatically accelerates this algorithm, compared to an arbitrary initial allocation. A by-product of our research is the discovery of a weighted potential function when link delays are {\em exponential\/} to their loads. This asserts the existence of pure Nash equilibria for these delay functions and extends the result of Fotakis et al. [2005].", acknowledgement = ack-nhfb, articleno = "2.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "congestion games; game theory; pure Nash equilibria", } @Article{Mohring:2006:PGS, author = "Rolf H. M{\"o}hring and Heiko Schilling and Birk Sch{\"u}tz and Dorothea Wagner and Thomas Willhalm", title = "Partitioning graphs to speedup {Dijkstra}'s algorithm", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.8:1--2.8:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216585", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study an acceleration method for point-to-point shortest-path computations in large and sparse directed graphs with given nonnegative arc weights. The acceleration method is called the {\em arc-flag approach\/} and is based on Dijkstra's algorithm. In the arc-flag approach, we allow a preprocessing of the network data to generate additional information, which is then used to speedup shortest-path queries. In the preprocessing phase, the graph is divided into regions and information is gathered on whether an arc is on a shortest path into a given region. The arc-flag method combined with an appropriate partitioning and a bidirected search achieves an average speedup factor of more than 500 compared to the standard algorithm of Dijkstra on large networks (1 million nodes, 2.5 million arcs). This combination narrows down the search space of Dijkstra's algorithm to almost the size of the corresponding shortest path for long-distance shortest-path queries. We conduct an experimental study that evaluates which partitionings are best suited for the arc-flag method. In particular, we examine partitioning algorithms from computational geometry and a multiway arc separator partitioning. The evaluation was done on German road networks. The impact of different partitions on the speedup of the shortest path algorithm are compared. Furthermore, we present an extension of the speedup technique to multiple levels of partitions. With this multilevel variant, the same speedup factors can be achieved with smaller space requirements. It can, therefore, be seen as a compression of the precomputed data that preserves the correctness of the computed shortest paths.", acknowledgement = ack-nhfb, articleno = "2.8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "acceleration method; Dijkstra's algorithm; road network; shortest path", } @Article{Boukerche:2006:ICC, author = "Azzedine Boukerche and Alba Cristina Magalhaes {Alves De Melo}", title = "Integrating coordinated checkpointing and recovery mechanisms into {DSM} synchronization barriers", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "11", pages = "2.9:1--2.9:??", month = "????", year = "2006", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1187436.1216586", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:06:20 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Distributed shared memory (DSM) creates an abstraction of a physical shared memory that parallel programmers can access. Most recent software DSM systems provide relaxed-memory models that guarantee consistency only at synchronization operations, such as locks and barriers. As the main goal of DSM systems is to provide support for long-term computation-intensive applications, checkpointing and recovery mechanisms are highly desirable. This article presents and evaluates the integration of a coordinated checkpointing mechanism to the barrier primitive that is usually provided with many DSM systems. Our results on some popular benchmarks and a real parallel application show that the overhead introduced during the failure-free execution is often small.", acknowledgement = ack-nhfb, articleno = "2.9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "barrier synchronization; distributed shared memory", } @Article{Anonymous:2008:EGC, author = "Anonymous", title = "Engineering graph clustering: {Models} and experimental evaluation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.1:1--1.1:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1227162", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A promising approach to graph clustering is based on the intuitive notion of intracluster density versus intercluster sparsity. As for the weighted case, clusters should accumulate lots of weight, in contrast to their connection to the remaining graph, which should be light. While both formalizations and algorithms focusing on particular aspects of this rather vague concept have been proposed, no conclusive argument on their appropriateness has been given. In order to deepen the understanding of particular concepts, including both quality assessment as well as designing new algorithms, we conducted an experimental evaluation of graph-clustering approaches. By combining proved techniques from graph partitioning and geometric clustering, we also introduce a new approach that compares favorably.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "clustering algorithms; experimental evaluation; graph clustering; quality measures", } @Article{Barsky:2008:GAT, author = "Marina Barsky and Ulrike Stege and Alex Thomo and Chris Upton", title = "A graph approach to the threshold all-against-all substring matching problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.10:1--1.10:??", month = jun, year = "2008", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1370596.1370601", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present a novel graph model and an efficient algorithm for solving the ``threshold all against all'' problem, which involves searching two strings (with length $M$ and $N$, respectively) for all maximal approximate substring matches of length at least $S$, with up to $K$ differences. Our algorithm solves the problem in time $O(MNK_3)$, which is a considerable improvement over the previous known bound for this problem. We also provide experimental evidence that, in practice, our algorithm exhibits a better performance than its worst-case running time.", acknowledgement = ack-nhfb, articleno = "1.10", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "bioinformatics; complexity; string matching", } @Article{Dietzfelbinger:2008:DIB, author = "Martin Dietzfelbinger and Martin H{\"u}hne and Christoph Weidling", title = "A dictionary implementation based on dynamic perfect hashing", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.11:1--1.11:??", month = jun, year = "2008", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1370596.1370602", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We describe experimental results on an implementation of a dynamic dictionary. The basis of our implementation is ``dynamic perfect hashing'' as described by Dietzfelbinger et al. ({\em SIAM J. Computing 23}, 1994, pp. 738--761), an extension of the storage scheme proposed by Fredman et al. ({\em J. ACM\/} 31, 1984, pp. 538--544). At the top level, a hash function is used to partition the keys to be stored into several sets. On the second level, there is a perfect hash function for each of these sets. This technique guarantees $O(1)$ worst-case time for lookup and expected $O(1)$ amortized time for insertion and deletion, while only linear space is required. We study the practical performance of dynamic perfect hashing and describe improvements of the basic scheme. The focus is on the choice of the hash function (both for integer and string keys), on the efficiency of rehashing, on the handling of small buckets, and on the space requirements of the implementation.", acknowledgement = ack-nhfb, articleno = "1.11", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "data structures; dictionaries; dynamic hashing; hash functions; implementation", } @Article{Maniscalco:2008:EVA, author = "Michael A. Maniscalco and Simon J. Puglisi", title = "An efficient, versatile approach to suffix sorting", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.2:1--1.2:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1278374", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Sorting the suffixes of a string into lexicographical order is a fundamental task in a number of contexts, most notably lossless compression (Burrows--Wheeler transformation) and text indexing (suffix arrays). Most approaches to suffix sorting produce a sorted array of suffixes directly, continually moving suffixes into their final place in the array until the ordering is complete. In this article, we describe a novel and resource-efficient (time and memory) approach to suffix sorting, which works in a complementary way --- by assigning each suffix its rank in the final ordering, before converting to a sorted array, if necessary, once all suffixes are ranked. We layer several powerful extensions on this basic idea and show experimentally that our approach is superior to other leading algorithms in a variety of real-world contexts.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "Burrows--Wheeler transform; suffix array; suffix sorting; suffix tree", } @Article{Aloul:2008:SBP, author = "Fadi A. Aloul and Arathi Ramani and Igor L. Markov and Karem A. Sakallah", title = "Symmetry breaking for pseudo-{Boolean} formulas", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.3:1--1.3:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1278375", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Many important tasks in design automation and artificial intelligence can be performed in practice via reductions to Boolean satisfiability (SAT). However, such reductions often omit application-specific structure, thus handicapping tools in their competition with creative engineers. Successful attempts to represent and utilize additional structure on Boolean variables include recent work on 0-1 integer linear programming (ILP) and symmetries in SAT. Those extensions gracefully accommodate well-known advances in SAT solving, however, no previous work has attempted to combine both extensions. Our work shows (i) how one can detect and use symmetries in instances of 0-1 ILP, and (ii) what benefits this may bring.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "graph automorphism", } @Article{Pellegrini:2008:EIT, author = "Marco Pellegrini and Giordano Fusco", title = "Efficient {IP} table lookup via adaptive stratified trees with selective reconstructions", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.4:1--1.4:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1278376", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "IP address lookup is a critical operation for high-bandwidth routers in packet-switching networks, such as Internet. The lookup is a nontrivial operation, since it requires searching for the longest prefix, among those stored in a (large) given table, matching the IP address. Ever increasing routing table size, traffic volume, and links speed demand new and more efficient algorithms. Moreover, the imminent move to IPv6 128-bit addresses will soon require a rethinking of previous technical choices. This article describes a the new data structure for solving the IP table lookup problem christened the adaptive stratified tree (AST). The proposed solution is based on casting the problem in geometric terms and on repeated application of efficient local geometric optimization routines. Experiments with this approach have shown that in terms of storage, query time, and update time the AST is at a par with state of the art algorithms based on data compression or string manipulations (and often it is better on some of the measured quantities).", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "data structures; IP table lookup", } @Article{Navarro:2008:DSA, author = "Gonzalo Navarro and Nora Reyes", title = "Dynamic spatial approximation trees", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.5:1--1.5:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1322337", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Metric space searching is an emerging technique to address the problem of efficient similarity searching in many applications, including multimedia databases and other repositories handling complex objects. Although promising, the metric space approach is still immature in several aspects that are well established in traditional databases. In particular, most indexing schemes are static, that is, few of them tolerate insertion or deletion of elements at reasonable cost over an existing index. The spatial approximation tree ({\em sa--tree\/}) has been experimentally shown to provide a good tradeoff between construction cost, search cost, and space requirement. However, the {\em sa--tree\/} is static, which renders it unsuitable for many database applications. In this paper, we study different methods to handle insertions and deletions on the {\em sa--tree\/} at low cost. In many cases, the dynamic construction (by successive insertions) is even faster than the previous static construction, and both are similar elsewhere. In addition, the dynamic version significantly improves the search performance of {\em sa--trees\/} in virtually all cases. The result is a much more practical data structure that can be useful in a wide range of database applications.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "multimedia databases; similarity or proximity search; spatial and multidimensional search; spatial approximation tree", } @Article{Li:2008:EAC, author = "Keqin Li", title = "Experimental average-case performance evaluation of online algorithms for routing and wavelength assignment and throughput maximization in {WDM} optical networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.7:1--1.7:??", month = jun, year = "2008", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1370596.1370598", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We investigate the problem of online routing and wavelength assignment and the related throughput maximization problem in wavelength division multiplexing optical networks. It is pointed out that these problems are highly inapproximable, that is, the competitive ratio of any algorithm is at least a polynomial. We evaluate the average-case performance of several online algorithms, which have no knowledge of future arriving connection requests when processing the current connection request. Our experimental results on a wide range of optical networks demonstrate that the average-case performance of these algorithms are very close to optimal.", acknowledgement = ack-nhfb, articleno = "1.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "average-case performance; competitive ratio; online algorithm; optical network; routing; wavelength assignment; wavelength division multiplexing", } @Article{Biggar:2008:ESS, author = "Paul Biggar and Nicholas Nash and Kevin Williams and David Gregg", title = "An experimental study of sorting and branch prediction", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.8:1--1.8:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1370599", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Sorting is one of the most important and well-studied problems in computer science. Many good algorithms are known which offer various trade-offs in efficiency, simplicity, memory use, and other factors. However, these algorithms do not take into account features of modern computer architectures that significantly influence performance. Caches and branch predictors are two such features and, while there has been a significant amount of research into the cache performance of general purpose sorting algorithms, there has been little research on their branch prediction properties. In this paper, we empirically examine the behavior of the branches in all the most common sorting algorithms. We also consider the interaction of cache optimization on the predictability of the branches in these algorithms. We find insertion sort to have the fewest branch mispredictions of any comparison-based sorting algorithm, that bubble and shaker sort operate in a fashion that makes their branches highly unpredictable, that the unpredictability of shellsort's branches improves its caching behavior, and that several cache optimizations have little effect on mergesort's branch mispredictions. We find also that optimizations to quicksort, for example the choice of pivot, have a strong influence on the predictability of its branches. We point out a simple way of removing branch instructions from a classic heapsort implementation and also show that unrolling a loop in a cache-optimized heapsort implementation improves the predicitability of its branches. Finally, we note that when sorting random data two-level adaptive branch predictors are usually no better than simpler bimodal predictors. This is despite the fact that two-level adaptive predictors are almost always superior to bimodal predictors, in general.", acknowledgement = ack-nhfb, articleno = "1.8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "branch prediction; caching; pipeline architectures; sorting", } @Article{Hazel:2008:TCL, author = "Thomas Hazel and Laura Toma and Jan Vahrenhold and Rajiv Wickremesinghe", title = "Terracost: {Computing} least-cost-path surfaces for massive grid terrains", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "1.9:1--1.9:??", month = jun, year = "2008", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1370596.1370600", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "This paper addresses the problem of computing least-cost-path surfaces for massive grid terrains. Consider a grid terrain $T$ and let $C$ be a cost grid for $T$ such that every point in $C$ stores a value that represents the cost of traversing the corresponding point in $T$. Given $C$ and a set of sources $S \in T$, a least-cost-path grid $\Delta$ for $T$ is a grid such that every point in $\Delta$ represents the distance to the source in $S$ that can be reached with minimal cost. We present a scalable approach to computing least-cost-path grids. Our algorithm, terracost, is derived from our previous work on I/O-efficient shortest paths on grids and uses $O(\hbox{sort}(n))$ I/Os, where $\hbox{sort}(n)$ is the complexity of sorting $n$ items of data in the I/O-model of Aggarwal and Vitter. We present the design, the analysis, and an experimental study of terracost. An added benefit of the algorithm underlying terracost is that it naturally lends itself to parallelization. We have implemented terracost in a distributed environment using our cluster management tool and report on experiments that show that it obtains speedup near-linear with the size of the cluster. To the best of our knowledge, this is the first experimental evaluation of a multiple-source least-cost-path algorithm in the external memory setting.", acknowledgement = ack-nhfb, articleno = "1.9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "data structures and algorithms; Dijkstra's algorithm; I/O-efficiency; shortest paths; terrain data", } @Article{Arge:2008:P, author = "Lars Arge and Giuseppe F. Italiano", title = "Preface", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "2.1:1--2.1:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1227163", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Brodal:2008:ECO, author = "Gerth St{\o}lting Brodal and Rolf Fagerberg and Kristoffer Vinther", title = "Engineering a cache-oblivious sorting algorithm", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "2.2:1--2.2:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1227164", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "This paper is an algorithmic engineering study of cache-oblivious sorting. We investigate by empirical methods a number of implementation issues and parameter choices for the cache-oblivious sorting algorithm Lazy Funnelsort and compare the final algorithm with Quicksort, the established standard for comparison-based sorting, as well as with recent cache-aware proposals. The main result is a carefully implemented cache-oblivious sorting algorithm, which, our experiments show, can be faster than the best Quicksort implementation we are able to find for input sizes well within the limits of RAM. It is also at least as fast as the recent cache-aware implementations included in the test. On disk, the difference is even more pronounced regarding Quicksort and the cache-aware algorithms, whereas the algorithm is slower than a careful implementation of multiway Mergesort, such as TPIE.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "cache-oblivious algorithms; funnelsort; quicksort", } @Article{Bender:2008:SSH, author = "Michael A. Bender and Bryan Bradley and Geetha Jagannathan and Krishnan Pillaipakkamnatt", title = "Sum-of-squares heuristics for bin packing and memory allocation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "2.3:1--2.3:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1227165", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The sum-of-squares algorithm (SS) was introduced by Csirik, Johnson, Kenyon, Shor, and Weber for online bin packing of integral-sized items into integral-sized bins. First, we show the results of experiments from two new variants of the SS algorithm. The first variant, which runs in time $O(n \sqrt{B \log B})$, appears to have almost identical expected waste as the sum-of-squares algorithm on all the distributions mentioned in the original papers on this topic. The other variant, which runs in $O(n \log B)$ time, performs well on most, but not on all of those distributions. We also apply SS to the online memory-allocation problem. Our experimental comparisons between SS and Best Fit indicate that neither algorithm is consistently better than the other. If the amount of randomness in item sizes is low, SS appears to have lower waste than Best Fit, whereas, if the amount of randomness is high Best Fit appears to have lower waste than SS. Our experiments suggest that in both real and synthetic traces, SS does not seem to have an asymptotic advantage over Best Fit, in contrast with the bin-packing problem.", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "bin packing; memory allocation; sum of squares", } @Article{Pyrga:2008:EMT, author = "Evangelia Pyrga and Frank Schulz and Dorothea Wagner and Christos Zaroliagis", title = "Efficient models for timetable information in public transportation systems", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "2.4:1--2.4:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1227166", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider two approaches that model timetable information in public transportation systems as shortest-path problems in weighted graphs. In the {\em time-expanded\/} approach, every event at a station, e.g., the departure of a train, is modeled as a node in the graph, while in the {\em time-dependent\/} approach the graph contains only one node per station. Both approaches have been recently considered for (a simplified version of) the earliest arrival problem, but little is known about their relative performance. Thus far, there are only theoretical arguments in favor of the time-dependent approach. In this paper, we provide the first extensive experimental comparison of the two approaches. Using several real-world data sets, we evaluate the performance of the basic models and of several new extensions towards realistic modeling. Furthermore, new insights on solving bicriteria optimization problems in both models are presented. The time-expanded approach turns out to be more robust for modeling more complex scenarios, whereas the time-dependent approach shows a clearly better performance.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "itinerary query; public transportation system; shortest path; timetable information", } @Article{Leaver-Fay:2008:FPH, author = "Andrew Leaver-Fay and Yuanxin Liu and Jack Snoeyink and Xueyi Wang", title = "Faster placement of hydrogens in protein structures by dynamic programming", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "2.5:1--2.5:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1227167", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "M. Word and coauthors from the Richardsons' 3D Protein Structure laboratory at Duke University propose {\em dot scores\/} to measure interatomic interactions in molecular structures. Their program REDUCE uses these scores in a brute-force search to solve instances of the {\em NP\/}-hard problem of finding the optimal placement of hydrogen atoms in molecular structures determined by X-ray crystallography. We capture the central combinatorial optimization in the hydrogen placement problem with an abstraction that we call an interaction (hyper)graph. REDUCE's dot-based scoring function cannot be decomposed into the sum of pair interactions, but because the function is short ranged we are able to decompose it into the sum of single, pair, triple, and quadruple interactions that we represent by graph hyperedges. Almost every interaction graph we have observed has had a small treewidth. This fact allows us to replace the brute-force search by dynamic programming, giving speedups of nearly ten orders of magnitude. This dynamic programming has been incorporated into REDUCE and is available for download.", acknowledgement = ack-nhfb, articleno = "2.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dynamic programming; hard-sphere model; hydrogen bonds; hydrogen placement; protein structure; treewidth", } @Article{Demetrescu:2008:PA, author = "Camil Demetrescu and Roberto Tamassia", title = "Papers from {ALENEX 2005}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "3.1:1--3.1:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1402293", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "3.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Brodal:2008:AQ, author = "Gerth St{\o}lting Brodal and Rolf Fagerberg and Gabriel Moruz", title = "On the adaptiveness of {Quicksort}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "3.2:1--3.2:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1402294", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Quicksort was first introduced in 1961 by Hoare. Many variants have been developed, the best of which are among the fastest generic-sorting algorithms available, as testified by the choice of Quicksort as the default sorting algorithm in most programming libraries. Some sorting algorithms are adaptive, i.e., they have a complexity analysis that is better for inputs, which are nearly sorted, according to some specified measure of presortedness. Quicksort is not among these, as it uses $\Omega (n \log n)$ comparisons even for sorted inputs. However, in this paper, we demonstrate empirically that the actual running time of Quicksort {\em is\/} adaptive with respect to the presortedness measure Inv. Differences close to a factor of two are observed between instances with low and high Inv value. We then show that for the randomized version of Quicksort, the number of element {\em swaps\/} performed is {\em provably\/} adaptive with respect to the measure $\hbox{Inv}$. More precisely, we prove that randomized Quicksort performs expected $O(n (1 + \log(1 + \hbox{Inv} / n)))$ element swaps, where $\hbox{Inv}$ denotes the number of inversions in the input sequence. This result provides a theoretical explanation for the observed behavior and gives new insights on the behavior of Quicksort. We also give some empirical results on the adaptive behavior of Heapsort and Mergesort.", acknowledgement = ack-nhfb, articleno = "3.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "adaptive sorting; branch mispredictions; Quicksort", } @Article{Codenotti:2008:ESD, author = "Bruno Codenotti and Benton Mccune and Sriram Pemmaraju and Rajiv Raman and Kasturi Varadarajan", title = "An experimental study of different approaches to solve the market equilibrium problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "3.3:1--3.3:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1402295", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Over the last few years, the problem of computing market equilibrium prices for exchange economies has received much attention in the theoretical computer science community. Such activity led to a flurry of polynomial time algorithms for various restricted, yet significant, settings. The most important restrictions arise either when the traders' utility functions satisfy a property known as {\em gross substitutability\/} or when the initial endowments are proportional (the Fisher model). In this paper, we experimentally compare the performance of some of these recent algorithms against that of the most used software packages. In particular, we evaluate the following approaches: (1) the solver PATH, available under GAMS/MPSGE, a popular tool for computing market equilibrium prices; (2) a discrete version of a simple iterative price update scheme called t{\^a}tonnement; (3) a discrete version of the welfare adjustment process; (4) convex feasibility programs that characterize the equilibrium in some special cases. We analyze the performance of these approaches on models of exchange economies where the consumers are equipped with utility functions, which are widely used in real world applications. The outcomes of our experiments consistently show that many market settings allow for an efficient computation of the equilibrium, well beyond the restrictions under which the theory provides polynomial time guarantees. For some of the approaches, we also identify models where they are prone to failure.", acknowledgement = ack-nhfb, articleno = "3.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "market equilibrium", } @Article{Dementiev:2008:BEM, author = "Roman Dementiev and Juha K{\"a}rkk{\"a}inen and Jens Mehnert and Peter Sanders", title = "Better external memory suffix array construction", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "3.4:1--3.4:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1402296", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Suffix arrays are a simple and powerful data structure for text processing that can be used for full text indexes, data compression, and many other applications, in particular, in bioinformatics. However, so far, it has appeared prohibitive to build suffix arrays for huge inputs that do not fit into main memory. This paper presents design, analysis, implementation, and experimental evaluation of several new and improved algorithms for suffix array construction. The algorithms are asymptotically optimal in the worst case or on average. Our implementation can construct suffix arrays for inputs of up to 4-GB in hours on a low-cost machine. As a tool of possible independent interest, we present a systematic way to design, analyze, and implement {\em pipelined\/} algorithms.", acknowledgement = ack-nhfb, articleno = "3.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithm engineering; algorithms for strings; external memory; I/O-efficient; large data sets; secondary memory; suffix array", } @Article{Swenson:2008:ATE, author = "Krister M. Swenson and Mark Marron and Joel V. Earnest-Deyoung and Bernard M. E. Moret", title = "Approximating the true evolutionary distance between two genomes", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "3.5:1--3.5:??", month = jun, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1227161.1402297", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "As more and more genomes are sequenced, evolutionary biologists are becoming increasingly interested in evolution at the level of whole genomes, in scenarios in which the genome evolves through insertions, duplications, deletions, and movements of genes along its chromosomes. In the mathematical model pioneered by Sankoff and others, a unichromosomal genome is represented by a signed permutation of a multiset of genes; Hannenhalli and Pevzner showed that the edit distance between two signed permutations of the same set can be computed in polynomial time when all operations are inversions. El-Mabrouk extended that result to allow deletions and a limited form of insertions (which forbids duplications); in turn we extended it to compute a nearly optimal edit sequence between an arbitrary genome and the identity permutation. In this paper we generalize our approach to compute distances between two arbitrary genomes, but focus on approximating the true evolutionary distance rather than the edit distance. We present experimental results showing that our algorithm produces excellent estimates of the true evolutionary distance up to a (high) threshold of saturation; indeed, the distances thus produced are good enough to enable the simple neighbor-joining procedure to reconstruct our test trees with high accuracy.", acknowledgement = ack-nhfb, articleno = "3.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "duplications; evolution; inversions; pairwise distances; whole-genome data", } @Article{Krommidas:2008:ESA, author = "Ioannis Krommidas and Christos Zaroliagis", title = "An experimental study of algorithms for fully dynamic transitive closure", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "12", pages = "16:1--16:??", month = jun, year = "2008", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1370596.1370597", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:17:58 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We have conducted an extensive experimental study on algorithms for fully dynamic transitive closure. We have implemented the recent fully dynamic algorithms by King [1999], Roditty [2003], Roditty and Zwick [2002, 2004], and Demetrescu and Italiano [2000, 2005] along with several variants and compared them to pseudo fully dynamic and simple-minded algorithms developed in a previous study [Frigioni et al. 2001]. We tested and compared these implementations on random inputs, synthetic (worst-case) inputs, and on inputs motivated by real-world graphs. Our experiments reveal that some of the dynamic algorithms can really be of practical value in many situations.", acknowledgement = ack-nhfb, articleno = "16", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "dynamic algorithm; path; reachability; transitive closure", } @Article{Gottlob:2008:BBA, author = "Georg Gottlob and Marko Samer", title = "A backtracking-based algorithm for hypertree decomposition", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "1:1--1:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412229", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Hypertree decompositions of hypergraphs are a generalization of tree decompositions of graphs. The corresponding hypertree-width is a measure for the acyclicity and therefore an indicator for the tractability of the associated computation problem. Several NP-hard decision and computation problems are known to be tractable on instances whose structure is represented by hypergraphs of bounded hypertree-width. Roughly speaking, the smaller the hypertree-width, the faster the computation problem can be solved. In this paper, we present the new backtracking-based algorithm det-$k$-decomp for computing hypertree decompositions of small width. Our benchmark evaluations have shown that det-$k$-decomp significantly outperforms opt- $k$-decomp, the only exact hypertree decomposition algorithm so far. Even compared to the best heuristic algorithm, we obtained competitive results as long as the hypergraphs are sufficiently simple.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "constraint satisfaction; hypertree decomposition", } @Article{Raman:2008:P, author = "Rajeev Raman and Matt Stallmann", title = "Preface", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "1:1--1:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412235", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gramm:2008:DRE, author = "Jens Gramm and Jiong Guo and Falk H{\"u}ffner and Rolf Niedermeier", title = "Data reduction and exact algorithms for clique cover", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "2:1--2:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412236", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "To cover the edges of a graph with a minimum number of cliques is an NP-hard problem with many applications. For this problem we develop efficient and effective polynomial-time data reduction rules that, combined with a search tree algorithm, allow for exact problem solutions in competitive time. This is confirmed by experiments with real-world and synthetic data. Moreover, we prove the fixed-parameter tractability of covering edges by cliques.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "clique cover; data reduction; fixed-parameter tractability", } @Article{Haran:2008:ESP, author = "Idit Haran and Dan Halperin", title = "An experimental study of point location in planar arrangements in {CGAL}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "3:1--3:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412237", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study the performance in practice of various point-location algorithms implemented in CGAL (the Computational Geometry Algorithms Library), including a newly devised {\em landmarks\/} algorithm. Among the other algorithms studied are: a na{\"\i}ve approach, a ``walk along a line'' strategy, and a trapezoidal decomposition-based search structure. The current implementation addresses general arrangements of planar curves, including arrangements of nonlinear segments (e.g., conic arcs) and allows for degenerate input (for example, more than two curves intersecting in a single point or overlapping curves). The algorithms use exact geometric computation and thus result in the correct point location. In our landmarks algorithm (a.k.a. jump \& walk), special points, ``landmarks,'' are chosen in a preprocessing stage, their place in the arrangement is found, and they are inserted into a data structure that enables efficient nearest-neighbor search. Given a query point, the nearest landmark is located and a ``walk'' strategy is applied from the landmark to the query point. We report on various experiments with arrangements composed of line segments or conic arcs. The results indicate that compared to the other algorithms tested, the landmarks approach is the most efficient, when the overall (amortized) cost of a query is taken into account, combining both preprocessing and query time. The simplicity of the algorithm enables an almost straightforward implementation and rather easy maintenance. The generic programming implementation allows versatility both in the selected type of landmarks and in the choice of the nearest-neighbor search structure. The end result is an efficient point-location algorithm that bypasses the alternative CGAL implementations in most practical aspects.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "arrangements; CGAL; computational geometry; generic programming; point location", } @Article{Lanthier:2008:CAC, author = "Mark A. Lanthier and Doron Nussbaum and Tsuo-Jung Wang", title = "Computing an approximation of the $1$-center problem on weighted terrain surfaces", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "3:1--3:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412231", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this article, we discuss the problem of determining a meeting point of a set of scattered robots $R = r_1, r _2, \ldots{}, r_s$ in a weighted terrain $P$, which has $n > s$ triangular faces. Our algorithmic approach is to produce a discretization of $P$ by producing a graph $G = \{V^G, E^G\}$, which lies on the surface of $P$. For a chosen vertex $p' \in V^G$, we define $|\Pi(r_i, p')|$ as the minimum weight cost of traveling from $r_i$ to $p'$. We show that min$_{p'} \in V^G$ \hbox{max}$_{1\leq i \leq s} |\Pi(r_i, p')| \leq \hbox{min}_p *\in P \hbox{max}_{1\leq i \leq s} |{\Pi}(r_i, p*)| + 2 W |L|$, where $L$ is the longest edge of $P$, $W$ is the maximum cost weight of a face of $P$, and $p*$ is the optimal solution. Our algorithm requires $O(s n m \log(s n m) + s n m^2)$ time to run, where $m = n$ in the Euclidean metric and $m = n^2$ in the weighted metric. However, we show, through experimentation, that only a constant value of $m$ is required (e.g., $m = 8$) in order to produce very accurate solutions.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "1-Center; algorithms; approximation; meeting point; robots; shortest path; terrain; weighted", } @Article{Hershberger:2008:SSD, author = "John Hershberger and Nisheeth Shrivastava and Subhash Suri", title = "Summarizing spatial data streams using {ClusterHulls}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "4:1--4:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412238", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the following problem: given an on-line, possibly unbounded stream of two-dimensional (2D) points, how can we summarize its spatial distribution or {\em shape\/} using a small, bounded amount of memory? We propose a novel scheme, called {\em ClusterHull}, which represents the shape of the stream as a dynamic collection of convex hulls, with a total of at most $m$ vertices, where $m$ is the size of the memory. The algorithm dynamically adjusts both the number of hulls and the number of vertices in each hull to best represent the stream using its fixed-memory budget. This algorithm addresses a problem whose importance is increasingly recognized, namely, the problem of summarizing real-time data streams to enable on-line analytical processing. As a motivating example, consider habitat monitoring using wireless sensor networks. The sensors produce a steady stream of geographic data, namely, the locations of objects being tracked. In order to conserve their limited resources (power, bandwidth, and storage), the sensors can compute, store, and exchange ClusterHull summaries of their data, without losing important geometric information. We are not aware of other schemes specifically designed for capturing shape information in geometric data streams and so we compare ClusterHull with some of the best general-purpose clustering schemes, such as CURE, $k$-medians, and LSEARCH. We show through experiments that ClusterHull is able to represent the shape of two-dimensional data streams more faithfully and flexibly than the stream versions of these clustering algorithms.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "convex hull; data streams; geometric data", } @Article{Safro:2008:MAL, author = "Ilya Safro and Dorit Ron and Achi Brandt", title = "Multilevel algorithms for linear ordering problems", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "4:1--4:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412232", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Linear ordering problems are combinatorial optimization problems that deal with the minimization of different functionals by finding a suitable permutation of the graph vertices. These problems are widely used and studied in many practical and theoretical applications. In this paper, we present a variety of linear--time algorithms for these problems inspired by the Algebraic Multigrid approach, which is based on weighted-edge contraction. The experimental result for four such problems turned out to be better than every known result in almost all cases, while the short (linear) running time of the algorithms enables testing very large graphs.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algebraic multigrid; linear ordering; multilevel algorithm", } @Article{Holzer:2008:EMO, author = "Martin Holzer and Frank Schulz and Dorothea Wagner", title = "Engineering multilevel overlay graphs for shortest-path queries", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "13", pages = "5:1--5:??", month = sep, year = "2008", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1412228.1412239", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Oct 6 16:18:31 MDT 2008", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "An overlay graph of a given graph $G$ = ($V$, $E$) on a subset $S \subseteq V$ is a graph with vertex set $S$ and edges corresponding to shortest paths in $G$. In particular, we consider variations of the multilevel overlay graph used in Schulz et al. [2002] to speed up shortest-path computation. In this work, we follow up and present several vertex selection criteria, along with two general strategies of applying these criteria, to determine a subset $S$ of a graph's vertices. The main contribution is a systematic experimental study where we investigate the impact of selection criteria and strategies on multilevel overlay graphs and the resulting speed-up achieved for shortest-path computation: Depending on selection strategy and graph type, a centrality index criterion, selection based on planar separators, and vertex degree turned out to perform best.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "Dijkstra's algorithm; hierarchical; multilevel; overlay graph; preprocessing; shortest path; speed-up technique; vertex selection", } @Article{Julstrom:2009:GHB, author = "Bryant A. Julstrom", title = "Greedy heuristics for the bounded diameter minimum spanning tree problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "1:1--1:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Given a connected, weighted, undirected graph G and a bound $D$, the bounded diameter minimum spanning tree problem seeks a spanning tree on $G$ of minimum weight among the trees in which no path between two vertices contains more than $D$ edges. In Prim's algorithm, the diameter of the growing spanning tree can always be known, so it is a good starting point from which to develop greedy heuristics for the bounded diameter problem. Abdalla, Deo, and Gupta described such an algorithm. It imitates Prim's algorithm but avoids edges whose inclusion in the spanning tree would violate the diameter bound. Running the algorithm from one start vertex requires time that is $O(n^3)$. A modification of this approach uses the start vertex as the center of the spanning tree (if $D$ is even) or as one of the two center vertices (if $D$ is odd). This yields a simpler algorithm whose time is $O(n^2)$. A further modification chooses each next vertex at random rather than greedily, though it still connects each vertex to the growing tree with the lowest-weight feasible edge. On Euclidean problem instances with small diameter bounds, the randomized heuristic is superior to the two fully greedy algorithms, though its advantage fades as the diameter bound grows. On instances whose edge weights have been chosen at random, the fully greedy algorithms outperform the randomized heuristic.", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Munro:2009:PSS, author = "J. Ian Munro and Dorothea Wagner", title = "Preface: Section 2 --- Selected Papers from {ALENEX 2008}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "1:1--1:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Dumitriu:2009:HMG, author = "Daniel Dumitriu and Stefan Funke and Martin Kutz and Nikola Milosavljevi{\'c}", title = "How much geometry it takes to reconstruct a $2$-manifold in {$R^3$}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "2:1--2:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Known algorithms for reconstructing a 2-manifold from a point sample in $R^3$ are naturally based on decisions/predicates that take the geometry of the point sample into account. Facing the always present problem of round-off errors that easily compromise the exactness of those predicate decisions, an exact and robust implementation of these algorithms is far from being trivial and typically requires employment of advanced datatypes for exact arithmetic, as provided by libraries like CORE, LEDA, or GMP. In this article, we present a new reconstruction algorithm, one whose main novelties is to throw away geometry information early on in the reconstruction process and to mainly operate combinatorially on a graph structure. More precisely, our algorithm only requires distances between the sample points and not the actual embedding in $R^3$. As such, it is less susceptible to robustness problems due to round-off errors and also benefits from not requiring expensive exact arithmetic by faster running times. A more theoretical view on our algorithm including correctness proofs under suitable sampling conditions can be found in a companion article.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Irving:2009:FLS, author = "Robert W. Irving and David F. Manlove", title = "Finding large stable matchings", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "2:1--2:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "When ties and incomplete preference lists are permitted in the stable marriage and hospitals/residents problems, stable matchings can have different sizes. The problem of finding a maximum cardinality stable matching in this context is known to be NP-hard, even under very severe restrictions on the number, size, and position of ties. In this article, we present two new heuristics for finding large stable matchings in variants of these problems in which ties are on one side only. We describe an empirical study involving these heuristics and the best existing approximation algorithm for this problem. Our results indicate that all three of these algorithms perform significantly better than naive tie-breaking algorithms when applied to real-world and randomly-generated data sets and that one of the new heuristics fares slightly better than the other algorithms, in most cases. This study, and these particular problem variants, are motivated by important applications in large-scale centralized matching schemes.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Basu:2009:GAO, author = "Amitabh Basu and Joseph S. B. Mitchell and Girish Kumar Sabhnani", title = "Geometric algorithms for optimal airspace design and air traffic controller workload balancing", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "3:1--3:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The National Airspace System (NAS) is designed to accommodate a large number of flights over North America. For purposes of workload limitations for air traffic controllers, the airspace is partitioned into approximately 600 sectors; each sector is observed by one or more controllers. In order to satisfy workload limitations for controllers, it is important that sectors be designed carefully according to the traffic patterns of flights, so that no sector becomes overloaded. We formulate and study the airspace sectorization problem from an algorithmic point-of-view, modeling the problem of optimal sectorization as a geometric partition problem with constraints. The novelty of the problem is that it partitions data consisting of trajectories of moving points, rather than static point set partitioning that is commonly studied. First, we formulate and solve the 1D version of the problem, showing how to partition a line into ``sectors'' (intervals) according to historical trajectory data. Then, we apply the 1D solution framework to design a 2D sectorization heuristic based on binary space partitions. We also devise partitions based on balanced ``pie partitions'' of a convex polygon. We evaluate our 2D algorithms experimentally, applying our algorithms to actual historical flight track data for the NAS. We compare the workload balance of our methods to that of the existing set of sectors for the NAS and find that our resectorization yields competitive and improved workload balancing. In particular, our methods yield an improvement by a factor between 2 and 3 over the current sectorization in terms of the time-average and the worst-case workloads of the maximum workload sector. An even better improvement is seen in the standard deviations (over all sectors) of both time-average and worst-case workloads.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Bauer:2009:SFR, author = "Reinhard Bauer and Daniel Delling", title = "{SHARC}: Fast and robust unidirectional routing", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "4:1--4:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "During recent years, impressive speed-up techniques for Dijkstra's have been developed. Unfortunately, the most advanced techniques use bidirectional search, which makes it hard to use them in scenarios where a backward search is prohibited. Even worse, such scenarios are widely spread (e.g., timetable-information systems or time-dependent networks). In this work, we present a unidirectional speed-up technique, which competes with bidirectional approaches. Moreover, we show how to exploit the advantage of unidirectional routing for fast exact queries in timetable information systems and for fast approximative queries in time-dependent scenarios. By running experiments on several inputs other than road networks, we show that our approach is very robust to the input.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Coleman:2009:RTL, author = "Tom Coleman and Anthony Wirth", title = "Ranking tournaments: Local search and a new algorithm", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "6:1--6:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Ranking is a fundamental activity for organizing and, later, understanding data. Advice of the form ``$a$ should be ranked before $b$'' is given. If this advice is consistent, and complete, then there is a total ordering on the data and the ranking problem is essentially a sorting problem. If the advice is consistent, but incomplete, then the problem becomes topological sorting. If the advice is inconsistent, then we have the feedback arc set (FAS) problem: The aim is then to rank a set of items to satisfy as much of the advice as possible. An instance in which there is advice about every pair of items is known as a tournament. This ranking task is equivalent to ordering the nodes of a given directed graph from left to right, while minimizing the number of arcs pointing left. In the past, much work focused on finding good, effective heuristics for solving the problem. Recently, a proof of the NP-completeness of the problem (even when restricted to tournaments) has accompanied new algorithms with approximation guarantees, culminating in the development of a PTAS (polynomial time approximation scheme) for solving FAS on tournaments. In this article, we reexamine many existing algorithms and develop some new techniques for solving FAS. The algorithms are tested on both synthetic and nonsynthetic datasets. We find that, in practice, local-search algorithms are very powerful, even though we prove that they do not have approximation guarantees. Our new algorithm is based on reversing arcs whose nodes have large in-degree differences, eventually leading to a total ordering. Combining this with a powerful local-search technique yields an algorithm that is as strong, or stronger than, existing techniques on a variety of data sets.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Cherkassky:2009:SPF, author = "Boris V. Cherkassky and Loukas Georgiadis and Andrew V. Goldberg and Robert E. Tarjan and Renato F. Werneck", title = "Shortest-path feasibility algorithms: An experimental evaluation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", pages = "7:1--7:??", month = may, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jun 3 16:21:43 MDT 2009", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "This is an experimental study of algorithms for the shortest-path feasibility problem: Given a directed weighted graph, find a negative cycle or present a short proof that none exists. We study previously known and new algorithms. Our testbed is more extensive than those previously used, including both static and incremental problems, as well as worst-case instances. We show that, while no single algorithm dominates, a small subset (including new algorithms) has very robust performance in practice. Our work advances the state of the art in the area.", acknowledgement = ack-nhfb, articleno = "7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Demetrescu:2009:P, author = "Camil Demetrescu", title = "Preface", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "1:1--1:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Serna:2009:PSS, author = "Maria Serna and Carme {\'A}lvarez", title = "Preface to special section of selected papers from {WEA 2006}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "1:1--1:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Maue:2009:GDS, author = "Jens Maue and Peter Sanders and Domagoj Matijevic", title = "Goal-directed shortest-path queries using precomputed cluster distances", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "2:1--2:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We demonstrate how Dijkstra's algorithm for shortest path queries can be accelerated by using precomputed shortest path distances. Our approach allows a completely flexible tradeoff between query time and space consumption for precomputed distances. In particular, sublinear space is sufficient to give the search a strong ``sense of direction''. We evaluate our approach experimentally using large, real-world road networks.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Valimaki:2009:ECS, author = "N. V{\"a}lim{\"a}ki and V. M{\"a}kinen and W. Gerlach and K. Dixit", title = "Engineering a compressed suffix tree implementation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "2:1--2:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Suffix tree is one of the most important data structures in string algorithms and biological sequence analysis. Unfortunately, when it comes to implementing those algorithms and applying them to real genomic sequences, often the main memory size becomes the bottleneck. This is easily explained by the fact that while a DNA sequence of length $n$ from alphabet $\Sigma = \{ A, C, G, T \}$ can be stored in $n \log |\Sigma| = 2 n$ bits, its suffix tree occupies $O(n \log n)$ bits. In practice, the size difference easily reaches factor 50. We report on an implementation of the compressed suffix tree very recently proposed by Sadakane (2007). The compressed suffix tree occupies space proportional to the text size, that is, $O(n \log |\Sigma|)$ bits, and supports all typical suffix tree operations with at most $\log n$ factor slowdown. Our experiments show that, for example, on a 10 MB DNA sequence, the compressed suffix tree takes 10\% of the space of the normal suffix tree. At the same time, a representative algorithm is slowed down by factor 30. Our implementation follows the original proposal in spirit, but some internal parts are tailored toward practical implementation. Our construction algorithm has time requirement $O(n \log n \log |\Sigma|)$ and uses closely the same space as the final structure while constructing it: on the 10MB DNA sequence, the maximum space usage during construction is only 1.5 times the final product size. As by-products, we develop a method to create Succinct Suffix Array directly from Burrows--Wheeler transform and a space-efficient version of the suffixes-insertion algorithm to build balanced parentheses representation of suffix tree from LCP information.", acknowledgement = ack-nhfb, articleno = "2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Eisenbrand:2009:ALO, author = "Friedrich Eisenbrand and Andreas Karrenbauer and Chihao Xu", title = "Algorithms for longer {OLED} lifetime", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "3:1--3:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider an optimization problem arising in the design of controllers for OLED displays. Our objective is to minimize the amplitude of the electrical current flowing through the diodes, which has a direct impact on the lifetime of such a display. The optimization problem consist of finding a decomposition of an image into subframes with special structural properties that allow the display driver to lower the stress on the diodes. For monochrome images, we present an algorithm that finds an optimal solution of this problem in linear time. Moreover, we consider an online version of the problem in which we have to take a decision for one row based on a constant number of rows in the lookahead. In this framework, this algorithm has a tight competitive ratio. A generalization of this algorithm computes near-optimal solutions of real-world instances in real time.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Englert:2009:EOS, author = "Matthias Englert and Heiko R{\"o}glin and Matthias Westermann", title = "Evaluation of online strategies for reordering buffers", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "3:1--3:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A sequence of objects that are characterized by their color has to be processed. Their processing order influences how efficiently they can be processed: Each color change between two consecutive objects produces costs. A reordering buffer, which is a random access buffer with storage capacity for k objects, can be used to rearrange this sequence online in such a way that the total costs are reduced. This concept is useful for many applications in computer science and economics. The strategy with the best-known competitive ratio is MAP. An upper bound of $O(\log k)$ on the competitive ratio of MAP is known and a nonconstant lower bound on the competitive ratio is not known. Based on theoretical considerations and experimental evaluations, we give strong evidence that the previously used proof techniques are not suitable to show an $o (\sqrt{\log k})$ upper bound on the competitive ratio of MAP. However, we also give some evidence that in fact MAP achieves a competitive ratio of $O(1)$. Further, we evaluate the performance of several strategies on random input sequences experimentally. MAP and its variants RC and RR clearly outperform the other strategies FIFO, LRU, and MCF. In particular, MAP, RC, and RR are the only known strategies whose competitive ratios do not depend on the buffer size. Furthermore, MAP achieves the smallest competitive ratio.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Farshi:2009:ESG, author = "Mohammad Farshi and Joachim Gudmundsson", title = "Experimental study of geometric $t$-spanners", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "3:1--3:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The construction of t -spanners of a given point set has received a lot of attention, especially from a theoretical perspective. In this article, we experimentally study the performance and quality of the most common construction algorithms for points in the Euclidean plane. We implemented the most well-known t -spanner algorithms and tested them on a number of different point sets. The experiments are discussed and compared to the theoretical results, and in several cases, we suggest modifications that are implemented and evaluated. The measures of quality that we consider are the number of edges, the weight, the maximum degree, the spanner diameter, and the number of crossings. This is the first time an extensive comparison has been made between the running times of construction algorithms of t -spanners and the quality of the generated spanners.", acknowledgement = ack-nhfb, articleno = "3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Cederman:2009:GQP, author = "Daniel Cederman and Philippas Tsigas", title = "{GPU-Quicksort}: a practical {Quicksort} algorithm for graphics processors", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "4:1--4:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this article, we describe GPU-Quicksort, an efficient Quicksort algorithm suitable for highly parallel multicore graphics processors. Quicksort has previously been considered an inefficient sorting solution for graphics processors, but we show that in CUDA, NVIDIA's programming platform for general-purpose computations on graphical processors, GPU-Quicksort performs better than the fastest-known sorting implementations for graphics processors, such as radix and bitonic sort. Quicksort can thus be seen as a viable alternative for sorting large quantities of data on graphics processors.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Chimani:2009:EEC, author = "Markus Chimani and Carsten Gutwenger and Petra Mutzel", title = "Experiments on exact crossing minimization using column generation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "4:1--4:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The crossing number of a graph G is the smallest number of edge crossings in any drawing of G into the plane. Recently, the first branch-and-cut approach for solving the crossing number problem has been presented in Buchheim et al. [2005]. Its major drawback was the huge number of variables out of which only very few were actually used in the optimal solution. This restricted the algorithm to rather small graphs with low crossing number. In this article, we discuss two column generation schemes; the first is based on traditional algebraic pricing, and the second uses combinatorial arguments to decide whether and which variables need to be added. The main focus of this article is the experimental comparison between the original approach and these two schemes. In addition, we evaluate the quality achieved by the best-known crossing number heuristic by comparing the new results with the results of the heuristic.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Putze:2009:CHS, author = "Felix Putze and Peter Sanders and Johannes Singler", title = "Cache-, hash-, and space-efficient {Bloom} filters", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "4:1--4:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A Bloom filter is a very compact data structure that supports approximate membership queries on a set, allowing false positives. We propose several new variants of Bloom filters and replacements with similar functionality. All of them have a better cache-efficiency and need less hash bits than regular Bloom filters. Some use SIMD functionality, while the others provide an even better space efficiency. As a consequence, we get a more flexible trade-off between false-positive rate, space-efficiency, cache-efficiency, hash-efficiency, and computational effort. We analyze the efficiency of Bloom filters and the proposed replacements in detail, in terms of the false-positive rate, the number of expected cache-misses, and the number of required hash bits. We also describe and experimentally evaluate the performance of highly tuned implementations. For many settings, our alternatives perform better than the methods proposed so far.", acknowledgement = ack-nhfb, articleno = "4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Chimani:2009:OOC, author = "Markus Chimani and Maria Kandyba and Ivana Ljubi{\'c} and Petra Mutzel", title = "Obtaining optimal $k$-cardinality trees fast", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "2.5:1--2.5:23", month = dec, year = "2009", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1498698.1537600", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Given an undirected graph G = (V, E) with edge weights and a positive integer number k, the k -cardinality tree problem consists of finding a subtree T of G with exactly k edges and the minimum possible weight. Many algorithms have been proposed to solve this NP-hard problem, resulting in mainly heuristic and metaheuristic approaches. In this article, we present an exact ILP-based algorithm using directed cuts. We mathematically compare the strength of our formulation to the previously known ILP formulations of this problem, and show the advantages of our approach. Afterwards, we give an extensive study on the algorithm's practical performance compared to the state-of-the-art metaheuristics. In contrast to the widespread assumption that such a problem cannot be efficiently tackled by exact algorithms for medium and large graphs (between 200 and 5,000 nodes), our results show that our algorithm not only has the advantage of proving the optimality of the computed solution, but also often outperforms the metaheuristic approaches in terms of running time.", acknowledgement = ack-nhfb, articleno = "2.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Frias:2009:LRC, author = "Leonor Frias and Jordi Petit and Salvador Roura", title = "Lists revisited: Cache-conscious {STL} lists", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "5:1--5:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present three cache-conscious implementations of STL standard compliant lists. Until now, one could either find simple doubly linked list implementations that easily cope with standard strict requirements, or theoretical approaches that do not take into account any of these requirements in their design. In contrast, we have merged both approaches, paying special attention to iterators constraints. In this article, the competitiveness of our implementations is evinced with an extensive experimental analysis. This shows, for instance, 5 to 10 times faster traversals and 3 to 5 times faster internal sort.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Holzer:2009:EPS, author = "Martin Holzer and Frank Schulz and Dorothea Wagner and Grigorios Prasinos and Christos Zaroliagis", title = "Engineering planar separator algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "5:1--5:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider classical linear-time planar separator algorithms, determining for a given planar graph a small subset of its nodes whose removal divides the graph into two components of similar size. These algorithms are based on planar separator theorems, which guarantee separators of size $O(\sqrt n)$ and remaining components of size at most $2 n / 3$ (where $n$ denotes the number of nodes in the graph). In this article, we present a comprehensive experimental study of the classical algorithms applied to a large variety of graphs, where our main goal is to find separators that do not only satisfy upper bounds, but also possess other desirable characteristics with respect to separator size and component balance. We achieve this by investigating a number of specific alternatives for the concrete implementation and fine-tuning of certain parts of the classical algorithms. It is also shown that the choice of several parameters influences the separation quality considerably. Moreover, we propose as planar separators the usage of fundamental cycles, whose size is at most twice the diameter of the graph: For graphs of small diameter, the guaranteed bound is better than the $O(\sqrt n)$ bounds, and it turns out that this simple strategy almost always outperforms the other algorithms, even for graphs with large diameter.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Tarjan:2009:DTP, author = "Robert E. Tarjan and Renato F. Werneck", title = "Dynamic trees in practice", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "5:1--5:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Dynamic tree data structures maintain forests that change over time through edge insertions and deletions. Besides maintaining connectivity information in logarithmic time, they can support aggregation of information over paths, trees, or both. We perform an experimental comparison of several versions of dynamic trees: ST-trees, ET-trees, RC-trees, and two variants of top trees (self-adjusting and worst-case). We quantify their strengths and weaknesses through tests with various workloads, most stemming from practical applications. We observe that a simple, linear-time implementation is remarkably fast for graphs of small diameter, and that worst-case and randomized data structures are best when queries are very frequent. The best overall performance, however, is achieved by self-adjusting ST-trees.", acknowledgement = ack-nhfb, articleno = "5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Avdil:2009:LSS, author = "Alaubek Avdil and Karsten Weihe", title = "Local search starting from an {LP} solution: Fast and quite good", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "6:1--6:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present and evaluate a specific way to generate good start solutions for local search. The start solution is computed from a certain LP, which is related to the underlying problem. We consider three optimization problems: the directed MAX-CUT problem with a source and a sink and two variations of the MAX- k -SAT problem with k = 2 and k = 3. To compare our technique, we run local search repeatedly with random start solutions. Our technique produces, consistently, final solutions whose objective values are not too far from the best solutions from repeated random starts. The surprising degree of stability and uniformity of this result throughout all of our experiments on various classes of instances strongly suggests that we have consistently achieved nearly optimal solutions. On the other hand, the runtime of our technique is rather small, so the technique is very efficient and probably quite accurate.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Deineko:2009:FMW, author = "Vladimir Deineko and Alexander Tiskin", title = "Fast minimum-weight double-tree shortcutting for metric {TSP}: Is the best one good enough?", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "6:1--6:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The Metric Traveling Salesman Problem (TSP) is a classical NP-hard optimization problem. The double-tree shortcutting method for Metric TSP yields an exponentially-sized space of TSP tours, each of which approximates the optimal solution within, at most, a factor of 2. We consider the problem of finding among these tours the one that gives the closest approximation, that is, the minimum-weight double-tree shortcutting. Burkard et al. gave an algorithm for this problem, running in time $O(n^3 + 2^d n^2)$ and memory $O(2^d n^2)$, where $d$ is the maximum node degree in the rooted minimum spanning tree. We give an improved algorithm for the case of small $d$ (including planar Euclidean TSP, where $d \leq 4$), running in time $O(4^d n^2)$ and memory $O(4^d n)$. This improvement allows one to solve the problem on much larger instances than previously attempted. Our computational experiments suggest that in terms of the time-quality trade-off, the minimum-weight double-tree shortcutting method provides one of the best existing tour-constructing heuristics.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Figueroa:2009:SSA, author = "Karina Figueroa and Edgar Chavez and Gonzalo Navarro and Rodrigo Paredes", title = "Speeding up spatial approximation search in metric spaces", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "6:1--6:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Proximity searching consists of retrieving from a database those elements that are similar to a query object. The usual model for proximity searching is a metric space where the distance, which models the proximity, is expensive to compute. An index uses precomputed distances to speedup query processing. Among all the known indices, the baseline for performance for about 20 years has been AESA. This index uses an iterative procedure, where at each iteration it first chooses the next promising element (``pivot'') to compare to the query, and then it discards database elements that can be proved not relevant to the query using the pivot. The next pivot in AESA is chosen as the one minimizing the sum of lower bounds to the distance to the query proved by previous pivots. In this article, we introduce the new index iAESA, which establishes a new performance baseline for metric space searching. The difference with AESA is the method to select the next pivot. In iAESA, each candidate sorts previous pivots by closeness to it, and chooses the next pivot as the candidate whose order is most similar to that of the query. We also propose a modification to AESA-like algorithms to turn them into probabilistic algorithms. Our empirical results confirm a consistent improvement in query performance. For example, we perform as few as 60\% of the distance evaluations of AESA in a database of documents, a very important and difficult real-life instance of the problem. For the probabilistic algorithm, we perform in a database of faces up to 40\% of the comparisons made by the best alternative algorithm to retrieve the same percentage of the correct answer. Based on the empirical results, we conjecture that the new probabilistic AESA-like algorithms will become, as AESA had been for exact algorithms, a reference point establishing, in practice, a lower bound on how good a probabilistic proximity search algorithm can be.", acknowledgement = ack-nhfb, articleno = "6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Barbay:2009:EIS, author = "J{\'e}r{\'e}my Barbay and Alejandro L{\'o}pez-Ortiz and Tyler Lu and Alejandro Salinger", title = "An experimental investigation of set intersection algorithms for text searching", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "7:1--7:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The intersection of large ordered sets is a common problem in the context of the evaluation of boolean queries to a search engine. In this article, we propose several improved algorithms for computing the intersection of sorted arrays, and in particular for searching sorted arrays in the intersection context. We perform an experimental comparison with the algorithms from the previous studies from Demaine, L{\'o}pez-Ortiz, and Munro [ALENEX 2001] and from Baeza-Yates and Salinger [SPIRE 2005]; in addition, we implement and test the intersection algorithm from Barbay and Kenyon [SODA 2002] and its randomized variant [SAGA 2003]. We consider both the random data set from Baeza-Yates and Salinger, the Google queries used by Demaine et al., a corpus provided by Google, and a larger corpus from the TREC Terabyte 2006 efficiency query stream, along with its own query log. We measure the performance both in terms of the number of comparisons and searches performed, and in terms of the CPU time on two different architectures. Our results confirm or improve the results from both previous studies in their respective context (comparison model on real data, and CPU measures on random data) and extend them to new contexts. In particular, we show that value-based search algorithms perform well in posting lists in terms of the number of comparisons performed.", acknowledgement = ack-nhfb, articleno = "7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Estivill-Castro:2009:RRD, author = "Vladimir Estivill-Castro and Apichat Heednacram and Francis Suraweera", title = "Reduction rules deliver efficient {FPT}-algorithms for covering points with lines", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "7:1--7:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present efficient algorithms to solve the Line Cover Problem exactly. In this NP-complete problem, the inputs are n points in the plane and a positive integer k, and we are asked to answer if we can cover these n points with at most k lines. Our approach is based on fixed-parameter tractability and, in particular, kernelization. We propose several reduction rules to transform instances of Line Cover into equivalent smaller instances. Once instances are no longer susceptible to these reduction rules, we obtain a problem kernel whose size is bounded by a polynomial function of the parameter k and does not depend on the size n of the input. Our algorithms provide exact solutions and are easy to implement. We also describe the design of algorithms to solve the corresponding optimization problem exactly. We experimentally evaluated ten variants of the algorithms to determine the impact and trade-offs of several reduction rules. We show that our approach provides tractability for a larger range of values of the parameter and larger inputs, improving the execution time by several orders of magnitude with respect to earlier algorithms that use less rules.", acknowledgement = ack-nhfb, articleno = "7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{DeLoera:2009:CMM, author = "Jes{\'u}s A. {De Loera} and David C. Haws and Jon Lee and Allison O'Hair", title = "Computation in multicriteria matroid optimization", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "14", number = "1", pages = "8:1--8:??", month = dec, year = "2009", CODEN = "????", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:04:28 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Motivated by recent work on algorithmic theory for nonlinear and multicriteria matroid optimization, we have developed algorithms and heuristics aimed at practical solution of large instances of some of these difficult problems. Our methods primarily use the local adjacency structure inherent in matroid polytopes to pivot to feasible solutions, which may or may not be optimal. We also present a modified breadth-first-search heuristic that uses adjacency to enumerate a subset of feasible solutions. We present other heuristics and provide computational evidence supporting our techniques. We implemented all of our algorithms in the software package MOCHA.", acknowledgement = ack-nhfb, articleno = "8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Jacobs:2010:ESR, author = "Tobias Jacobs", title = "An experimental study of recent hotlink assignment algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "1.1:1--1.1:??", month = mar, year = "2010", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1671970.1671971", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:05:50 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The concept of {\em hotlink assignment\/} aims at enhancing the structure of Web sites such that the user's expected navigation effort is minimized. We concentrate on sites that are representable by trees and assume that each leaf carries a weight representing its popularity.\par The problem of optimally adding at most one additional outgoing edge (``hotlink'') to each inner node has been widely studied. A considerable number of approximation algorithms have been proposed and worst-case bounds for the quality of the computed solutions have been given. However, only little is known about the practical behavior of most of these algorithms.\par This article contributes to closing this gap by evaluating all recently proposed strategies experimentally. Our experiments are based on trees extracted from real Web sites, as well as on synthetic instances. The latter are generated by a new method that simulates the growth of a Web site over time. Finally, we present a new heuristic that is easy to implement and exhibits excellent behavior in practice.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "approximation; hotlink; Search tree", } @Article{Spence:2010:SGS, author = "Ivor Spence", title = "{{\tt sgen1}}: a generator of small but difficult satisfiability benchmarks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "1.2:1--1.2:??", month = mar, year = "2010", CODEN = "????", DOI = "http://doi.acm.org/10.1145/1671970.1671972", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:05:50 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The satisfiability problem is known to be NP-Complete; therefore, there should be relatively small problem instances that take a very long time to solve. However, most of the smaller benchmarks that were once thought challenging, especially the satisfiable ones, can be processed quickly by modern SAT-solvers. We describe and make available a generator that produces both unsatisfiable and, more significantly, satisfiable formulae that take longer to solve than any others known. At the two most recent international SAT Competitions, the smallest unsolved benchmarks were created by this generator. We analyze the results of all solvers in the most recent competition when applied to these benchmarks and also present our own more focused experiments.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "SAT-solvers; Satisfiability benchmarks", } @Article{Langguth:2010:HIB, author = "Johannes Langguth and Fredrik Manne and Peter Sanders", title = "Heuristic initialization for bipartite matching problems", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "1.3:1--1.3:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1712655.1712656", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Mar 15 12:05:50 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "It is a well-established result that improved pivoting in linear solvers can be achieved by computing a bipartite matching between matrix entries and positions on the main diagonal. With the availability of increasingly faster linear solvers, the speed of bipartite matching computations must keep up to avoid slowing down the main computation. Fast algorithms for bipartite matching, which are usually initialized with simple heuristics, have been known for a long time. However, the performance of these algorithms is largely dependent on the quality of the heuristic. We compare combinations of several known heuristics and exact algorithms to find fast combined methods, using real-world matrices as well as randomly generated instances. In addition, we present a new heuristic aimed at obtaining high-quality matchings and compare its impact on bipartite matching algorithms with that of other heuristics. The experiments suggest that its performance compares favorably to the best-known heuristics, and that it is especially suited for application in linear solvers.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "Heuristics; matching", } @Article{Delbot:2010:AEC, author = "Fran{\c{c}}ois Delbot and Christian Laforest", title = "Analytical and experimental comparison of six algorithms for the vertex cover problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "14:1--14:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1865970.1865971", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The vertex cover is a well-known NP-complete minimization problem in graphs that has received a lot of attention these last decades. Many algorithms have been proposed to construct vertex cover in different contexts (offline, online, list algorithms, etc.) leading to solutions of different level of quality. This quality is traditionally measured in terms of approximation ratio, that is, the worst possible ratio between the quality of the solution constructed and the optimal one. For the vertex cover problem the range of such known ratios are between 2 (conjectured as being the smallest constant ratio) and $\Delta$, the maximum degree of the graph. Based on this measure of quality, the hierarchy is almost clear (the smaller the ratio is, the better the algorithm is). In this article, we show that this measure, although of great importance, is too macroscopic and does not reflect the practical behavior of the methods. We prove this by analyzing (known and recent) algorithms running on a particular class of graphs: the paths. We obtain closed and exact formulas for the mean of the sizes of vertex cover constructed by these different algorithms. Then, we assess their quality experimentally in several well-chosen class of graphs (random, regular, trees, BHOSLIB benchmarks, trap graphs, etc.). The synthesis of all these results lead us to formulate a ``practical hierarchy'' of the algorithms. We remark that it is, more or less, the opposite to the one only based on approximation ratios, showing that worst-case analysis only gives partial information on the quality of an algorithm.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Arroyuelo:2010:PAR, author = "Diego Arroyuelo and Gonzalo Navarro", title = "Practical approaches to reduce the space requirement of {Lempel--Ziv}-based compressed text indices", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "15:1--15:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1883684", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Given a text $T [1.\,.n]$ over an alphabet of size $\sigma$, the full-text search problem consists in locating the occ occurrences of a given pattern $P[1.\,.m]$ in $T$. Compressed full-text self-indices are space-efficient representations of the text that provide direct access to and indexed search on it.\par The LZ-index of Navarro is a compressed full-text self-index based on the LZ78 compression algorithm. This index requires about 5 times the size of the compressed text (in theory, $4 n H_k(T) + o(n \log \sigma)$ bits of space, where $H_k(T)$ is the $k$-th order empirical entropy of $T$). In practice, the average locating complexity of the LZ-index is $O(\sigma m \log_\sigma n + {\rm occ} \sigma^{m / 2})$, where {\em occ} is the number of occurrences of $P$. It can extract text substrings of length $l$ in $O(l)$ time. This index outperforms competing schemes both to locate short patterns and to extract text snippets. However, the LZ-index can be up to 4 times larger than the smallest existing indices (which use $n H_k(T) + o(n \log \sigma)$ bits in theory), and it does not offer space/time tuning options. This limits its applicability.\par In this article, we study practical ways to reduce the space of the LZ-index. We obtain new LZ-index variants that require $2(1 + \epsilon) n H_k(T) + o(n \log \sigma)$ bits of space, for any $0 < \epsilon < 1$. They have an average locating time of $O(1 / \epsilon (m \log n + {\rm occ} \sigma^{m / 2}))$, while extracting takes $O(l)$ time.\par We perform extensive experimentation and conclude that our schemes are able to reduce the space of the original LZ-index by a factor of $2/3$, that is, around $3$ times the compressed text size. Our schemes are able to extract about 1 to 2 MB of the text per second, being twice as fast as the most competitive alternatives. Pattern occurrences are located at a rate of up to 1 to 4 million per second. This constitutes the best space\slash time trade-off when indices are allowed to use 4 times the size of the compressed text or more.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Ullmann:2010:BVA, author = "Julian R. Ullmann", title = "Bit-vector algorithms for binary constraint satisfaction and subgraph isomorphism", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "16:1--16:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1921702", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A solution to a binary constraint satisfaction problem is a set of discrete values, one in each of a given set of domains, subject to constraints that allow only prescribed pairs of values in specified pairs of domains. Solutions are sought by backtrack search interleaved with a process that removes from domains those values that are currently inconsistent with provisional choices already made in the course of search. For each value in a given domain, a bit-vector shows which values in another domain are or are not permitted in a solution. Bit-vector representation of constraints allows bit-parallel, therefore fast, operations for editing domains during search. This article revises and updates bit-vector algorithms published in the 1970's, and introduces focus search, which is a new bit-vector algorithm relying more on search and less on domain-editing than previous algorithms. Focus search is competitive within a limited family of constraint satisfaction problems. Determination of subgraph isomorphism is a specialized binary constraint satisfaction problem for which bit-vector algorithms have been widely used since the 1980s, particularly for matching molecular structures. This article very substantially updates the author's 1976 subgraph isomorphism algorithm, and reports experimental results with random and real-life data.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Askitis:2010:RSH, author = "Nikolas Askitis and Justin Zobel", title = "Redesigning the string hash table, burst trie, and {BST} to exploit cache", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "17:1--17:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1921704", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A key decision when developing in-memory computing applications is choice of a mechanism to store and retrieve strings. The most efficient current data structures for this task are the hash table with move-to-front chains and the burst trie, both of which use linked lists as a substructure, and variants of binary search tree. These data structures are computationally efficient, but typical implementations use large numbers of nodes and pointers to manage strings, which is not efficient in use of cache. In this article, we explore two alternatives to the standard representation: the simple expedient of including the string in its node, and, for linked lists, the more drastic step of replacing each list of nodes by a contiguous array of characters. Our experiments show that, for large sets of strings, the improvement is dramatic. For hashing, in the best case the total space overhead is reduced to less than 1 bit per string. For the burst trie, over 300MB of strings can be stored in a total of under 200MB of memory with significantly improved search time. These results, on a variety of data sets, show that cache-friendly variants of fundamental data structures can yield remarkable gains in performance.", acknowledgement = ack-nhfb, articleno = "1.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{McGeoch:2010:P, author = "Catherine C. McGeoch", title = "Preface", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "2.1:1--2.1:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1671974", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Chimani:2010:LFU, author = "Markus Chimani and Carsten Gutwenger and Petra Mutzel and Hoi-Ming Wong", title = "Layer-free upward crossing minimization", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "2.2:1--2.2:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1671975", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "An upward drawing of a DAG $G$ is a drawing of $G$ in which all arcs are drawn as curves increasing monotonically in the vertical direction. In this article, we present a new approach for upward crossing minimization, that is, finding an upward drawing of a DAG $G$ with as few crossings as possible. Our algorithm is based on a two-stage upward planarization approach, which computes a feasible upward planar subgraph in the first step and reinserts the remaining arcs by computing constraint-feasible upward insertion paths. An experimental study shows that the new algorithm leads to much better results than existing algorithms for upward crossing minimization, including the classical Sugiyama approach.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "Crossing number; planarization approach; upward drawing; upward planarization", } @Article{Bauer:2010:CHG, author = "Reinhard Bauer and Daniel Delling and Peter Sanders and Dennis Schieferdecker and Dominik Schultes and Dorothea Wagner", title = "Combining hierarchical and goal-directed speed-up techniques for {Dijkstra}'s algorithm", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "2.3:1--2.3:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1671976", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In recent years, highly effective hierarchical and goal-directed speed-up techniques for routing in large road networks have been developed. This article makes a systematic study of combinations of such techniques. These combinations turn out to give the best results in many scenarios, including graphs for unit disk graphs, grid networks, and time-expanded timetables. Besides these quantitative results, we obtain general insights for successful combinations.", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "Dijkstra's algorithm; speed-up technique", } @Article{Nash:2010:CID, author = "Nicholas Nash and David Gregg", title = "Comparing integer data structures for 32- and 64-bit keys", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "2.4:1--2.4:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1671977", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this article, we experimentally compare a number of data structures operating over keys that are 32- and 64-bit integers. We examine traditional comparison-based search trees as well as data structures that take advantage of the fact that the keys are integers such as van Emde Boas trees and various trie-based data structures. We propose a variant of a burst trie that performs better in time than all the alternative data structures. In addition, even for small sets of keys, this burst trie variant occupies less space than comparison-based data structures such as red-black trees and $B$-trees. Burst tries have previously been shown to provide a very efficient base for implementing cache efficient string sorting algorithms. We find that with suitable engineering, they also perform excellently as a dynamic ordered data structure operating over integer keys. We provide experimental results when the data structures operate over uniform random data. We also present experimental results for other types of data, including datasets arising from {\em Valgrind}, a widely used suite of tools for the dynamic binary instrumentation of programs.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "Integer keys; level compression; searching; trees; tries", } @Article{Sinha:2010:EBT, author = "Ranjan Sinha and Anthony Wirth", title = "Engineering burstsort: Toward fast in-place string sorting", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "15", number = "1", pages = "2.5:1--2.5:??", month = mar, year = "2010", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1671970.1671978", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon Dec 10 09:03:03 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Burstsort is a trie-based string sorting algorithm that distributes strings into small buckets whose contents are then sorted in cache. This approach has earlier been demonstrated to be efficient on modern cache-based processors [Sinha \& Zobel, JEA 2004]. In this article, we introduce improvements that reduce by a significant margin the memory requirement of Burstsort: It is now less than 1\% greater than an in-place algorithm. These techniques can be applied to existing variants of Burstsort, as well as other string algorithms such as for string management.\par We redesigned the buckets, introducing sub-buckets and an index structure for them, which resulted in an order-of-magnitude space reduction. We also show the practicality of moving some fields from the trie nodes to the insertion point (for the next string pointer) in the bucket; this technique reduces memory usage of the trie nodes by one-third. Importantly, the trade-off for the reduction in memory use is only a very slight increase in the running time of Burstsort on real-world string collections. In addition, during the bucket-sorting phase, the string suffixes are copied to a small buffer to improve their spatial locality, lowering the running time of Burstsort by up to 30\%. These memory usage enhancements have enabled the copy-based approach [Sinha et al., JEA 2006] to also reduce the memory usage with negligible impact on speed.", acknowledgement = ack-nhfb, articleno = "2.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", keywords = "algorithms; cache; experimental algorithms; Sorting; string management; tries", } @Article{Boytsov:2011:IMA, author = "Leonid Boytsov", title = "Indexing methods for approximate dictionary searching: Comparative analysis", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "1.1:1--1.1:??", month = may, year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.1963191", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon May 30 08:26:05 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The primary goal of this article is to survey state-of-the-art indexing methods for approximate dictionary searching. To improve understanding of the field, we introduce a taxonomy that classifies all methods into direct methods and sequence-based filtering methods. We focus on infrequently updated dictionaries, which are used primarily for retrieval. Therefore, we consider indices that are optimized for retrieval rather than for update. The indices are assumed to be associative, that is, capable of storing and retrieving auxiliary information, such as string identifiers. All solutions are lossless and guarantee retrieval of strings within a specified edit distance $k$. Benchmark results are presented for the practically important cases of $k = 1, 2$, and $3$.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Biro:2011:SMC, author = "P{\'e}ter Bir{\'o} and Robert W. Irving and Ildik{\'o} Schlotter", title = "Stable matching with couples: an empirical study", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "12:1--12:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.1970372", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In practical applications, algorithms for the classic version of the hospitals residents problem (the many-one version of the stable marriage problem) may have to be extended to accommodate the needs of couples who wish to be allocated to (geographically) compatible places. Such an extension has been in operation in the National Resident Matching Problem (NRMP) matching scheme in the United States for a number of years. In this setting, a stable matching need not exist, and it is an NP-complete problem to decide if one does. However, the only previous empirical study in this context (focused on the NRMP algorithm), together with information from NRMP, suggest that, in practice, stable matchings do exist and that an appropriate heuristic can be used to find such a matching. The study presented here was motivated by the recent decision to accommodate couples in the Scottish Foundation Allocation Scheme (SFAS), the Scottish equivalent of the NRMP. Here, the problem is a special case, since hospital preferences are derived from a ``master list'' of resident scores, but we show that the existence problem remains NP-complete in this case. We describe the algorithm used in SFAS and contrast it with a version of the algorithm that forms the basis of the NRMP approach. We also propose a third simpler algorithm based on satisfying blocking pairs, and an FPT algorithm when the number of couples is viewed as a parameter. We present an empirical study of the performance of a number of variants of these algorithms using a range of datasets. The results indicate that, not surprisingly, increasing the ratio of couples to single applicants typically makes it harder to find a stable matching (and, by inference, less likely that a stable matching exists). However, the likelihood of finding a stable matching is very high for realistic values of this ratio, and especially so for particular variants of the algorithms.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Huber:2011:MGS, author = "Stefan Huber and Martin Held", title = "Motorcycle graphs: Stochastic properties motivate an efficient yet simple implementation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "13:1--13:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2019578", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this article, we study stochastic properties of a geometric setting that underpins random motorcycle graphs and use it to motivate a simple but very efficient algorithm for computing motorcycle graphs. An analysis of the mean trace length of $n$ random motorcycles suggests that, on average, a motorcycle crosses only a constant number of cells within a $\sqrt n \times \sqrt n$ rectangular grid, provided that the motorcycles are distributed sufficiently uniformly over the area covered by the grid. This analysis motivates a simple algorithm for computing motorcycle graphs: We use the standard priority-queue--based algorithm and enhance it with geometric hashing by means of a rectangular grid. If the motorcycles are distributed sufficiently uniformly, then our stochastic analysis predicts an $O(n \log n)$ runtime. Indeed, extensive experiments run on 22,000 synthetic and real-world datasets confirm a runtime of less than $10^{-5} n \log n$ seconds for the vast majority of our datasets on a standard PC. Further experiments with our software, Moca, also confirm the mean trace length and average number of cells crossed by a motorcycle, as predicted by our analysis. This makes Moca the first implementation that is efficient enough to be applied in practice for computing motorcycle graphs of large datasets. Actually, it is easy to extend Moca to make it compute a generalized version of the original motorcycle graph, thus enabling a significantly larger field of applications.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Michail:2011:ECS, author = "Dimitrios Michail", title = "An experimental comparison of single-sided preference matching algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "14:1--14:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2019579", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We experimentally study the problem of assigning applicants to posts. Each applicant provides a preference list, which may contain ties, ranking a subset of the posts. Different optimization criteria may be defined, which depend on the desired solution properties. The main focus of this work is to assess the quality of matchings computed by rank-maximal and popular matching algorithms and compare this with the minimum weight matching algorithm, which is a standard matching algorithm that is used in practice. Both rank-maximal and popular matching algorithms use common algorithmic techniques, which makes them excellent candidates for a running time comparison. Since popular matchings do not always exist, we also study the unpopularity of matchings computed by the aforementioned algorithms. Finally, extra criteria like total weight and cardinality are included, due to their importance in practice. All experiments are performed using structured random instances as well as instances created using real-world datasets.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Kot:2011:ECP, author = "Andriy Kot and Andrey N. Chernikov and Nikos P. Chrisochoides", title = "Effective out-of-core parallel {Delaunay} mesh refinement using off-the-shelf software", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "15:1--15:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2019580", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present three related out-of-core parallel mesh generation algorithms and their implementations for small size computational clusters. Computing out-of-core permits to solve larger problems than otherwise possible on the same hardware setup. Also, when using shared computing resources with high demand, a problem can take longer to compute in terms of wall-clock time when using an in-core algorithm on many nodes instead of using an out-of-core algorithm on few nodes. The difference is due to wait-in-queue delays that can grow exponentially to the number of requested nodes. In one specific case, using our best method and only 16 nodes it can take several times less wall-clock time to generate a 2 billion element mesh than to generate the same size mesh in-core with 121 nodes. Although our best out-of-core method exhibits unavoidable overheads (could be as low as 19\% in some cases) over the corresponding in-core method (for mesh sizes that fit completely in-core), this is a modest and expected performance penalty. We evaluated our methods on traditional clusters of workstations as well as presented preliminary performance evaluation on [the] emerging BlueWaters supercomputer.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Prosser:2011:LDS, author = "Patrick Prosser and Chris Unsworth", title = "Limited discrepancy search revisited", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "16:1--16:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2019581", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Harvey and Ginsberg's limited discrepancy search (LDS) is based on the assumption that costly heuristic mistakes are made early in the search process. Consequently, LDS repeatedly probes the state space, going against the heuristic (i.e., taking discrepancies) a specified number of times in all possible ways and attempts to take those discrepancies as early as possible. LDS was improved by Richard Korf, to become improved LDS (ILDS), but in doing so, discrepancies were taken as late as possible, going against the original assumption. Many subsequent algorithms have faithfully inherited Korf's interpretation of LDS, and take discrepancies late. This then raises the question: Should we take our discrepancies late or early? We repeat the original experiments performed by Harvey and Ginsberg and those by Korf in an attempt to answer this question. We also investigate the early stopping condition of the YIELDS algorithm, demonstrating that it is simple, elegant and efficient.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Tabourier:2011:GCR, author = "Lionel Tabourier and Camille Roth and Jean-Philippe Cointet", title = "Generating constrained random graphs using multiple edge switches", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "17:1--17:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2063515", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The generation of random graphs using edge swaps provides a reliable method to draw uniformly random samples of sets of graphs respecting some simple constraints (e.g., degree distributions). However, in general, it is not necessarily possible to access all graphs obeying some given constraints through a classical switching procedure calling on pairs of edges. Therefore, we propose to get around this issue by generalizing this classical approach through the use of higher-order edge switches. This method, which we denote by ``$k$-edge switching,'' makes it possible to progressively improve the covered portion of a set of constrained graphs, thereby providing an increasing, asymptotically certain confidence on the statistical representativeness of the obtained sample.", acknowledgement = ack-nhfb, articleno = "1.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Tsourakakis:2011:AAS, author = "Charalampos E. Tsourakakis and Richard Peng and Maria A. Tsiarli and Gary L. Miller and Russell Schwartz", title = "Approximation algorithms for speeding up dynamic programming and denoising {aCGH} data", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "18:1--18:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2063517", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The development of cancer is largely driven by the gain or loss of subsets of the genome, promoting uncontrolled growth or disabling defenses against it. Denoising array-based Comparative Genome Hybridization (aCGH) data is an important computational problem central to understanding cancer evolution. In this article, we propose a new formulation of the denoising problem that we solve with a ``vanilla'' dynamic programming algorithm, which runs in $O(n^2)$ units of time. Then, we propose two approximation techniques. Our first algorithm reduces the problem into a well-studied geometric problem, namely halfspace emptiness queries, and provides an $\epsilon$ additive approximation to the optimal objective value in $\tilde{O}(n ^{4 / 3 + \delta} \log (U / \epsilon))$ time, where $\delta$ is an arbitrarily small positive constant and $U = \max\{\sqrt C, (|P_i|)_{i = 1,\ldots{}, n}\} (P = (P_1, P_2, \ldots{}, P_n), P_i \in \mathbb{R})$, is the vector of the noisy aCGH measurements, $C$ a normalization constant. The second algorithm provides a $(1 \pm \epsilon)$ approximation (multiplicative error) and runs in $O(n \log n / \epsilon)$ time. The algorithm decomposes the initial problem into a small (logarithmic) number of Monge optimization subproblems that we can solve in linear time using existing techniques. Finally, we validate our model on synthetic and real cancer datasets. Our method consistently achieves superior precision and recall to leading competitors on the data with ground truth. In addition, it finds several novel markers not recorded in the benchmarks but supported in the oncology literature.", acknowledgement = ack-nhfb, articleno = "1.8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Vahrenhold:2011:P, author = "Jan Vahrenhold", title = "Preface", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "21:1--21:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.1970374", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Wang:2011:CEM, author = "Bei Wang and Herbert Edelsbrunner and Dmitriy Morozov", title = "Computing elevation maxima by searching the {Gauss} sphere", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "22:1--22:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.1970375", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The elevation function on a smoothly embedded 2-manifold in R$^3$ reflects the multiscale topography of cavities and protrusions as local maxima. The function has been useful in identifying coarse docking configurations for protein pairs. Transporting the concept from the smooth to the piecewise linear category, this article describes an algorithm for finding all local maxima. While its worst-case running time is the same as of the algorithm used in prior work, its performance in practice is orders of magnitudes superior. We cast light on this improvement by relating the running time to the total absolute Gaussian curvature of the 2-manifold.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Rotta:2011:MLS, author = "Randolf Rotta and Andreas Noack", title = "Multilevel local search algorithms for modularity clustering", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "23:1--23:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.1970376", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Modularity is a widely used quality measure for graph clusterings. Its exact maximization is NP-hard and prohibitively expensive for large graphs. Popular heuristics first perform a coarsening phase, where local search starting from singleton clusters is used to compute a preliminary clustering, and then optionally a refinement phase, where this clustering is improved by moving vertices between clusters. As a generalization, multilevel heuristics coarsen in several stages, and refine by moving entire clusters from each of these stages, not only individual vertices. This article organizes existing and new single-level and multilevel heuristics into a coherent design space, and compares them experimentally with respect to their effectiveness (achieved modularity) and runtime. For coarsening by iterated cluster joining, it turns out that the most widely used criterion for joining clusters (modularity increase) is outperformed by other simple criteria, that a recent multistep algorithm [Schuetz and Caflisch 2008] is no improvement over simple single-step coarsening for these criteria, and that the recent multilevel coarsening by iterated vertex moving [Blondel et al. 2008] is somewhat faster but slightly less effective (with refinement). The new multilevel refinement is significantly more effective than the conventional single-level refinement or no refinement, in reasonable runtime. A comparison with published benchmark results and algorithm implementations shows that multilevel local search heuristics, despite their relative simplicity, are competitive with the best algorithms in the literature.", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Bertasi:2011:PYA, author = "Paolo Bertasi and Marco Bressan and Enoch Peserico", title = "{{\tt psort}}, yet another fast stable sorting software", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "24:1--24:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.1970377", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "{\tt psort} is the fastest sorting software according to the PennySort benchmark, sorting 181GB of data in 2008 and 224GB in 2009 for \$0.01 of computer time. This article details its internals, and the careful fitting of its architecture to the structure of modern PC-class platforms, allowing it to outperform state-of-the-art sorting software such as STXXL sort.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Finocchi:2011:GEF, author = "Irene Finocchi and John Hershberger", title = "Guest editors' foreword", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "31:1--31:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2025377", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "3.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Belazzougui:2011:TPM, author = "Djamal Belazzougui and Paolo Boldi and Rasmus Pagh and Sebastiano Vigna", title = "Theory and practice of monotone minimal perfect hashing", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "32:1--32:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2025378", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Minimal perfect hash functions have been shown to be useful to compress data in several data management tasks. In particular, order-preserving minimal perfect hash functions (Fox et al. 1991) have been used to retrieve the position of a key in a given list of keys; however, the ability to preserve any given order leads to an unavoidable $\Omega(n \log n)$ lower bound on the number of bits required to store the function. Recently, it was observed (Belazzougui et al. 2009) that very frequently the keys to be hashed are sorted in their intrinsic (i.e., lexicographical) order. This is typically the case of dictionaries of search engines, list of URLs of Web graphs, and so on. We refer to this restricted version of the problem as monotone minimal perfect hashing. We analyze experimentally the data structures proposed in Belazzougui et al. [2009], and along our way we propose some new methods that, albeit asymptotically equivalent or worse, perform very well in practice and provide a balance between access speed, ease of construction, and space usage.", acknowledgement = ack-nhfb, articleno = "3.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Doerr:2011:QRS, author = "Benjamin Doerr and Tobias Friedrich and Marvin K{\"u}nnemann and Thomas Sauerwald", title = "Quasirandom rumor spreading: an experimental analysis", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "33:1--33:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2025379", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We empirically analyze two versions of the well-known ``randomized rumor spreading'' protocol to disseminate a piece of information in networks. In the classical model, in each round, each informed node informs a random neighbor. In the recently proposed quasirandom variant, each node has a (cyclic) list of its neighbors. Once informed, it starts at a random position of the list, but from then on informs its neighbors in the order of the list. While for sparse random graphs a better performance of the quasirandom model could be proven, all other results show that, independent of the structure of the lists, the same asymptotic performance guarantees hold as for the classical model. In this work, we compare the two models experimentally. Not only does this show that the quasirandom model generally is faster, but it also shows that the runtime is more concentrated around the mean. This is surprising given that much fewer random bits are used in the quasirandom process. These advantages are also observed in a lossy communication model, where each transmission does not reach its target with a certain probability, and in an asynchronous model, where nodes send at random times drawn from an exponential distribution. We also show that typically the particular structure of the lists has little influence on the efficiency.", acknowledgement = ack-nhfb, articleno = "3.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Haverkort:2011:FDH, author = "Herman Haverkort and Freek V. Walderveen", title = "Four-dimensional {Hilbert} curves for {$R$}-trees", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "34:1--34:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2025380", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Two-dimensional R-trees are a class of spatial index structures in which objects are arranged to enable fast window queries: report all objects that intersect a given query window. One of the most successful methods of arranging the objects in the index structure is based on sorting the objects according to the positions of their centers along a two-dimensional Hilbert space-filling curve. Alternatively, one may use the coordinates of the objects' bounding boxes to represent each object by a four-dimensional point, and sort these points along a four-dimensional Hilbert-type curve. In experiments by Kamel and Faloutsos and by Arge et al., the first solution consistently outperformed the latter when applied to point data, while the latter solution clearly outperformed the first on certain artificial rectangle data. These authors did not specify which four-dimensional Hilbert-type curve was used; many exist. In this article, we show that the results of the previous articles can be explained by the choice of the four-dimensional Hilbert-type curve that was used and by the way it was rotated in four-dimensional space. By selecting a curve that has certain properties and choosing the right rotation, one can combine the strengths of the two-dimensional and the four-dimensional approach into one, while avoiding their apparent weaknesses. The effectiveness of our approach is demonstrated with experiments on various datasets. For real data taken from VLSI design, our new curve yields R-trees with query times that are better than those of R-trees that were obtained with previously used curves.", acknowledgement = ack-nhfb, articleno = "3.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Negrucseri:2011:SMF, author = "Cosmin Silvestru Negrucseri and Mircea Bogdan Pacsosi and Barbara Stanley and Clifford Stein and Cristian George Strat", title = "Solving maximum flow problems on real-world bipartite graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "35:1--35:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2025381", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this article, we present an experimental study of several maximum-flow algorithms in the context of unbalanced bipartite networks. Our experiments are motivated by a real-world problem of managing reservation-based inventory in Google content ad systems. We are interested in observing the performance of several push-relabel algorithms on our real-world datasets and also on some generated ones. Previous work suggested an important improvement for push-relabel algorithms on unbalanced bipartite networks: the two-edge push rule. We show how the two-edge push rule improves the running time. While no single algorithm dominates the results, we show there is one that has very robust performance in practice.", acknowledgement = ack-nhfb, articleno = "3.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Tazari:2011:DLH, author = "Siamak Tazari and Matthias M{\"u}ller-Hannemann", title = "Dealing with large hidden constants: engineering a {Planar Steiner Tree (PTAS)}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "16", number = "1", pages = "36:1--36:??", year = "2011", CODEN = "????", DOI = "http://dx.doi.org/10.1145/1963190.2025382", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Sat Feb 25 18:02:18 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We present the first attempt on implementing a highly theoretical polynomial-time approximation scheme (PTAS) with huge hidden constants, namely, the PTAS for Steiner tree in planar graphs by Borradaile, Klein, and Mathieu (2009). Whereas this result, and several other PTAS results of the recent years, are of high theoretical importance, no practical applications or even implementation attempts have been known to date, due to the extremely large constants that are involved in them. We describe techniques on how to circumvent the challenges in implementing such a scheme. With today's limitations on processing power and space, we still have to sacrifice approximation guarantees for improved running times by choosing some parameters empirically. But our experiments show that with our choice of parameters, we do get the desired approximation ratios, suggesting that a much tighter analysis might be possible. Our computational experiments with benchmark instances from SteinLib and large artificial instances well exceeded our own expectations. We demonstrate that we are able to handle instances with up to a million nodes and several hundreds of terminals in 1.5 hours on a standard PC. On the rectilinear preprocessed instances from SteinLib, we observe a monotonous improvement for smaller values of $\epsilon$, with an average gap below 1\% for $\epsilon = 0.1$. We compare our implementation against the well-known batched $1$-Steiner heuristic and observe that on very large instances, we are able to produce comparable solutions much faster. We also present a thorough experimental evaluation of the influence of the various parameters of the PTAS and thus obtain a better understanding of their empirical effects.", acknowledgement = ack-nhfb, articleno = "3.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Reams:2012:AFD, author = "Charles Reams", title = "{Anatree}: a Fast Data Structure for Anagrams", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "17", number = "1", pages = "1.1:1--1.1:??", month = mar, year = "2012", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2133803.2133804", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jan 21 07:42:23 MST 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Natural language is a rich source of constraint satisfaction problems (CSPs), with a uniquely structured solution domain. We describe a number of approaches to satisfying the particular case of unordered letter-level constraints, including anagrams, but also relevant to typographical error correction, password security and word puzzles among other fields. We define the anatree, a data structure that can solve many such problems in constant time with respect to the size of the lexicon. The structure represents the lexicon of a language in a format somewhat analogous to a binary decision diagram (BDD) and, as with BDDs, construction heuristics allow the real average-case performance to vastly exceed the theoretical worst case. We compare anatrees and their alternatives empirically, explore the behavior of the construction heuristics, and characterize the tasks for which each is best suited.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Geisberger:2012:RPF, author = "Robert Geisberger and Michael N. Rice and Peter Sanders and Vassilis J. Tsotras", title = "Route planning with flexible edge restrictions", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "17", number = "1", pages = "1.2:1--1.2:??", month = mar, year = "2012", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2133803.2133805", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed Jan 21 07:42:23 MST 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this work, we explore a new type of flexible shortest-path query, in which the query can be dynamically parameterized to constrain the type of edges that may be included in the resulting shortest path (e.g., find the shortest path in a road network that avoids toll roads and low overpasses, respective of the specified vehicle height). We extend the hierarchical preprocessing technique known as Contraction Hierarchies to efficiently support such flexible queries. We also present several effective algorithmic optimizations for further improving the overall scalability and query times of this approach, including the addition of goal-directed search techniques, search space pruning techniques, and generalizing the constraints of the local search. Experiments are presented for both the North American and the European road networks, showcasing the general effectiveness and scalability of our proposed methodology to large-scale, real-world graphs.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Abraham:2013:ARR, author = "Ittai Abraham and Daniel Delling and Andrew V. Goldberg and Renato F. Werneck", title = "Alternative routes in road networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "1.3:1--1.3:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2444019", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon May 6 18:55:51 MDT 2013", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study the problem of finding good alternative routes in road networks. We look for routes that are substantially different from the shortest path, have small stretch, and are locally optimal. We formally define the problem of finding alternative routes with a single via vertex, develop efficient algorithms for it, and evaluate them experimentally. Our algorithms are efficient enough for practical use and compare favorably with previous methods in both speed and solution quality.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Batz:2013:MTD, author = "G. Veit Batz and Robert Geisberger and Peter Sanders and Christian Vetter", title = "Minimum time-dependent travel times with contraction hierarchies", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "1.4:1--1.4:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2444020", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon May 6 18:55:51 MDT 2013", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Time-dependent road networks are represented as weighted graphs, where the weight of an edge depends on the time one passes through that edge. This way, we can model periodic congestions during rush hour and similar effects. In this work we deal with the special case where edge weights are time-dependent travel times. Namely, we consider two problems in this setting: Earliest arrival queries ask for a minimum travel time route for a start and a destination depending on a given departure time. Travel time profile queries ask for the travel time profile for a start, a destination, and an interval of possible departure times. For an instance representing the German road network, for example, we can answer earliest arrival queries in less than 1.5ms. For travel time profile queries, which are much harder to answer, we need less than 40ms if the interval of possible departure times has a width of 24 hours. For inexact travel time profiles with an allowed error of about 1\% this even reduces to 3.2ms. The underlying hierarchical representations of the road network, which are variants of a time-dependent contraction hierarchy (TCH), need less than 1GiB of space and can be generated in about 30 minutes. As far as we know, TCHs are currently the only method being able to answer travel time profile queries efficiently. Altogether, with TCHs, web servers with massive request traffic are able to provide fast time-dependent earliest arrival route planning and computation of travel time profiles.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Bonami:2013:BRC, author = "Pierre Bonami and Jon Lee and Sven Leyffer and Andreas W{\"a}chter", title = "On branching rules for convex mixed-integer nonlinear optimization", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "2.6:1--2.6:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2532568", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Branch-and-Bound (B\&B) is perhaps the most fundamental algorithm for the global solution of convex Mixed-Integer Nonlinear Programming (MINLP) problems. It is well-known that carrying out branching in a nonsimplistic manner can greatly enhance the practicality of B\&B in the context of Mixed-Integer Linear Programming (MILP). No detailed study of branching has heretofore been carried out for MINLP. In this article, we study and identify useful sophisticated branching methods for MINLP, including novel approaches based on approximations of the nonlinear relaxations by linear and quadratic programs.", acknowledgement = ack-nhfb, articleno = "2.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Canzar:2013:PDA, author = "Stefan Canzar and Khaled Elbassioni and Juli{\'a}n Mestre", title = "A polynomial-delay algorithm for enumerating approximate solutions to the interval constrained coloring problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "2.2:1--2.2:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2493372", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study the interval constrained coloring problem, a combinatorial problem arising in the interpretation of data on protein structure emanating from experiments based on hydrogen/deuterium exchange and mass spectrometry. The problem captures the challenging task of increasing the spatial resolution of experimental data in order to get a better picture of the protein structure. Since solutions proposed by any algorithmic framework have to ultimately be verified by biochemists, it is important to provide not just a single solution, but a valuable set of candidate solutions. Our contribution is a polynomial-delay, polynomial-space algorithm for enumerating all exact solutions plus further approximate solutions, which are guaranteed to be within an absolute error of two of the optimum within fragments of the protein, that is, within sets of consecutive residues. Our experiments indicate that the quality of the approximate solutions is comparable to the optimal ones in terms of deviation from the underlying true solution. In addition, the experiments also confirm the effectiveness of the method in reducing the delay between two consecutive solutions considerably, compared to what it takes an integer programming solver to produce the next exact solution.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Delort:2013:HDP, author = "Charles Delort and Olivier Spanjaard", title = "A hybrid dynamic programming approach to the biobjective binary knapsack problem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "1.2:1--1.2:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2444018", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "This article is devoted to a study of the impact of using bound sets in biobjective dynamic programming. This notion, introduced by Villareal and Karwan [1981], has been independently revisited by Ehrgott and Gandibleux [2007], as well as by Sourd and Spanjaard [2008]. The idea behind it is very general and can, therefore, be adapted to a wide range of biobjective combinatorial problems. We focus here on the biobjective binary knapsack problem. We show that using bound sets to perform a hybrid dynamic programming procedure embedded in a two-phase method [Ulungu and Teghem 1995] yields numerical results that outperform previous dynamic programming approaches to the problem, both in execution times and memory requirements.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Eppstein:2013:LAM, author = "David Eppstein and Maarten L{\"o}ffler and Darren Strash", title = "Listing All Maximal Cliques in Large Sparse Real-World Graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "??", pages = "3.1:1--3.1:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2543629", ISSN = "1084-6654", bibdate = "Wed Jan 21 07:35:03 MST 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "3.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Ferraro-Petrillo:2013:DSR, author = "Umberto Ferraro-Petrillo and Fabrizio Grandoni and Giuseppe F. Italiano", title = "Data structures resilient to memory faults: an experimental study of dictionaries", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "1.6:1--1.6:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2444022", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon May 6 18:55:51 MDT 2013", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We address the problem of implementing data structures resilient to memory faults, which may arbitrarily corrupt memory locations. In this framework, we focus on the implementation of dictionaries and perform a thorough experimental study using a testbed that we designed for this purpose. Our main discovery is that the best-known (asymptotically optimal) resilient data structures have very large space overheads. More precisely, most of the space used by these data structures is not due to key storage. This might not be acceptable in practice, since resilient data structures are meant for applications where a huge amount of data (often of the order of terabytes) has to be stored. Exploiting techniques developed in the context of resilient (static) sorting and searching, in combination with some new ideas, we designed and engineered an alternative implementation, which, while still guaranteeing optimal asymptotic time and space bounds, performs much better in terms of memory without compromising the time efficiency.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Festa:2013:FSI, author = "Paola Festa", title = "Foreword to the special issue {SEA 2010}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "1.1:1--1.1:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2444017", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gog:2013:CST, author = "Simon Gog and Enno Ohlebusch", title = "Compressed suffix trees: Efficient computation and storage of {LCP}-values", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "2.1:1--2.1:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2461327", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The suffix tree is a very important data structure in string processing, but typical implementations suffer from huge space consumption. In large-scale applications, compressed suffix trees (CSTs) are therefore used instead. A CST consists of three (compressed) components: the suffix array, the longest common prefix (LCP)-array and data structures for simulating navigational operations on the suffix tree. The LCP-array stores the lengths of the LCPs of lexicographically adjacent suffixes, and it can be computed in linear time. In this article, we present a new LCP-array construction algorithm that is fast and very space efficient. In practice, our algorithm outperforms alternative algorithms. Moreover, we introduce a new compressed representation of LCP-arrays.", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gorke:2013:DGC, author = "Robert G{\"o}rke and Pascal Maillard and Andrea Schumm and Christian Staudt and Dorothea Wagner", title = "Dynamic graph clustering combining modularity and smoothness", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "1.5:1--1.5:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2444021", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Mon May 6 18:55:51 MDT 2013", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Maximizing the quality index modularity has become one of the primary methods for identifying the clustering structure within a graph. Since many contemporary networks are not static but evolve over time, traditional static approaches can be inappropriate for specific tasks. In this work, we pioneer the NP-hard problem of online dynamic modularity maximization. We develop scalable dynamizations of the currently fastest and the most widespread static heuristics and engineer a heuristic dynamization of an optimal static algorithm. Our algorithms efficiently maintain a modularity -based clustering of a graph for which dynamic changes arrive as a stream. For our quickest heuristic we prove a tight bound on its number of operations. In an experimental evaluation on both a real-world dynamic network and on dynamic clustered random graphs, we show that the dynamic maintenance of a clustering of a changing graph yields higher modularity than recomputation, guarantees much smoother clustering dynamics, and requires much lower runtimes. We conclude with giving sound recommendations for the choice of an algorithm.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Hofri:2013:OSS, author = "Micha Hofri", title = "Optimal selection and sorting via dynamic programming", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "2.3:1--2.3:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2444016.2493373", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We show how to find optimal algorithms for the selection of one or more order statistics over a small set of numbers, and as an extreme case, complete sorting. The criterion is using the smallest number of comparisons; separate derivations are performed for minimization on the average (over all permutations) or in the worst case. When the computational process establishes the optimal values, it also generates C-language functions that implement policies which achieve those optimal values. The search for the algorithms is driven by a Markov decision process, and the program provides the optimality proof as well.", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Kouri:2013:FRM, author = "Tina M. Kouri and Dinesh P. Mehta", title = "Faster reaction mapping through improved naming techniques", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "2.5:1--2.5:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2532569", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Automated reaction mapping is an important tool in cheminformatics where it may be used to classify reactions or validate reaction mechanisms. The reaction mapping problem is known to be NP-Hard and may be formulated as an optimization problem. In this article, we present four algorithms that continue to obtain optimal solutions to this problem, but with significantly improved runtimes over the previous Constructive Count Vector (CCV) algorithm. Our algorithmic improvements include (i) the use of a fast (but not 100\% accurate) canonical labeling algorithm, (ii) name reuse (i.e., storing intermediate results rather than recomputing), and (iii) an incremental approach to canonical name computation. The time to map the reactions from the Kegg/Ligand database previously took over 2 days using CCV, but now it takes fewer than 4 hours to complete. Experimental results on chemical reaction databases demonstrate our 2-CCV FDN MS algorithm usually performs over fifteen times faster than previous automated reaction mapping algorithms.", acknowledgement = ack-nhfb, articleno = "2.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Kouzinopoulos:2013:EOT, author = "Charalampos S. Kouzinopoulos and Konstantinos G. Margaritis", title = "Exact online two-dimensional pattern matching using multiple pattern matching algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "1", pages = "2.4:1--2.4:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2513148", ISSN = "1084-6654", ISSN-L = "1084-6654", bibdate = "Wed May 21 14:36:03 MDT 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Baker and Bird and Baeza-Yates and Regnier are two of the most efficient and widely used algorithms for exact online two-dimensional pattern matching. Both use the automaton of the Aho--Corasick multiple pattern matching algorithm to locate all the occurrences of a two-dimensional pattern in a two-dimensional input string, a data structure that is considered by many as inefficient, especially when used to process long patterns or data using large alphabet sizes. This article presents variants of the Baker and Bird and the Baeza-Yates and Regnier algorithms that use the data structures of the Set Horspool, Wu-Manber, Set Backward Oracle Matching, and SOG multiple pattern matching algorithms in place of the automaton of Aho--Corasick and evaluates their performance experimentally in terms of preprocessing and searching time.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Nagarajan:2013:EEI, author = "Chandrashekhar Nagarajan and David P. Williamson", title = "An Experimental Evaluation of Incremental and Hierarchical $k$-Median Algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "18", number = "??", pages = "3.2:1--3.2:??", month = dec, year = "2013", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2543628", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:23:43 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In this article, we consider different incremental and hierarchical $k$-median algorithms with provable performance guarantees and compare their running times and quality of output solutions on different benchmark $k$-median datasets. We determine that the quality of solutions output by these algorithms for all the datasets is much better than their performance guarantees suggest. Since some of the incremental $k$-median algorithms require approximate solutions for the $k$-median problem, we also compare some of the existing $k$-median algorithms running times and quality of solutions obtained on these datasets.", acknowledgement = ack-nhfb, articleno = "3.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gonzalez:2015:LCS, author = "Rodrigo Gonz{\'a}lez and Gonzalo Navarro and H{\'e}ctor Ferrada", title = "Locally Compressed Suffix Arrays", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.1:1--1.1:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2594408", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/datacompression.bib; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We introduce a compression technique for suffix arrays. It is sensitive to the compressibility of the text and local, meaning that random portions of the suffix array can be decompressed by accessing mostly contiguous memory areas. This makes decompression very fast, especially when various contiguous cells must be accessed. Our main technical contributions are the following. First, we show that runs of consecutive values that are known to appear in function $ \Psi (i) = A^{-1} [A [i] + 1] $ of suffix arrays $A$ of compressible texts also show up as repetitions in the differential suffix array $ A'[i] = A [i] - A [i - 1]$. Second, we use Re-Pair, a grammar-based compressor, to compress the differential suffix array, and upper bound its compression ratio in terms of the number of runs. Third, we show how to compact the space used by the grammar rules by up to 50\%, while still permitting direct access to the rules. Fourth, we develop specific variants of Re-Pair that work using knowledge of $ \Psi $, and use much less space than the general Re-Pair compressor, while achieving almost the same compression ratios. Fifth, we implement the scheme and compare it exhaustively with previous work, including the first implementations of previous theoretical proposals.", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Doerr:2015:RRP, author = "Benjamin Doerr and Magnus Wahlstr{\"o}m", title = "Randomized Rounding in the Presence of a Cardinality Constraint", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.2:1--1.2:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2594409", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the problem of generating randomized roundings that satisfy a single cardinality constraint and admit Chernoff-type large deviation bounds for weighted sums of the variables. That this can be done efficiently was proven by Srinivasan [2001], a different approach was later given by the first author [Doerr 2006]. In this work, we (a) present an improved version of the bitwise derandomization given by Doerr, (b) give the first derandomization of Srinivasan's tree-based randomized approach and prove its correctness, and (c) experimentally compare the resulting algorithms. Our experiments show that adding a single cardinality constraint typically reduces the rounding errors and only moderately increases the running times. In general, our derandomization of the tree-based approach is superior to the derandomized bitwise one, while the two randomized versions produce very similar rounding errors. When implementing the derandomized tree-based approach, however, the choice of the tree is important.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Auer:2015:EMC, author = "B. O. Fagginger Auer and R. H. Bisseling", title = "Efficient Matching for Column Intersection Graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.3:1--1.3:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2616587", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "To improve the quality and efficiency of hypergraph-based matrix partitioners, we investigate high-quality matchings in column intersection graphs of large sparse binary matrices. We show that such algorithms have a natural decomposition in an integer-weighted graph-matching function and a neighbor-finding function and study the performance of 16 combinations of these functions. We improve upon the original matching algorithm of the Mondriaan matrix partitioner: by using PGA', we improve the average matching quality from 95.3\% to 97.4\% of the optimum value; by using our new neighbor-finding heuristic, we obtain comparable quality and speedups of up to a factor of 19.6.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Angione:2015:SMB, author = "Claudio Angione and Annalisa Occhipinti and Giuseppe Nicosia", title = "Satisfiability by {Maxwell--Boltzmann} and {Bose--Einstein} Statistical Distributions", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.4:1--1.4:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2629498", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Recent studies in theoretical computer science have exploited new algorithms and methodologies based on statistical physics for investigating the structure and the properties of the Satisfiability (SAT) problem. We propose a characterization of the SAT problem as a physical system, using both quantum and classical statistical-physical models. We associate a graph to an SAT instance and we prove that a Bose--Einstein condensation occurs in the instance with higher probability if the quantum distribution is adopted in the generation of the graph. Conversely, the fit-get-rich behavior is more likely if we adopt the Maxwell--Boltzmann distribution. Our method allows a comprehensive analysis of the SAT problem based on a new definition of entropy of an instance, without requiring the computation of its truth assignments. The entropy of an SAT instance increases in the satisfiability region as the number of free variables in the instance increases. Finally, we develop six new solvers for the MaxSAT problem based on quantum and classical statistical distributions, and we test them on random SAT instances, with competitive results. We experimentally prove that the performance of the solvers based on the two distributions depends on the criterion used to flag clauses as satisfied in the SAT solving process.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Frieder:2015:ESA, author = "Asaf Frieder and Liam Roditty", title = "An Experimental Study on Approximating $k$ Shortest Simple Paths", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.5:1--1.5:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2630068", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We have conducted an extensive experimental study on approximation algorithms for computing $k$ shortest simple paths in weighted directed graphs. Very recently, Bernstein [2010] presented an algorithm that computes a $1 + \epsilon$ approximated $k$ shortest simple path in $O(\epsilon^{-1} k(m + n \log n) \log^2 n)$ time. We have implemented Bernstein's algorithm and tested it on synthetic inputs and real-world graphs (road maps). Our results reveal that Bernstein's algorithm has a practical value in many scenarios. Moreover, it produces in most of the cases exact paths rather than approximated. We also present a new variant for Bernstein's algorithm. We prove that our new variant has the same upper bounds for the running time and approximation as Bernstein's original algorithm. We have implemented and tested this variant as well. Our testing shows that this variant, which is based on a simple theoretical observation, is better than Bernstein's algorithm in practice.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gorke:2015:EDC, author = "Robert G{\"o}rke and Andrea Kappes and Dorothea Wagner", title = "Experiments on Density-Constrained Graph Clustering", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.6:1--1.6:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2638551", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Clustering a graph means identifying internally dense subgraphs that are only sparsely interconnected. Formalizations of this notion lead to measures that quantify the quality of a clustering and to algorithms that actually find clusterings. Since, most generally, corresponding optimization problems are hard, heuristic clustering algorithms are used in practice, or other approaches that are not based on an objective function. In this work, we conduct a comprehensive experimental evaluation of the qualitative behavior of greedy bottom-up heuristics driven by cut-based objectives and constrained by intracluster density, using both real-world data and artificial instances. Our study documents that a greedy strategy based on local movement is superior to one based on merging. We further reveal that the former approach generally outperforms alternative setups and reference algorithms from the literature in terms of its own objective, while a modularity-based algorithm competes surprisingly well. Finally, we exhibit which combinations of cut-based inter- and intracluster measures are suitable for identifying a hidden reference clustering in synthetic random graphs and discuss the skewness of the resulting cluster size distributions. Our results serve as a guideline to the usage of bicriterial, cut-based measures for graph clusterings.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Galvao:2015:ATG, author = "Gustavo Rodrigues Galv{\~a}o and Zanoni Dias", title = "An Audit Tool for Genome Rearrangement Algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.7:1--1.7:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2661633", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the combinatorial problem of sorting a permutation using a minimum number of rearrangement events, which finds application in the estimation of evolutionary distance between species. Many variants of this problem, which we generically refer to as the rearrangement sorting problem, have been tackled in the literature, and for most of them, the best known algorithms are approximations or heuristics. In this article, we present a tool, called GRAAu, to aid in the evaluation of the results produced by these algorithms. To illustrate its application, we use GRAAu to evaluate the results of four approximation algorithms regarding two variants of the rearrangement sorting problem: the problem of sorting by prefix reversals and the problem of sorting by prefix transpositions. As a result, we show that the approximation ratios of three algorithms are tight and conjecture that the approximation ratio of the remaining one is also tight.", acknowledgement = ack-nhfb, articleno = "1.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Grossi:2015:FCT, author = "Roberto Grossi and Giuseppe Ottaviano", title = "Fast Compressed Tries through Path Decompositions", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.8:1--1.8:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2656332", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/datacompression.bib; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Tries are popular data structures for storing a set of strings, where common prefixes are represented by common root-to-node paths. More than 50 years of usage have produced many variants and implementations to overcome some of their limitations. We explore new succinct representations of path-decomposed tries and experimentally evaluate the corresponding reduction in space usage and memory latency, comparing with the state of the art. We study the following applications: compressed string dictionary and monotone minimal perfect hash for strings. In compressed string dictionary, we obtain data structures that outperform other state-of-the-art compressed dictionaries in space efficiency while obtaining predictable query times that are competitive with data structures preferred by the practitioners. On real-world datasets, our compressed tries obtain the smallest space (except for one case) and have the fastest lookup times, whereas access times are within 20\% slower than the best-known solutions. In monotone minimal perfect hash for strings, our compressed tries perform several times faster than other trie-based monotone perfect hash functions while occupying nearly the same space. On real-world datasets, our tries are approximately 2 to 5 times faster than previous solutions, with a space occupancy less than 10\% larger.", acknowledgement = ack-nhfb, articleno = "1.8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Jurkiewicz:2015:MVA, author = "Tomasz Jurkiewicz and Kurt Mehlhorn", title = "On a Model of Virtual Address Translation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "1.9:1--1.9:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2656337", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Modern computers are not Random Access Machines (RAMs). They have a memory hierarchy, multiple cores, and a virtual memory. We address the computational cost of the address translation in the virtual memory. The starting point for our work on virtual memory is the observation that the analysis of some simple algorithms (random scan of an array, binary search, heapsort) in either the RAM model or the External Memory (EM) model does not correctly predict growth rates of actual running times. We propose the Virtual Address Translation (VAT) model to account for the cost of address translations and analyze the algorithms mentioned and others in the model. The predictions agree with the measurements. We also analyze the VAT-cost of cache-oblivious algorithms.", acknowledgement = ack-nhfb, articleno = "1.9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Klasing:2015:E, author = "Ralf Klasing", title = "Editorial", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "2.1:1--2.1:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2677196", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Safro:2015:ACS, author = "Ilya Safro and Peter Sanders and Christian Schulz", title = "Advanced Coarsening Schemes for Graph Partitioning", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "2.2:1--2.2:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2670338", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The graph partitioning problem is widely used and studied in many practical and theoretical applications. Today, multilevel strategies represent one of the most effective and efficient generic frameworks for solving this problem on large-scale graphs. Most of the attention in designing multilevel partitioning frameworks has been on the refinement phase. In this work, we focus on the coarsening phase, which is responsible for creating structures similar to the original but smaller graphs. We compare different matching- and AMG-based coarsening schemes, experiment with the algebraic distance between nodes, and demonstrate computational results on several classes of graphs that emphasize the running time and quality advantages of different coarsening schemes.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Navarro:2015:GDR, author = "Gonzalo Navarro and Simon J. Puglisi and Daniel Valenzuela", title = "General Document Retrieval in Compact Space", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "2.3:1--2.3:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2670128", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/datacompression.bib; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Given a collection of documents and a query pattern, document retrieval is the problem of obtaining documents that are relevant to the query. The collection is available beforehand so that a data structure, called an index, can be built on it to speed up queries. While initially restricted to natural language text collections, document retrieval problems arise nowadays in applications like bioinformatics, multimedia databases, and web mining. This requires a more general setup where text and pattern can be general sequences of symbols, and the classical inverted indexes developed for words cannot be applied. While linear-space time-optimal solutions have been developed for most interesting queries in this general case, space usage is a serious problem in practice. In this article, we develop compact data structures that solve various important document retrieval problems on general text collections. More specifically, we provide practical solutions for listing the documents where a query pattern appears, together with its frequency in each document, and for listing $k$ documents where a query pattern appears most frequently. Some of our techniques build on existing theoretical proposals, while others are new. In particular, we introduce a novel grammar-based compressed bitmap representation that may be of independent interest when dealing with repetitive sequences. Ours are the first practical indexes that use less space when the text collection is compressible. Our experimental results show that, in various real-life text collections, our data structures are significantly smaller than the most space-efficient previous solutions, using up to half the space without noticeably increasing the query time. Overall, document listing can be carried out in 10 to 40 milliseconds for patterns that appear 100 to 10,000 times in the collection, whereas top-$k$ retrieval is carried out in $k$ to $ 10 k$ milliseconds.", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Moruz:2015:EEP, author = "Gabriel Moruz and Andrei Negoescu and Christian Neumann and Volker Weichert", title = "Engineering Efficient Paging Algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "2.4:1--2.4:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2670127", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In the field of online algorithms, paging is a well-studied problem. LRU is a simple paging algorithm that incurs few cache misses and supports efficient implementations. Algorithms outperforming LRU in terms of cache misses exist but are in general more complex and thus not automatically better, since their increased runtime might annihilate the gains in cache misses. In this article, we focus on efficient implementations for the O nOPT class described in Moruz and Negoescu [2012], particularly on an algorithm in this class, denoted RDM, that was shown to typically incur fewer misses than LRU. We provide experimental evidence on a wide range of cache traces showing that our implementation of RDM is competitive to LRU with respect to runtime. In a scenario incurring realistic time penalties for cache misses, we show that our implementation consistently outperforms LRU, even if the runtime of LRU is set to zero.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Kirchler:2015:ECS, author = "Dominik Kirchler and Leo Liberti and Roberto Wolfler Calvo", title = "Efficient Computation of Shortest Paths in Time-Dependent Multi-Modal Networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "2.5:1--2.5:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2670126", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider shortest paths on time-dependent multimodal transportation networks in which restrictions or preferences on the use of certain modes of transportation may arise. We model restrictions and preferences by means of regular languages. Methods for solving the corresponding problem (called the regular language constrained shortest path problem ) already exist. We propose a new algorithm, called State Dependent ALT (SDALT), which runs considerably faster in many scenarios. Speed-up magnitude depends on the type of constraints. We present different versions of SDALT, including unidirectional and bidirectional search. We also provide extensive experimental results on realistic multimodal transportation networks.", acknowledgement = ack-nhfb, articleno = "2.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Manlove:2015:PAK, author = "David F. Manlove and Gregg O'malley", title = "Paired and Altruistic Kidney Donation in the {UK}: Algorithms and Experimentation", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "2.6:1--2.6:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2670129", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study the computational problem of identifying optimal sets of kidney exchanges in the UK. We show how to expand an integer programming-based formulation due to Roth et al. [2007] in order to model the criteria that constitute the UK definition of optimality. The software arising from this work has been used by the National Health Service Blood and Transplant to find optimal sets of kidney exchanges for their National Living Donor Kidney Sharing Schemes since July 2008. We report on the characteristics of the solutions that have been obtained in matching runs of the scheme since this time. We then present empirical results arising from experiments on the real datasets that stem from these matching runs, with the aim of establishing the extent to which the particular optimality criteria that are present in the UK influence the structure of the solutions that are ultimately computed. A key observation is that allowing four-way exchanges would be likely to lead to a moderate number of additional transplants.", acknowledgement = ack-nhfb, articleno = "2.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Luxen:2015:CSA, author = "Dennis Luxen and Dennis Schieferdecker", title = "Candidate Sets for Alternative Routes in Road Networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "2.7:1--2.7:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2674395", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We study the computation of good alternatives to the shortest path in road networks. Our approach is based on single via-node routing on top of contraction hierarchies and achieves superior quality and efficiency compared to previous methods. We present a fast preprocessing method for computing multiple good alternatives and apply this result in an online setting. This setting makes our result applicable in legacy systems with negligible memory overhead. An extensive experimental analysis on a continental-sized real- world road network proves the performance of our algorithm and supports the general systematic algorithm engineering approach. We also show how to combine our results with the competing concept of alternative graphs that encode many alternative paths at once.", acknowledgement = ack-nhfb, articleno = "2.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Bader:2015:ISI, author = "David A. Bader and Petra Mutzel", title = "Introduction to Special Issue {ALENEX'12}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "3.1:1--3.1:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2721893", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "3.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Dibbelt:2015:UCM, author = "Julian Dibbelt and Thomas Pajor and Dorothea Wagner", title = "User-Constrained Multimodal Route Planning", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "19", number = "??", pages = "3.2:1--3.2:??", month = feb, year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2699886", ISSN = "1084-6654", bibdate = "Fri Apr 3 16:22:03 MDT 2015", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In the multimodal route planning problem, we are given multiple transportation networks (e.g., pedestrian, road, public transit) and ask for a best integrated journey between two points. The main challenge is that a seemingly optimal journey may have changes between networks that do not reflect the user's modal preferences. In fact, quickly computing reasonable multimodal routes remains a challenging problem: previous approaches either suffer from poor query performance or their available choices of modal preferences during query time is limited. In this work, we focus on computing exact multimodal journeys that can be restricted by specifying arbitrary modal sequences at query time. For example, a user can say whether he or she wants to only use public transit, prefers to also use a taxi or walking at the beginning or end of the journey, or has no restrictions at all. By carefully adapting node contraction, a common ingredient to many speedup techniques on road networks, we are able to compute point-to-point queries on a continental network combined of cars, railroads, and flights several orders of magnitude faster than Dijkstra's algorithm. Thereby, we require little space overhead and obtain fast preprocessing times.", acknowledgement = ack-nhfb, articleno = "3.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Chimani:2015:UPT, author = "Markus Chimani and Robert Zeranski", title = "Upward Planarity Testing in Practice: {SAT} Formulations and Comparative Study", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "20", number = "??", pages = "1.2:1--1.2:??", year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2699875", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:47:42 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A directed acyclic graph (DAG) is upward planar if it can be drawn without any crossings while all edges-when following them in their direction-are drawn with strictly monotonously increasing y -coordinates. Testing whether a graph allows such a drawing is known to be NP-complete, and while the problem is polynomial-time solvable for special graph classes, there is not much known about solving the problem for general graphs in practice. The only attempt so far has been a branch-and-bound algorithm over the graph's triconnectivity structure, which was able to solve small graphs. Furthermore, there are some known FPT algorithms to deal with the problem. In this article, we propose two fundamentally different approaches based on the seemingly novel concept of ordered embeddings and on the concept of a Hanani--Tutte-type characterization of monotone drawings. In both approaches, we model the problem as special SAT instances, that is, logic formulae for which we check satisfiability. Solving these SAT instances allows us to decide upward planarity for arbitrary graphs. For the first time, we give an extensive experimental comparison between virtually all known approaches to the problem. To this end, we also investigate implementation issues and different variants of the known algorithms as well as of our SAT approaches and evaluate all algorithms on real-world as well as on constructed instances. We also give a detailed performance study of the novel SAT approaches. We show that the SAT formulations outperform all known approaches for graphs with up to 400 edges. For even larger graphs, a modified branch-and-bound algorithm becomes competitive.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Cohen:2015:CGH, author = "Nathann Cohen and David Coudert and Aur{\'e}lien Lancin", title = "On Computing the {Gromov} Hyperbolicity", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "20", number = "??", pages = "1.6:1--1.6:??", year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2780652", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:47:42 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The Gromov hyperbolicity is an important parameter for analyzing complex networks which expresses how the metric structure of a network looks like a tree. It is for instance used to provide bounds on the expected stretch of greedy-routing algorithms in Internet-like graphs. However, the best-known theoretical algorithm computing this parameter runs in O ( n$^{3.69}$ ) time, which is prohibitive for large-scale graphs. In this article, we propose an algorithm for determining the hyperbolicity of graphs with tens of thousands of nodes. Its running time depends on the distribution of distances and on the actual value of the hyperbolicity. Although its worst case runtime is O ( n$^4$ ), it is in practice much faster than previous proposals as observed in our experimentations. Finally, we propose a heuristic algorithm that can be used on graphs with millions of nodes. Our algorithms are all evaluated on benchmark instances.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{DAndrea:2015:DMS, author = "Annalisa D'Andrea and Mattia D'Emidio and Daniele Frigioni and Stefano Leucci and Guido Proietti", title = "Dynamic Maintenance of a Shortest-Path Tree on Homogeneous Batches of Updates: New Algorithms and Experiments", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "20", number = "??", pages = "1.5:1--1.5:??", year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2786022", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:47:42 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A dynamic graph algorithm is called batch if it is able to update efficiently the solution of a given graph problem after multiple updates at a time (i.e., a batch) take place on the input graph. In this article, we study batch algorithms for maintaining a single-source shortest-path tree in graphs with positive real edge weights. In particular, we focus our attention on homogeneous batches, that is, either incremental (containing only edge insertion and weight decrease operations) or decremental (containing only edge deletion and weight increase operations) batches, which model realistic dynamic scenarios like transient vertex failures in communication networks and traffic congestion/decongestion phenomena in road networks. We propose two new algorithms to process either incremental or decremental batches, respectively, and a combination of these two algorithms that is able to process arbitrary sequences of incremental and decremental batches. All these algorithms are update sensitive; namely, they are efficient with respect to the number of vertices in the shortest-path tree that change their parents and/or their distances from the source as a consequence of a batch. This makes unfeasible an effective comparison on a theoretical basis of our new algorithms with the solutions known in the literature, which in turn are analyzed with respect to others and different parameters. For this reason, in order to evaluate the quality of our approach, we provide also an extensive experimental study including our new algorithms and the most efficient previous batch algorithms. Our experimental results complement previous studies and show that the various solutions can be consistently ranked on the basis of the type of homogeneous batch and of the underlying network. As a result, our work can be helpful in selecting a proper solution depending on the specific application scenario.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Finocchi:2015:CCM, author = "Irene Finocchi and Marco Finocchi and Emanuele G. Fusco", title = "Clique Counting in {MapReduce}: Algorithms and Experiments", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "20", number = "??", pages = "1.7:1--1.7:??", year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2794080", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:47:42 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We tackle the problem of counting the number q$_k$ of k -cliques in large-scale graphs, for any constant k {$>$}= 3. Clique counting is essential in a variety of applications, including social network analysis. Our algorithms make it possible to compute q$_k$ for several real-world graphs and shed light on its growth rate as a function of k. Even for small values of k, the number q$_k$ of k -cliques can be in the order of tens or hundreds of trillions. As k increases, different graph instances show different behaviors: while on some graphs $q_{k + 1} < q_k$, on other benchmarks $q_{k + 1} \fg q_k$, up to two orders of magnitude in our observations. Graphs with steep clique growth rates represent particularly tough instances in practice. Due to the computationally intensive nature of the clique counting problem, we settle for parallel solutions in the MapReduce framework, which has become in the last few years a de facto standard for batch processing of massive datasets. We give both theoretical and experimental contributions. On the theory side, we design the first exact scalable algorithm for counting (and listing) $k$-cliques in MapReduce. Our algorithm uses $O (m^{3 / 2})$ total space and $O(m^{k / 2})$ work, where $m$ is the number of graph edges. This matches the best-known bounds for triangle listing when $k = 3$ and is work optimal in the worst case for any $k$, while keeping the communication cost independent of $k$. We also design sampling-based estimators that can dramatically reduce the running time and space requirements of the exact approach, while providing very accurate solutions with high probability. We then assess the effectiveness of different clique counting approaches through an extensive experimental analysis over the Amazon EC2 platform, considering both our algorithms and their state-of-the-art competitors. The experimental results clearly highlight the algorithm of choice in different scenarios and prove our exact approach to be the most effective when the number of $k$-cliques is large, gracefully scaling to nontrivial values of $k$ even on clusters of small/medium size. Our approximation algorithms achieve extremely accurate estimates and large speedups, especially on the toughest instances for the exact algorithms.", acknowledgement = ack-nhfb, articleno = "1.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Hedtke:2015:UST, author = "Ivo Hedtke", title = "Upgrading Subgroup Triple-Product-Property Triples", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "20", number = "??", pages = "1.1:1--1.1:??", year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2699877", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:47:42 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "In 2003, Cohn and Umans introduced a group-theoretic approach to fast matrix multiplication. This involves finding large subsets of a group satisfying the Triple Product Property (TPP) as a means to bound the exponent of matrix multiplication. Recently, Hedtke and Murthy discussed several methods to find TPP triples. Because the search space for subset triples is too large, it is only possible to focus on subgroup triples. We present methods to upgrade a given TPP triple to a bigger TPP triple. If no upgrade is possible, we use reduction methods (based on random experiments and heuristics) to create a smaller TPP triple that can be used as input for the upgrade methods. If we apply the upgrade process for subset triples after one step with the upgrade method for subgroup triples for the known maximal subgroup TPP triples in groups of order up to 1,000, we achieve an enlargement of the triple size of 100\% in the best case. Further, we test the upgrade process with all examples from the 2003 and 2005 papers from Cohn et al. and are able to increase the triple size by 595\% in the best case (in the group D$^5_6$ ).", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Spence:2015:WCC, author = "Ivor Spence", title = "Weakening Cardinality Constraints Creates Harder Satisfiability Benchmarks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "20", number = "??", pages = "1.4:1--1.4:??", year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2746239", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:47:42 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "For some time, the satisfiability formulae that have been the most difficult to solve for their size have been crafted to be unsatisfiable by the use of cardinality constraints. Recent solvers have introduced explicit checking of such constraints, rendering previously difficult formulae trivial to solve. A family of unsatisfiable formulae is described that is derived from the sgen4 family but cannot be solved using cardinality constraints detection and reasoning alone. These formulae were found to be the most difficult during the SAT2014 competition by a significant margin and include the shortest unsolved benchmark in the competition, sgen6-1200-5-1.cnf.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Ullmann:2015:DRL, author = "Julian R. Ullmann", title = "Degree Reduction in Labeled Graph Retrieval", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "20", number = "??", pages = "1.3:1--1.3:??", year = "2015", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2699878", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:47:42 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Within a given collection of graphs, a graph retrieval system may seek all graphs that contain a given graph, or may instead seek all graphs that are contained within a given graph. Although subgraph isomorphism is worst-case exponential, it may be average-case polynomial if graphs are labeled so as to restrict possible correspondences between vertices of included and includer graphs. Degree reduction is a procedure that uses logical inference to preclude some such correspondences, thereby substantially increasing the size of includer graphs that can be processed, without preventing any existent isomorphism from being found. Degree reduction works only with labeled graphs, which may be directed or undirected, with or without edge labels. Inexact or approximate isomorphism is accommodated by reducing strictness of conditions for perfect isomorphism. Disk-based degree reduction, which is an order of magnitude slower than memory-based degree reduction, has successfully processed graphs that have millions of vertices. Although the principle of degree reduction is simple and fundamental, its efficient practical implementation involves intricate procedural detail. Its average-case complexity analysis is currently intractable, so cost-benefit assessment has to be experimental.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Bergner:2016:BPC, author = "Martin Bergner and Marco E. L{\"u}bbecke and Jonas T. Witt", title = "A Branch-Price-and-Cut Algorithm for Packing Cuts in Undirected Graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.2:1--1.2:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2851492", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The cut packing problem in an undirected graph is to find a largest cardinality collection of pairwise edge-disjoint cuts. We provide the first experimental study of this NP-hard problem that is interesting from a pure theorist's viewpoint as well as from the standpoint of scientific applications (e.g., in bioinformatics and network reliability). So far it could not be solved exactly. We propose a branch-price-and-cut algorithm to optimally solve instances from various graph classes, random and from the literature, with up to several hundred vertices. In particular, we investigate how complexity results match computational experience and how combinatorial properties help improve the algorithm's performance.", acknowledgement = ack-nhfb, articleno = "1.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Bingmann:2016:ISL, author = "Timo Bingmann and Johannes Fischer and Vitaly Osipov", title = "Inducing Suffix and {LCP} Arrays in External Memory", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "2.3:1--2.3:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2975593", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider full text index construction in external memory (EM). Our first contribution is an inducing algorithm for suffix arrays in external memory, which runs in sorting complexity. Practical tests show that this algorithm outperforms the previous best EM suffix sorter [Dementiev et al., JEA 2008] by a factor of about two in time and I/O volume. Our second contribution is to augment the first algorithm to also construct the array of longest common prefixes (LCPs). This yields a new internal memory LCP array construction algorithm and the first EM construction algorithm for LCP arrays. The overhead in time and I/O volume for this extended algorithm over plain suffix array construction is roughly two. Our algorithms scale far beyond problem sizes previously considered in the literature (text size of 80GiB using only 4GiB of RAM in our experiments).", acknowledgement = ack-nhfb, articleno = "2.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Coudert:2016:EEB, author = "David Coudert and Dorian Mazauric and Nicolas Nisse", title = "Experimental Evaluation of a Branch-and-Bound Algorithm for Computing Pathwidth and Directed Pathwidth", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.3:1--1.3:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2851494", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Path decompositions of graphs are an important ingredient of dynamic programming algorithms for solving efficiently many NP-hard problems. Therefore, computing the pathwidth and associated path decomposition of graphs has both a theoretical and practical interest. In this article, we design a branch-and-bound algorithm that computes the exact pathwidth of graphs and a corresponding path decomposition. Our main contribution consists of several nontrivial techniques to reduce the size of the input graph (preprocessing) and to cut the exploration space during the search phase of the algorithm. We evaluate experimentally our algorithm by comparing it to existing algorithms of the literature. It appears from the simulations that our algorithm offers a significant gain with respect to previous work. In particular, it is able to compute the exact pathwidth of any graph with less than 60 nodes in a reasonable running time ($\leq$ 10min on a standard laptop). Moreover, our algorithm achieves good performance when used as a heuristic (i.e., when returning best result found within bounded time limit). Our algorithm is not restricted to undirected graphs since it actually computes the directed pathwidth that generalizes the notion of pathwidth to digraphs.", acknowledgement = ack-nhfb, articleno = "1.3", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Dibbelt:2016:CCH, author = "Julian Dibbelt and Ben Strasser and Dorothea Wagner", title = "Customizable Contraction Hierarchies", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.5:1--1.5:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2886843", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the problem of quickly computing shortest paths in weighted graphs. Often, this is achieved in two phases: (1) derive auxiliary data in an expensive preprocessing phase, and (2) use this auxiliary data to speed up the query phase. By adding a fast weight-customization phase, we extend Contraction Hierarchies to support a three-phase workflow. The expensive preprocessing is split into a phase exploiting solely the unweighted topology of the graph and a lightweight phase that adapts the auxiliary data to a specific weight. We achieve this by basing our Customizable Contraction Hierarchies (CCHs) on nested dissection orders. We provide an in-depth experimental analysis on large road and game maps showing that CCHs are a very practicable solution in scenarios where edge weights often change.", acknowledgement = ack-nhfb, articleno = "1.5", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Efentakis:2016:REH, author = "Alexandros Efentakis and Dieter Pfoser", title = "{ReHub}: Extending Hub Labels for Reverse $k$-Nearest Neighbor Queries on Large-Scale Networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.13:1--1.13:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2990192", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Quite recently, the algorithmic community has focused on solving multiple shortest-path query problems beyond simple vertex-to-vertex queries, especially in the context of road networks. Unfortunately, those advanced query-processing techniques cannot be applied to large-scale graphs, such as social or collaboration networks, or to efficiently answer reverse k -nearest neighbor (R k NN) queries, which are of practical relevance to a wide range of applications. To remedy this, we propose ReHub, a novel main-memory algorithm that extends the hub labeling technique to efficiently answer R k NN queries on large-scale networks. Our experimentation will show that ReHub is the best overall solution for this type of queries, requiring only minimal additional preprocessing and providing very fast query times in all cases.", acknowledgement = ack-nhfb, articleno = "1.13", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Fox-Epstein:2016:SSC, author = "Eli Fox-Epstein and Shay Mozes and Phitchaya Mangpo Phothilimthana and Christian Sommer", title = "Short and Simple Cycle Separators in Planar Graphs", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "2.2:1--2.2:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2957318", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We provide an implementation of an algorithm that, given a triangulated planar graph with m edges, returns a simple cycle that is a 3/4-balanced separator consisting of at most $\sqrt{8 m}$ edges. An efficient construction of a short and balanced separator that forms a simple cycle is essential in numerous planar graph algorithms, for example, for computing shortest paths, minimum cuts, or maximum flows. To the best of our knowledge, this is the first implementation of such a cycle separator algorithm with a worst-case guarantee on the cycle length. We evaluate the performance of our algorithm and compare it to the planar separator algorithms recently studied by Holzer et al. [2009]. Out of these algorithms, only the Fundamental Cycle Separator (FCS) produces a simple cycle separator. However, FCS does not provide a worst-case size guarantee. We demonstrate that (1) our algorithm is competitive across all test cases in terms of running time, balance, and cycle length; (2) it provides worst-case guarantees on the cycle length, significantly outperforming FCS on some instances; and (3) it scales to large graphs.", acknowledgement = ack-nhfb, articleno = "2.2", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gemsa:2016:ELS, author = "Andreas Gemsa and Martin N{\"o}llenburg and Ignaz Rutter", title = "Evaluation of Labeling Strategies for Rotating Maps", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.4:1--1.4:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2851493", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "We consider the following problem of labeling points in a dynamic map that allows rotation. We are given a set of feature points in the plane labeled by a set of mutually disjoint labels, where each label is an axis-aligned rectangle attached with one corner to its respective point. We require that each label remains horizontally aligned during the map rotation, and our goal is to find a set of mutually nonoverlapping visible labels for every rotation angle $\alpha \in [0, 2 \pi)$ so that the number of visible labels over a full map rotation of $2 \pi$ is maximized. We discuss and experimentally evaluate several labeling strategies that define additional consistency constraints on label visibility to reduce flickering effects during monotone map rotation. We introduce three heuristic algorithms and compare them experimentally to an existing approximation algorithm and exact solutions obtained from an integer linear program. Our results show that on the one hand, low flickering can be achieved at the expense of only a small reduction in the objective value, and on the other hand, the proposed heuristics achieve a high labeling quality significantly faster than the other methods.", acknowledgement = ack-nhfb, articleno = "1.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Glantz:2016:TBC, author = "Roland Glantz and Henning Meyerhenke and Christian Schulz", title = "Tree-Based Coarsening and Partitioning of Complex Networks", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.6:1--1.6:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2851496", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "A hierarchy of increasingly coarse versions of a network allows one to represent the network on multiple scales at the same time. Often, the elementary operation for generating a hierarchy on a network is merging adjacent vertices, an operation that can be realized through contracting the edge between the two vertices. Such a hierarchy is defined by the selection of the edges to be contracted between a level and the next coarser level. The selection may involve (i) rating the edges, (ii) constraining the selection (e.g., that the selected edges form a matching), as well as (iii) maximizing the total rate of the selected edges under the constraints. Hierarchies of this kind are, among others, involved in multilevel methods for partitioning networks-a prerequisite for processing in parallel with distributed memory. In this article, we propose a new edge rating by (i) defining weights for the edges of a network that express the edges' importance for connectivity via shortest paths, (ii) computing a minimum weight spanning tree with respect to these weights, and (iii) rating the network edges based on the conductance values of the tree's fundamental cuts. To make the computation of our new edge rating efficient, we develop the first optimal linear-time algorithm to compute the conductance values of all fundamental cuts of a given spanning tree. We integrate the new edge rating into a leading multilevel graph partitioner and equip the latter also with a new greedy postprocessing for optimizing the Maximum Communication Volume (MCV) of a partition. Our experiments, in which we bipartition frequently used benchmark networks, show that the postprocessing reduces MCV by 11.3\%. Our new edge rating, here used for matching-based coarsening, further reduces MCV by 10.3\% compared to the previously best rating with MCV postprocessing in place for both ratings. In total, with a modest increase in running time, our new approach reduces the MCV of complex network partitions by 20.4\%.", acknowledgement = ack-nhfb, articleno = "1.6", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Gudmundsson:2016:ESS, author = "Joachim Gudmundsson and Jyrki Katajainen", title = "Editorial, {SEA 2014} Special Issue", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.1:1--1.1:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2854021", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "1.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Karkkainen:2016:LAC, author = "Juha K{\"a}rkk{\"a}inen and Dominik Kempa", title = "{LCP} Array Construction in External Memory", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.7:1--1.7:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2851491", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "One of the most important data structures for string processing-the suffix array-needs to be augmented with the longest-common-prefix (LCP) array in numerous applications. We describe the first external memory algorithm for constructing the LCP array given the suffix array as input. The only previous way to compute the LCP array for data that is bigger than the RAM is to use an external memory suffix array construction algorithm (SACA) with complex modifications to produce the LCP array as a by-product. Compared to the best prior method, our algorithm needs much less disk space (by more than a factor of three) and is significantly faster. Furthermore, our algorithm can be combined with any SACA, including a better one developed in the future.", acknowledgement = ack-nhfb, articleno = "1.7", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Karkkainen:2016:LLZ, author = "Juha K{\"a}rkk{\"a}inen and Dominik Kempa and Simon J. Puglisi", title = "Lazy {Lempel--Ziv} Factorization Algorithms", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "2.4:1--2.4:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2699876", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/datacompression.bib; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "For decades the Lempel--Ziv (LZ77) factorization has been a cornerstone of data compression and string processing algorithms, and uses for it are still being uncovered. For example, LZ77 is central to several recent text indexing data structures designed to search highly repetitive collections. However, in many applications computation of the factorization remains a bottleneck in practice. In this article, we describe a number of simple and fast LZ77 factorization algorithms, which consistently outperform all previous methods in practice, use less memory, and still offer strong worst-case performance guarantees. A common feature of the new algorithms is that they compute longest common prefix information in a lazy fashion, with the degree of laziness in preprocessing characterizing different algorithms.", acknowledgement = ack-nhfb, articleno = "2.4", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Marinov:2016:PAF, author = "Martin Marinov and Nicholas Nash and David Gregg", title = "Practical Algorithms for Finding Extremal Sets", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.9:1--1.9:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2893184", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "The minimal sets within a collection of sets are defined as the ones that do not have a proper subset within the collection, and the maximal sets are the ones that do not have a proper superset within the collection. Identifying extremal sets is a fundamental problem with a wide range of applications in SAT solvers, data mining, and social network analysis. In this article, we present two novel improvements of the high-quality extremal set identification algorithm, AMS-Lex, described by Bayardo and Panda. The first technique uses memoization to improve the execution time of the single-threaded variant of the AMS-Lex, while our second improvement uses parallel programming methods. In a subset of the presented experiments, our memoized algorithm executes more than 400 times faster than the highly efficient publicly available implementation of AMS-Lex. Moreover, we show that our modified algorithm's speedup is not bounded above by a constant and that it increases as the length of the common prefixes in successive input itemsets increases. We provide experimental results using both real-world and synthetic datasets, and show our multithreaded variant algorithm outperforming AMS-Lex by 3 to 6 times. We find that on synthetic input datasets, when executed using 16 CPU cores of a 32-core machine, our multithreaded program executes about as fast as the state-of-the-art parallel GPU-based program using an NVIDIA GTX 580 graphics processing unit.", acknowledgement = ack-nhfb, articleno = "1.9", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Mcgeoch:2016:MDJ, author = "Catherine Mcgeoch", title = "In Memoriam: {David S. Johnson}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.1:1--1.1:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2907073", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "1.1e", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Navarro:2016:FCS, author = "Gonzalo Navarro and Alberto Ord{\'o}{\~n}ez Pereira", title = "Faster Compressed Suffix Trees for Repetitive Collections", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.8:1--1.8:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2851495", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Recent compressed suffix trees targeted to highly repetitive sequence collections reach excellent compression performance, but operation times are very high. We design a new suffix tree representation for this scenario that still achieves very low space usage, only slightly larger than the best previous one, but supports the operations orders of magnitude faster. Our suffix tree is still orders of magnitude slower than general-purpose compressed suffix trees, but these use several times more space when the collection is repetitive. Our main novelty is a practical grammar-compressed tree representation with full navigation functionality, which is useful in all applications where large trees with repetitive topology must be represented.", acknowledgement = ack-nhfb, articleno = "1.8", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Peethambaran:2016:ESR, author = "Jiju Peethambaran and Amal Dev Parakkat and Ramanathan Muthuganapathy", title = "An Empirical Study on Randomized Optimal Area Polygonization of Planar Point Sets", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.10:1--1.10:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2896849", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "While random polygon generation from a set of planar points has been widely investigated in the literature, very few works address the construction of a simple polygon with minimum area (MINAP) or maximum area (MAXAP) from a set of fixed planar points. Currently, no deterministic algorithms are available to compute MINAP/MAXAP, as the problems have been shown to be NP-complete. In this article, we present a greedy heuristic for computing the approximate MINAP of any given planar point set using the technique of randomized incremental construction. For a given point set of n points, the proposed algorithm takes O ( n$^2$ log n ) time and O ( n ) space. It is rather simplistic in nature, hence very easy for implementation and maintenance. We report on various experimental studies on the behavior of a randomized heuristic on different point set instances. Test data have been taken from the SPAETH cluster data base and TSPLIB library. Experimental results indicate that the proposed algorithm outperforms its counterparts for generating a tighter upper bound on the optimal minimum area polygon for large-sized point sets.", acknowledgement = ack-nhfb, articleno = "1.10", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Rosenbrock:2016:NAP, author = "Conrad W. Rosenbrock and Wiley S. Morgan and Gus L. W. Hart and Stefano Curtarolo and Rodney W. Forcade", title = "Numerical Algorithm for {P{\'o}lya} Enumeration Theorem", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.11:1--1.11:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2955094", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "Although the P{\'o}lya enumeration theorem has been used extensively for decades, an optimized, purely numerical algorithm for calculating its coefficients is not readily available. We present such an algorithm for finding the number of unique colorings of a finite set under the action of a finite group.", acknowledgement = ack-nhfb, articleno = "1.11", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Sanders:2016:ISI, author = "Peter Sanders and Norbert Zeh", title = "Introduction to Special Issue {ALENEX 2013}", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "2.1:1--2.1:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2966922", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", acknowledgement = ack-nhfb, articleno = "2.1", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", } @Article{Toda:2016:IEA, author = "Takahisa Toda and Takehide Soh", title = "Implementing Efficient All Solutions {SAT} Solvers", journal = j-ACM-J-EXP-ALGORITHMICS, volume = "21", number = "1", pages = "1.12:1--1.12:??", month = nov, year = "2016", CODEN = "????", DOI = "http://dx.doi.org/10.1145/2975585", ISSN = "1084-6654", bibdate = "Fri Nov 4 16:46:55 MDT 2016", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/jea.bib", abstract = "All solutions SAT (AllSAT for short) is a variant of the propositional satisfiability problem. AllSAT has been relatively unexplored compared to other variants despite its significance. We thus survey and discuss major techniques of AllSAT solvers. We accurately implemented them and conducted comprehensive experiments using a large number of instances and various types of solvers including a few publicly available software. The experiments revealed the solvers' characteristics. We made our implemented solvers publicly available so that other researchers can easily develop their solvers by modifying our code and comparing it with existing methods.", acknowledgement = ack-nhfb, articleno = "1.12", fjournal = "Journal of Experimental Algorithmics (JEA)", journal-URL = "http://portal.acm.org/browse_dl.cfm?idx=J430", }