Publications

2018
[1]
A. Abboud, A. Backurs, K. Bringmann, and M. Künnemann, “Fine-Grained Complexity of Analyzing Compressed Data: Quantifying Improvements over Decompress-And-Solve,” 2018. [Online]. Available: http://arxiv.org/abs/1803.00796. (arXiv: 1803.00796)
Abstract
Can we analyze data without decompressing it? As our data keeps growing, understanding the time complexity of problems on compressed inputs, rather than in convenient uncompressed forms, becomes more and more relevant. Suppose we are given a compression of size $n$ of data that originally has size $N$, and we want to solve a problem with time complexity $T(\cdot)$. The naive strategy of "decompress-and-solve" gives time $T(N)$, whereas "the gold standard" is time $T(n)$: to analyze the compression as efficiently as if the original data was small. We restrict our attention to data in the form of a string (text, files, genomes, etc.) and study the most ubiquitous tasks. While the challenge might seem to depend heavily on the specific compression scheme, most methods of practical relevance (Lempel-Ziv-family, dictionary methods, and others) can be unified under the elegant notion of Grammar Compressions. A vast literature, across many disciplines, established this as an influential notion for Algorithm design. We introduce a framework for proving (conditional) lower bounds in this field, allowing us to assess whether decompress-and-solve can be improved, and by how much. Our main results are: - The $O(nN\sqrt{\log{N/n}})$ bound for LCS and the $O(\min\{N \log N, nM\})$ bound for Pattern Matching with Wildcards are optimal up to $N^{o(1)}$ factors, under the Strong Exponential Time Hypothesis. (Here, $M$ denotes the uncompressed length of the compressed pattern.) - Decompress-and-solve is essentially optimal for Context-Free Grammar Parsing and RNA Folding, under the $k$-Clique conjecture. - We give an algorithm showing that decompress-and-solve is not optimal for Disjointness.
Export
BibTeX
@online{Abboud_arXiv1803.00796, TITLE = {Fine-Grained Complexity of Analyzing Compressed Data: Quantifying Improvements over Decompress-And-Solve}, AUTHOR = {Abboud, Amir and Backurs, Arturs and Bringmann, Karl and K{\"u}nnemann, Marvin}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1803.00796}, EPRINT = {1803.00796}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {Can we analyze data without decompressing it? As our data keeps growing, understanding the time complexity of problems on compressed inputs, rather than in convenient uncompressed forms, becomes more and more relevant. Suppose we are given a compression of size $n$ of data that originally has size $N$, and we want to solve a problem with time complexity $T(\cdot)$. The naive strategy of "decompress-and-solve" gives time $T(N)$, whereas "the gold standard" is time $T(n)$: to analyze the compression as efficiently as if the original data was small. We restrict our attention to data in the form of a string (text, files, genomes, etc.) and study the most ubiquitous tasks. While the challenge might seem to depend heavily on the specific compression scheme, most methods of practical relevance (Lempel-Ziv-family, dictionary methods, and others) can be unified under the elegant notion of Grammar Compressions. A vast literature, across many disciplines, established this as an influential notion for Algorithm design. We introduce a framework for proving (conditional) lower bounds in this field, allowing us to assess whether decompress-and-solve can be improved, and by how much. Our main results are: -- The $O(nN\sqrt{\log{N/n}})$ bound for LCS and the $O(\min\{N \log N, nM\})$ bound for Pattern Matching with Wildcards are optimal up to $N^{o(1)}$ factors, under the Strong Exponential Time Hypothesis. (Here, $M$ denotes the uncompressed length of the compressed pattern.) -- Decompress-and-solve is essentially optimal for Context-Free Grammar Parsing and RNA Folding, under the $k$-Clique conjecture. -- We give an algorithm showing that decompress-and-solve is not optimal for Disjointness.}, }
Endnote
%0 Report %A Abboud, Amir %A Backurs, Arturs %A Bringmann, Karl %A Künnemann, Marvin %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fine-Grained Complexity of Analyzing Compressed Data: Quantifying Improvements over Decompress-And-Solve : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3E38-C %U http://arxiv.org/abs/1803.00796 %D 2018 %X Can we analyze data without decompressing it? As our data keeps growing, understanding the time complexity of problems on compressed inputs, rather than in convenient uncompressed forms, becomes more and more relevant. Suppose we are given a compression of size $n$ of data that originally has size $N$, and we want to solve a problem with time complexity $T(\cdot)$. The naive strategy of "decompress-and-solve" gives time $T(N)$, whereas "the gold standard" is time $T(n)$: to analyze the compression as efficiently as if the original data was small. We restrict our attention to data in the form of a string (text, files, genomes, etc.) and study the most ubiquitous tasks. While the challenge might seem to depend heavily on the specific compression scheme, most methods of practical relevance (Lempel-Ziv-family, dictionary methods, and others) can be unified under the elegant notion of Grammar Compressions. A vast literature, across many disciplines, established this as an influential notion for Algorithm design. We introduce a framework for proving (conditional) lower bounds in this field, allowing us to assess whether decompress-and-solve can be improved, and by how much. Our main results are: - The $O(nN\sqrt{\log{N/n}})$ bound for LCS and the $O(\min\{N \log N, nM\})$ bound for Pattern Matching with Wildcards are optimal up to $N^{o(1)}$ factors, under the Strong Exponential Time Hypothesis. (Here, $M$ denotes the uncompressed length of the compressed pattern.) - Decompress-and-solve is essentially optimal for Context-Free Grammar Parsing and RNA Folding, under the $k$-Clique conjecture. - We give an algorithm showing that decompress-and-solve is not optimal for Disjointness. %K Computer Science, Computational Complexity, cs.CC,Computer Science, Data Structures and Algorithms, cs.DS
[2]
A. Abboud and K. Bringmann, “Tighter Connections Between Formula-SAT and Shaving Logs,” 2018. [Online]. Available: http://arxiv.org/abs/1804.08978. (arXiv: 1804.08978)
Abstract
A noticeable fraction of Algorithms papers in the last few decades improve the running time of well-known algorithms for fundamental problems by logarithmic factors. For example, the $O(n^2)$ dynamic programming solution to the Longest Common Subsequence problem (LCS) was improved to $O(n^2/\log^2 n)$ in several ways and using a variety of ingenious tricks. This line of research, also known as "the art of shaving log factors", lacks a tool for proving negative results. Specifically, how can we show that it is unlikely that LCS can be solved in time $O(n^2/\log^3 n)$? Perhaps the only approach for such results was suggested in a recent paper of Abboud, Hansen, Vassilevska W. and Williams (STOC'16). The authors blame the hardness of shaving logs on the hardness of solving satisfiability on Boolean formulas (Formula-SAT) faster than exhaustive search. They show that an $O(n^2/\log^{1000} n)$ algorithm for LCS would imply a major advance in circuit lower bounds. Whether this approach can lead to tighter barriers was unclear. In this paper, we push this approach to its limit and, in particular, prove that a well-known barrier from complexity theory stands in the way for shaving five additional log factors for fundamental combinatorial problems. For LCS, regular expression pattern matching, as well as the Fr\'echet distance problem from Computational Geometry, we show that an $O(n^2/\log^{7+\varepsilon} n)$ runtime would imply new Formula-SAT algorithms. Our main result is a reduction from SAT on formulas of size $s$ over $n$ variables to LCS on sequences of length $N=2^{n/2} \cdot s^{1+o(1)}$. Our reduction is essentially as efficient as possible, and it greatly improves the previously known reduction for LCS with $N=2^{n/2} \cdot s^c$, for some $c \geq 100$.
Export
BibTeX
@online{Abboud_arXiv1804.08978, TITLE = {Tighter Connections Between Formula-{SAT} and Shaving Logs}, AUTHOR = {Abboud, Amir and Bringmann, Karl}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1804.08978}, EPRINT = {1804.08978}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {A noticeable fraction of Algorithms papers in the last few decades improve the running time of well-known algorithms for fundamental problems by logarithmic factors. For example, the $O(n^2)$ dynamic programming solution to the Longest Common Subsequence problem (LCS) was improved to $O(n^2/\log^2 n)$ in several ways and using a variety of ingenious tricks. This line of research, also known as "the art of shaving log factors", lacks a tool for proving negative results. Specifically, how can we show that it is unlikely that LCS can be solved in time $O(n^2/\log^3 n)$? Perhaps the only approach for such results was suggested in a recent paper of Abboud, Hansen, Vassilevska W. and Williams (STOC'16). The authors blame the hardness of shaving logs on the hardness of solving satisfiability on Boolean formulas (Formula-SAT) faster than exhaustive search. They show that an $O(n^2/\log^{1000} n)$ algorithm for LCS would imply a major advance in circuit lower bounds. Whether this approach can lead to tighter barriers was unclear. In this paper, we push this approach to its limit and, in particular, prove that a well-known barrier from complexity theory stands in the way for shaving five additional log factors for fundamental combinatorial problems. For LCS, regular expression pattern matching, as well as the Fr\'echet distance problem from Computational Geometry, we show that an $O(n^2/\log^{7+\varepsilon} n)$ runtime would imply new Formula-SAT algorithms. Our main result is a reduction from SAT on formulas of size $s$ over $n$ variables to LCS on sequences of length $N=2^{n/2} \cdot s^{1+o(1)}$. Our reduction is essentially as efficient as possible, and it greatly improves the previously known reduction for LCS with $N=2^{n/2} \cdot s^c$, for some $c \geq 100$.}, }
Endnote
%0 Report %A Abboud, Amir %A Bringmann, Karl %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Tighter Connections Between Formula-SAT and Shaving Logs : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3DF7-5 %U http://arxiv.org/abs/1804.08978 %D 2018 %X A noticeable fraction of Algorithms papers in the last few decades improve the running time of well-known algorithms for fundamental problems by logarithmic factors. For example, the $O(n^2)$ dynamic programming solution to the Longest Common Subsequence problem (LCS) was improved to $O(n^2/\log^2 n)$ in several ways and using a variety of ingenious tricks. This line of research, also known as "the art of shaving log factors", lacks a tool for proving negative results. Specifically, how can we show that it is unlikely that LCS can be solved in time $O(n^2/\log^3 n)$? Perhaps the only approach for such results was suggested in a recent paper of Abboud, Hansen, Vassilevska W. and Williams (STOC'16). The authors blame the hardness of shaving logs on the hardness of solving satisfiability on Boolean formulas (Formula-SAT) faster than exhaustive search. They show that an $O(n^2/\log^{1000} n)$ algorithm for LCS would imply a major advance in circuit lower bounds. Whether this approach can lead to tighter barriers was unclear. In this paper, we push this approach to its limit and, in particular, prove that a well-known barrier from complexity theory stands in the way for shaving five additional log factors for fundamental combinatorial problems. For LCS, regular expression pattern matching, as well as the Fr\'echet distance problem from Computational Geometry, we show that an $O(n^2/\log^{7+\varepsilon} n)$ runtime would imply new Formula-SAT algorithms. Our main result is a reduction from SAT on formulas of size $s$ over $n$ variables to LCS on sequences of length $N=2^{n/2} \cdot s^{1+o(1)}$. Our reduction is essentially as efficient as possible, and it greatly improves the previously known reduction for LCS with $N=2^{n/2} \cdot s^c$, for some $c \geq 100$. %K Computer Science, Computational Complexity, cs.CC,Computer Science, Data Structures and Algorithms, cs.DS
[3]
M. Abrahamsen, A. Adamaszek, K. Bringmann, V. Cohen-Addad, M. Mehr, E. Rotenberg, A. Roytman, and M. Thorup, “Fast Fencing,” 2018. [Online]. Available: http://arxiv.org/abs/1804.00101. (arXiv: 1804.00101)
Abstract
We consider very natural "fence enclosure" problems studied by Capoyleas, Rote, and Woeginger and Arkin, Khuller, and Mitchell in the early 90s. Given a set $S$ of $n$ points in the plane, we aim at finding a set of closed curves such that (1) each point is enclosed by a curve and (2) the total length of the curves is minimized. We consider two main variants. In the first variant, we pay a unit cost per curve in addition to the total length of the curves. An equivalent formulation of this version is that we have to enclose $n$ unit disks, paying only the total length of the enclosing curves. In the other variant, we are allowed to use at most $k$ closed curves and pay no cost per curve. For the variant with at most $k$ closed curves, we present an algorithm that is polynomial in both $n$ and $k$. For the variant with unit cost per curve, or unit disks, we present a near-linear time algorithm. Capoyleas, Rote, and Woeginger solved the problem with at most $k$ curves in $n^{O(k)}$ time. Arkin, Khuller, and Mitchell used this to solve the unit cost per curve version in exponential time. At the time, they conjectured that the problem with $k$ curves is NP-hard for general $k$. Our polynomial time algorithm refutes this unless P equals NP.
Export
BibTeX
@online{Abrahamsen_arXiv1804.00101, TITLE = {Fast Fencing}, AUTHOR = {Abrahamsen, Mikkel and Adamaszek, Anna and Bringmann, Karl and Cohen-Addad, Vincent and Mehr, Mehran and Rotenberg, Eva and Roytman, Alan and Thorup, Mikkel}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1804.00101}, EPRINT = {1804.00101}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {We consider very natural "fence enclosure" problems studied by Capoyleas, Rote, and Woeginger and Arkin, Khuller, and Mitchell in the early 90s. Given a set $S$ of $n$ points in the plane, we aim at finding a set of closed curves such that (1) each point is enclosed by a curve and (2) the total length of the curves is minimized. We consider two main variants. In the first variant, we pay a unit cost per curve in addition to the total length of the curves. An equivalent formulation of this version is that we have to enclose $n$ unit disks, paying only the total length of the enclosing curves. In the other variant, we are allowed to use at most $k$ closed curves and pay no cost per curve. For the variant with at most $k$ closed curves, we present an algorithm that is polynomial in both $n$ and $k$. For the variant with unit cost per curve, or unit disks, we present a near-linear time algorithm. Capoyleas, Rote, and Woeginger solved the problem with at most $k$ curves in $n^{O(k)}$ time. Arkin, Khuller, and Mitchell used this to solve the unit cost per curve version in exponential time. At the time, they conjectured that the problem with $k$ curves is NP-hard for general $k$. Our polynomial time algorithm refutes this unless P equals NP.}, }
Endnote
%0 Report %A Abrahamsen, Mikkel %A Adamaszek, Anna %A Bringmann, Karl %A Cohen-Addad, Vincent %A Mehr, Mehran %A Rotenberg, Eva %A Roytman, Alan %A Thorup, Mikkel %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations External Organizations External Organizations %T Fast Fencing : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3DFE-E %U http://arxiv.org/abs/1804.00101 %D 2018 %X We consider very natural "fence enclosure" problems studied by Capoyleas, Rote, and Woeginger and Arkin, Khuller, and Mitchell in the early 90s. Given a set $S$ of $n$ points in the plane, we aim at finding a set of closed curves such that (1) each point is enclosed by a curve and (2) the total length of the curves is minimized. We consider two main variants. In the first variant, we pay a unit cost per curve in addition to the total length of the curves. An equivalent formulation of this version is that we have to enclose $n$ unit disks, paying only the total length of the enclosing curves. In the other variant, we are allowed to use at most $k$ closed curves and pay no cost per curve. For the variant with at most $k$ closed curves, we present an algorithm that is polynomial in both $n$ and $k$. For the variant with unit cost per curve, or unit disks, we present a near-linear time algorithm. Capoyleas, Rote, and Woeginger solved the problem with at most $k$ curves in $n^{O(k)}$ time. Arkin, Khuller, and Mitchell used this to solve the unit cost per curve version in exponential time. At the time, they conjectured that the problem with $k$ curves is NP-hard for general $k$. Our polynomial time algorithm refutes this unless P equals NP. %K Computer Science, Computational Geometry, cs.CG
[4]
A. Adamaszek, P. Chalermsook, A. Ene, and A. Wiese, “Submodular Unsplittable Flow on Trees,” Mathematical Programming / B, 2018.
Export
BibTeX
@article{Adamaszek2018, TITLE = {Submodular Unsplittable Flow on Trees}, AUTHOR = {Adamaszek, Anna and Chalermsook, Parinya and Ene, Alina and Wiese, Andreas}, LANGUAGE = {eng}, ISSN = {0025-5610}, DOI = {10.1007/s10107-017-1218-4}, PUBLISHER = {Springer}, ADDRESS = {Berlin}, YEAR = {2018}, JOURNAL = {Mathematical Programming / B}, }
Endnote
%0 Journal Article %A Adamaszek, Anna %A Chalermsook, Parinya %A Ene, Alina %A Wiese, Andreas %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Submodular Unsplittable Flow on Trees : %G eng %U http://hdl.handle.net/21.11116/0000-0000-73B6-1 %R 10.1007/s10107-017-1218-4 %7 2018-01-17 %D 2018 %8 17.01.2018 %J Mathematical Programming / B %I Springer %C Berlin %@ false
[5]
J. Baldus and K. Bringmann, “A Fast Implementation of Near Neighbors Queries for Fréchet Distance (GIS Cup),” 2018. [Online]. Available: http://arxiv.org/abs/1803.00806. (arXiv: 1803.00806)
Abstract
This paper describes an implementation of fast near-neighbours queries (also known as range searching) with respect to the Fr\'echet distance. The algorithm is designed to be efficient on practical data such as GPS trajectories. Our approach is to use a quadtree data structure to enumerate all curves in the database that have similar start and endpoints as the query curve. On these curves we run positive and negative filters to narrow the set of potential results. Only for those trajectories where these heuristics fail, we compute the Fr\'echet distance exactly, by running a novel recursive variant of the classic free-space diagram algorithm. Our implementation won the ACM SIGSPATIAL GIS Cup 2017.
Export
BibTeX
@online{Baldus_arXiv1803.00806, TITLE = {A Fast Implementation of Near Neighbors Queries for {F}réchet Distance ({GIS Cup})}, AUTHOR = {Baldus, Julian and Bringmann, Karl}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1803.00806}, EPRINT = {1803.00806}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {This paper describes an implementation of fast near-neighbours queries (also known as range searching) with respect to the Fr\'echet distance. The algorithm is designed to be efficient on practical data such as GPS trajectories. Our approach is to use a quadtree data structure to enumerate all curves in the database that have similar start and endpoints as the query curve. On these curves we run positive and negative filters to narrow the set of potential results. Only for those trajectories where these heuristics fail, we compute the Fr\'echet distance exactly, by running a novel recursive variant of the classic free-space diagram algorithm. Our implementation won the ACM SIGSPATIAL GIS Cup 2017.}, }
Endnote
%0 Report %A Baldus, Julian %A Bringmann, Karl %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Fast Implementation of Near Neighbors Queries for Fréchet Distance (GIS Cup) : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3E1A-E %U http://arxiv.org/abs/1803.00806 %D 2018 %X This paper describes an implementation of fast near-neighbours queries (also known as range searching) with respect to the Fr\'echet distance. The algorithm is designed to be efficient on practical data such as GPS trajectories. Our approach is to use a quadtree data structure to enumerate all curves in the database that have similar start and endpoints as the query curve. On these curves we run positive and negative filters to narrow the set of potential results. Only for those trajectories where these heuristics fail, we compute the Fr\'echet distance exactly, by running a novel recursive variant of the classic free-space diagram algorithm. Our implementation won the ACM SIGSPATIAL GIS Cup 2017. %K Computer Science, Computational Geometry, cs.CG
[6]
G. Ballard, C. Ikenmeyer, J. M. Landsberg, and N. Ryder, “The Geometry of Rank Decompositions of Matrix Multiplication II: 3 x 3 matrices,” 2018. [Online]. Available: http://arxiv.org/abs/1801.00843. (arXiv: 1801.00843)
Abstract
This is the second in a series of papers on rank decompositions of the matrix multiplication tensor. We present new rank $23$ decompositions for the $3\times 3$ matrix multiplication tensor $M_{\langle 3\rangle}$. All our decompositions have symmetry groups that include the standard cyclic permutation of factors but otherwise exhibit a range of behavior. One of them has 11 cubes as summands and admits an unexpected symmetry group of order 12. We establish basic information regarding symmetry groups of decompositions and outline two approaches for finding new rank decompositions of $M_{\langle n\rangle}$ for larger $n$.
Export
BibTeX
@online{Ballard_arXiv1801.00843, TITLE = {The geometry of rank decompositions of matrix multiplication II: $3\times 3$ matrices}, AUTHOR = {Ballard, Grey and Ikenmeyer, Christian and Landsberg, J. M. and Ryder, Nick}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1801.00843}, EPRINT = {1801.00843}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {This is the second in a series of papers on rank decompositions of the matrix multiplication tensor. We present new rank $23$ decompositions for the $3\times 3$ matrix multiplication tensor $M_{\langle 3\rangle}$. All our decompositions have symmetry groups that include the standard cyclic permutation of factors but otherwise exhibit a range of behavior. One of them has 11 cubes as summands and admits an unexpected symmetry group of order 12. We establish basic information regarding symmetry groups of decompositions and outline two approaches for finding new rank decompositions of $M_{\langle n\rangle}$ for larger $n$.}, }
Endnote
%0 Report %A Ballard, Grey %A Ikenmeyer, Christian %A Landsberg, J. M. %A Ryder, Nick %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T The Geometry of Rank Decompositions of Matrix Multiplication II: 3 x 3 matrices : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3F64-9 %U http://arxiv.org/abs/1801.00843 %D 2018 %X This is the second in a series of papers on rank decompositions of the matrix multiplication tensor. We present new rank $23$ decompositions for the $3\times 3$ matrix multiplication tensor $M_{\langle 3\rangle}$. All our decompositions have symmetry groups that include the standard cyclic permutation of factors but otherwise exhibit a range of behavior. One of them has 11 cubes as summands and admits an unexpected symmetry group of order 12. We establish basic information regarding symmetry groups of decompositions and outline two approaches for finding new rank decompositions of $M_{\langle n\rangle}$ for larger $n$. %K Computer Science, Computational Complexity, cs.CC,
[7]
R. Becker, M. Sagraloff, V. Sharma, and C. Yap, “A Simple Near-Optimal Subdivision Algorithm for Complex Root Isolation based on the Pellet Test and Newton Iteration,” Journal of Symbolic Computation, vol. 86, 2018.
Export
BibTeX
@article{Becker2017JSC, TITLE = {A Simple Near-Optimal Subdivision Algorithm for Complex Root Isolation based on the {Pellet} Test and {Newton} Iteration}, AUTHOR = {Becker, Ruben and Sagraloff, Michael and Sharma, Vikram and Yap, Chee}, LANGUAGE = {eng}, ISSN = {0747-7171}, DOI = {10.1016/j.jsc.2017.03.009}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2018}, DATE = {2018}, JOURNAL = {Journal of Symbolic Computation}, VOLUME = {86}, PAGES = {51--96}, }
Endnote
%0 Journal Article %A Becker, Ruben %A Sagraloff, Michael %A Sharma, Vikram %A Yap, Chee %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T A Simple Near-Optimal Subdivision Algorithm for Complex Root Isolation based on the Pellet Test and Newton Iteration : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5717-8 %R 10.1016/j.jsc.2017.03.009 %7 2017-03-29 %D 2018 %J Journal of Symbolic Computation %V 86 %& 51 %P 51 - 96 %I Elsevier %C Amsterdam %@ false
[8]
A. Bhattacharya, D. Issac, R. Jaiswal, and A. Kumar, “Sampling in Space Restricted Settings,” Algorithmica, vol. 80, no. 5, 2018.
Export
BibTeX
@article{Bhattacharya2018, TITLE = {Sampling in Space Restricted Settings}, AUTHOR = {Bhattacharya, Anup and Issac, Davis and Jaiswal, Ragesh and Kumar, Amit}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-017-0335-z}, PUBLISHER = {Springer-Verlag}, ADDRESS = {New York}, YEAR = {2018}, DATE = {2018}, JOURNAL = {Algorithmica}, VOLUME = {80}, NUMBER = {5}, PAGES = {1439--1458}, }
Endnote
%0 Journal Article %A Bhattacharya, Anup %A Issac, Davis %A Jaiswal, Ragesh %A Kumar, Amit %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Sampling in Space Restricted Settings : %G eng %U http://hdl.handle.net/21.11116/0000-0001-2C37-1 %R 10.1007/s00453-017-0335-z %7 2017 %D 2018 %J Algorithmica %V 80 %N 5 %& 1439 %P 1439 - 1458 %I Springer-Verlag %C New York %@ false
[9]
M. Bläser, C. Ikenmeyer, G. Jindal, and V. Lysikov, “Generalized Matrix Completion and Algebraic Natural Proofs Contact Add Comment RSS-Feed,” Electronic Colloquium on Computational Complexity (ECCC): Report Series, vol. 18–064, 2018.
Export
BibTeX
@article{BlaeserCCC18_064, TITLE = {Generalized Matrix Completion and Algebraic Natural Proofs Contact Add Comment {RSS}-Feed}, AUTHOR = {Bl{\"a}ser, Markus and Ikenmeyer, Christian and Jindal, Gorav and Lysikov, Vladimir}, LANGUAGE = {eng}, ISSN = {1433-8092}, PUBLISHER = {Hasso-Plattner-Institut f{\"u}r Softwaretechnik GmbH}, ADDRESS = {Potsdam}, YEAR = {2018}, JOURNAL = {Electronic Colloquium on Computational Complexity (ECCC): Report Series}, VOLUME = {18-064}, PAGES = {1--27}, }
Endnote
%0 Journal Article %A Bläser, Markus %A Ikenmeyer, Christian %A Jindal, Gorav %A Lysikov, Vladimir %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Generalized Matrix Completion and Algebraic Natural Proofs Contact Add Comment RSS-Feed : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3F5F-0 %7 2018 %D 2018 %J Electronic Colloquium on Computational Complexity (ECCC): Report Series %V 18-064 %& 1 %P 1 - 27 %I Hasso-Plattner-Institut für Softwaretechnik GmbH %C Potsdam %@ false %U https://eccc.weizmann.ac.il/report/2018/064/
[10]
K. Bringmann, P. Gawrychowski, S. Mozes, and O. Weimann, “Tree Edit Distance Cannot be Computed in Strongly Subcubic Time (unless APSP can),” in SODA’18, Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 2018.
Export
BibTeX
@inproceedings{Bringmann_SODA18b, TITLE = {Tree Edit Distance Cannot be Computed in Strongly Subcubic Time (unless {APSP} can)}, AUTHOR = {Bringmann, Karl and Gawrychowski, Pawe{\l} and Mozes, Shay and Weimann, Oren}, LANGUAGE = {eng}, ISBN = {978-1-61197-503-1}, DOI = {10.1137/1.9781611975031.77}, PUBLISHER = {SIAM}, YEAR = {2018}, DATE = {2018}, BOOKTITLE = {SODA'18, Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms}, EDITOR = {Czumaj, Artur}, PAGES = {1190--1206}, ADDRESS = {New Orleans, LA, USA}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Gawrychowski, Paweł %A Mozes, Shay %A Weimann, Oren %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Tree Edit Distance Cannot be Computed in Strongly Subcubic Time (unless APSP can) : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F13-5 %R 10.1137/1.9781611975031.77 %D 2018 %B Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2018-01-07 - 2018-01-10 %C New Orleans, LA, USA %B SODA'18 %E Czumaj, Artur %P 1190 - 1206 %I SIAM %@ 978-1-61197-503-1
[11]
K. Bringmann, S. Cabello, and M. T. M. Emmerich, “Maximum Volume Subset Selection for Anchored Boxes,” 2018. [Online]. Available: http://arxiv.org/abs/1803.00849. (arXiv: 1803.00849)
Abstract
Let $B$ be a set of $n$ axis-parallel boxes in $\mathbb{R}^d$ such that each box has a corner at the origin and the other corner in the positive quadrant of $\mathbb{R}^d$, and let $k$ be a positive integer. We study the problem of selecting $k$ boxes in $B$ that maximize the volume of the union of the selected boxes. This research is motivated by applications in skyline queries for databases and in multicriteria optimization, where the problem is known as the hypervolume subset selection problem. It is known that the problem can be solved in polynomial time in the plane, while the best known running time in any dimension $d \ge 3$ is $\Omega\big(\binom{n}{k}\big)$. We show that: - The problem is NP-hard already in 3 dimensions. - In 3 dimensions, we break the bound $\Omega\big(\binom{n}{k}\big)$, by providing an $n^{O(\sqrt{k})}$ algorithm. - For any constant dimension $d$, we present an efficient polynomial-time approximation scheme.
Export
BibTeX
@online{escidoc:2584984, TITLE = {Maximum Volume Subset Selection for Anchored Boxes}, AUTHOR = {Bringmann, Karl and Cabello, Sergio and Emmerich, Michael T. M.}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1803.00849}, DOI = {Bringmann_arXiv1803.00849}, EPRINT = {1803.00849}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {Let $B$ be a set of $n$ axis-parallel boxes in $\mathbb{R}^d$ such that each box has a corner at the origin and the other corner in the positive quadrant of $\mathbb{R}^d$, and let $k$ be a positive integer. We study the problem of selecting $k$ boxes in $B$ that maximize the volume of the union of the selected boxes. This research is motivated by applications in skyline queries for databases and in multicriteria optimization, where the problem is known as the hypervolume subset selection problem. It is known that the problem can be solved in polynomial time in the plane, while the best known running time in any dimension $d \ge 3$ is $\Omega\big(\binom{n}{k}\big)$. We show that: -- The problem is NP-hard already in 3 dimensions. -- In 3 dimensions, we break the bound $\Omega\big(\binom{n}{k}\big)$, by providing an $n^{O(\sqrt{k})}$ algorithm. -- For any constant dimension $d$, we present an efficient polynomial-time approximation scheme.}, }
Endnote
%0 Report %A Bringmann, Karl %A Cabello, Sergio %A Emmerich, Michael T. M. %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Maximum Volume Subset Selection for Anchored Boxes : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3E08-2 %U http://arxiv.org/abs/1803.00849 %R Bringmann_arXiv1803.00849 %D 2018 %X Let $B$ be a set of $n$ axis-parallel boxes in $\mathbb{R}^d$ such that each box has a corner at the origin and the other corner in the positive quadrant of $\mathbb{R}^d$, and let $k$ be a positive integer. We study the problem of selecting $k$ boxes in $B$ that maximize the volume of the union of the selected boxes. This research is motivated by applications in skyline queries for databases and in multicriteria optimization, where the problem is known as the hypervolume subset selection problem. It is known that the problem can be solved in polynomial time in the plane, while the best known running time in any dimension $d \ge 3$ is $\Omega\big(\binom{n}{k}\big)$. We show that: - The problem is NP-hard already in 3 dimensions. - In 3 dimensions, we break the bound $\Omega\big(\binom{n}{k}\big)$, by providing an $n^{O(\sqrt{k})}$ algorithm. - For any constant dimension $d$, we present an efficient polynomial-time approximation scheme. %K Computer Science, Computational Geometry, cs.CG,Computer Science, Data Structures and Algorithms, cs.DS
[12]
K. Bringmann and P. Wellnitz, “Clique-Based Lower Bounds for Parsing Tree-Adjoining Grammars,” 2018. [Online]. Available: http://arxiv.org/abs/1803.00804. (arXiv: 1803.00804)
Abstract
Tree-adjoining grammars are a generalization of context-free grammars that are well suited to model human languages and are thus popular in computational linguistics. In the tree-adjoining grammar recognition problem, given a grammar $\Gamma$ and a string $s$ of length $n$, the task is to decide whether $s$ can be obtained from $\Gamma$. Rajasekaran and Yooseph's parser (JCSS'98) solves this problem in time $O(n^{2\omega})$, where $\omega < 2.373$ is the matrix multiplication exponent. The best algorithms avoiding fast matrix multiplication take time $O(n^6)$. The first evidence for hardness was given by Satta (J. Comp. Linguist.'94): For a more general parsing problem, any algorithm that avoids fast matrix multiplication and is significantly faster than $O(|\Gamma| n^6)$ in the case of $|\Gamma| = \Theta(n^{12})$ would imply a breakthrough for Boolean matrix multiplication. Following an approach by Abboud et al. (FOCS'15) for context-free grammar recognition, in this paper we resolve many of the disadvantages of the previous lower bound. We show that, even on constant-size grammars, any improvement on Rajasekaran and Yooseph's parser would imply a breakthrough for the $k$-Clique problem. This establishes tree-adjoining grammar parsing as a practically relevant problem with the unusual running time of $n^{2\omega}$, up to lower order factors.
Export
BibTeX
@online{Bringmann_arXiv1803.00804, TITLE = {Clique-Based Lower Bounds for Parsing Tree-Adjoining Grammars}, AUTHOR = {Bringmann, Karl and Wellnitz, Philip}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1803.00804}, EPRINT = {1803.00804}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {Tree-adjoining grammars are a generalization of context-free grammars that are well suited to model human languages and are thus popular in computational linguistics. In the tree-adjoining grammar recognition problem, given a grammar $\Gamma$ and a string $s$ of length $n$, the task is to decide whether $s$ can be obtained from $\Gamma$. Rajasekaran and Yooseph's parser (JCSS'98) solves this problem in time $O(n^{2\omega})$, where $\omega < 2.373$ is the matrix multiplication exponent. The best algorithms avoiding fast matrix multiplication take time $O(n^6)$. The first evidence for hardness was given by Satta (J. Comp. Linguist.'94): For a more general parsing problem, any algorithm that avoids fast matrix multiplication and is significantly faster than $O(|\Gamma| n^6)$ in the case of $|\Gamma| = \Theta(n^{12})$ would imply a breakthrough for Boolean matrix multiplication. Following an approach by Abboud et al. (FOCS'15) for context-free grammar recognition, in this paper we resolve many of the disadvantages of the previous lower bound. We show that, even on constant-size grammars, any improvement on Rajasekaran and Yooseph's parser would imply a breakthrough for the $k$-Clique problem. This establishes tree-adjoining grammar parsing as a practically relevant problem with the unusual running time of $n^{2\omega}$, up to lower order factors.}, }
Endnote
%0 Report %A Bringmann, Karl %A Wellnitz, Philip %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Clique-Based Lower Bounds for Parsing Tree-Adjoining Grammars : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3E2A-C %U http://arxiv.org/abs/1803.00804 %D 2018 %X Tree-adjoining grammars are a generalization of context-free grammars that are well suited to model human languages and are thus popular in computational linguistics. In the tree-adjoining grammar recognition problem, given a grammar $\Gamma$ and a string $s$ of length $n$, the task is to decide whether $s$ can be obtained from $\Gamma$. Rajasekaran and Yooseph's parser (JCSS'98) solves this problem in time $O(n^{2\omega})$, where $\omega < 2.373$ is the matrix multiplication exponent. The best algorithms avoiding fast matrix multiplication take time $O(n^6)$. The first evidence for hardness was given by Satta (J. Comp. Linguist.'94): For a more general parsing problem, any algorithm that avoids fast matrix multiplication and is significantly faster than $O(|\Gamma| n^6)$ in the case of $|\Gamma| = \Theta(n^{12})$ would imply a breakthrough for Boolean matrix multiplication. Following an approach by Abboud et al. (FOCS'15) for context-free grammar recognition, in this paper we resolve many of the disadvantages of the previous lower bound. We show that, even on constant-size grammars, any improvement on Rajasekaran and Yooseph's parser would imply a breakthrough for the $k$-Clique problem. This establishes tree-adjoining grammar parsing as a practically relevant problem with the unusual running time of $n^{2\omega}$, up to lower order factors. %K Computer Science, Computational Complexity, cs.CC,Computer Science, Data Structures and Algorithms, cs.DS
[13]
K. Bringmann and S. Krinninger, “A Note on Hardness of Diameter Approximation,” Information Processing Letters, vol. 133, 2018.
Export
BibTeX
@article{Bringmann2018, TITLE = {A Note on Hardness of Diameter Approximation}, AUTHOR = {Bringmann, Karl and Krinninger, Sebastian}, LANGUAGE = {eng}, ISSN = {0020-0190}, DOI = {10.1016/j.ipl.2017.12.010}, PUBLISHER = {Elsevier}, YEAR = {2018}, DATE = {2018}, JOURNAL = {Information Processing Letters}, VOLUME = {133}, PAGES = {10--15}, }
Endnote
%0 Journal Article %A Bringmann, Karl %A Krinninger, Sebastian %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T A Note on Hardness of Diameter Approximation : %G eng %U http://hdl.handle.net/21.11116/0000-0001-2C44-2 %R 10.1016/j.ipl.2017.12.010 %7 2018 %D 2018 %J Information Processing Letters %V 133 %& 10 %P 10 - 15 %I Elsevier %@ false
[14]
K. Bringmann and M. Künnemann, “Multivariate Fine-Grained Complexity of Longest Common Subsequence,” 2018. [Online]. Available: http://arxiv.org/abs/1803.00938. (arXiv: 1803.00938)
Abstract
We revisit the classic combinatorial pattern matching problem of finding a longest common subsequence (LCS). For strings $x$ and $y$ of length $n$, a textbook algorithm solves LCS in time $O(n^2)$, but although much effort has been spent, no $O(n^{2-\varepsilon})$-time algorithm is known. Recent work indeed shows that such an algorithm would refute the Strong Exponential Time Hypothesis (SETH) [Abboud, Backurs, Vassilevska Williams + Bringmann, K\"unnemann FOCS'15]. Despite the quadratic-time barrier, for over 40 years an enduring scientific interest continued to produce fast algorithms for LCS and its variations. Particular attention was put into identifying and exploiting input parameters that yield strongly subquadratic time algorithms for special cases of interest, e.g., differential file comparison. This line of research was successfully pursued until 1990, at which time significant improvements came to a halt. In this paper, using the lens of fine-grained complexity, our goal is to (1) justify the lack of further improvements and (2) determine whether some special cases of LCS admit faster algorithms than currently known. To this end, we provide a systematic study of the multivariate complexity of LCS, taking into account all parameters previously discussed in the literature: the input size $n:=\max\{|x|,|y|\}$, the length of the shorter string $m:=\min\{|x|,|y|\}$, the length $L$ of an LCS of $x$ and $y$, the numbers of deletions $\delta := m-L$ and $\Delta := n-L$, the alphabet size, as well as the numbers of matching pairs $M$ and dominant pairs $d$. For any class of instances defined by fixing each parameter individually to a polynomial in terms of the input size, we prove a SETH-based lower bound matching one of three known algorithms. Specifically, we determine the optimal running time for LCS under SETH as $(n+\min\{d, \delta \Delta, \delta m\})^{1\pm o(1)}$. [...]
Export
BibTeX
@online{Bringmann_arXiv1803.00938, TITLE = {Multivariate Fine-Grained Complexity of Longest Common Subsequence}, AUTHOR = {Bringmann, Karl and K{\"u}nnemann, Marvin}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1803.00938}, EPRINT = {1803.00938}, EPRINTTYPE = {arXiv}, YEAR = {2018}, ABSTRACT = {We revisit the classic combinatorial pattern matching problem of finding a longest common subsequence (LCS). For strings $x$ and $y$ of length $n$, a textbook algorithm solves LCS in time $O(n^2)$, but although much effort has been spent, no $O(n^{2-\varepsilon})$-time algorithm is known. Recent work indeed shows that such an algorithm would refute the Strong Exponential Time Hypothesis (SETH) [Abboud, Backurs, Vassilevska Williams + Bringmann, K\"unnemann FOCS'15]. Despite the quadratic-time barrier, for over 40 years an enduring scientific interest continued to produce fast algorithms for LCS and its variations. Particular attention was put into identifying and exploiting input parameters that yield strongly subquadratic time algorithms for special cases of interest, e.g., differential file comparison. This line of research was successfully pursued until 1990, at which time significant improvements came to a halt. In this paper, using the lens of fine-grained complexity, our goal is to (1) justify the lack of further improvements and (2) determine whether some special cases of LCS admit faster algorithms than currently known. To this end, we provide a systematic study of the multivariate complexity of LCS, taking into account all parameters previously discussed in the literature: the input size $n:=\max\{|x|,|y|\}$, the length of the shorter string $m:=\min\{|x|,|y|\}$, the length $L$ of an LCS of $x$ and $y$, the numbers of deletions $\delta := m-L$ and $\Delta := n-L$, the alphabet size, as well as the numbers of matching pairs $M$ and dominant pairs $d$. For any class of instances defined by fixing each parameter individually to a polynomial in terms of the input size, we prove a SETH-based lower bound matching one of three known algorithms. Specifically, we determine the optimal running time for LCS under SETH as $(n+\min\{d, \delta \Delta, \delta m\})^{1\pm o(1)}$. [...]}, }
Endnote
%0 Report %A Bringmann, Karl %A K&#252;nnemann, Marvin %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Multivariate Fine-Grained Complexity of Longest Common Subsequence : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3E02-8 %U http://arxiv.org/abs/1803.00938 %D 2018 %X We revisit the classic combinatorial pattern matching problem of finding a longest common subsequence (LCS). For strings $x$ and $y$ of length $n$, a textbook algorithm solves LCS in time $O(n^2)$, but although much effort has been spent, no $O(n^{2-\varepsilon})$-time algorithm is known. Recent work indeed shows that such an algorithm would refute the Strong Exponential Time Hypothesis (SETH) [Abboud, Backurs, Vassilevska Williams + Bringmann, K\"unnemann FOCS'15]. Despite the quadratic-time barrier, for over 40 years an enduring scientific interest continued to produce fast algorithms for LCS and its variations. Particular attention was put into identifying and exploiting input parameters that yield strongly subquadratic time algorithms for special cases of interest, e.g., differential file comparison. This line of research was successfully pursued until 1990, at which time significant improvements came to a halt. In this paper, using the lens of fine-grained complexity, our goal is to (1) justify the lack of further improvements and (2) determine whether some special cases of LCS admit faster algorithms than currently known. To this end, we provide a systematic study of the multivariate complexity of LCS, taking into account all parameters previously discussed in the literature: the input size $n:=\max\{|x|,|y|\}$, the length of the shorter string $m:=\min\{|x|,|y|\}$, the length $L$ of an LCS of $x$ and $y$, the numbers of deletions $\delta := m-L$ and $\Delta := n-L$, the alphabet size, as well as the numbers of matching pairs $M$ and dominant pairs $d$. For any class of instances defined by fixing each parameter individually to a polynomial in terms of the input size, we prove a SETH-based lower bound matching one of three known algorithms. Specifically, we determine the optimal running time for LCS under SETH as $(n+\min\{d, \delta \Delta, \delta m\})^{1\pm o(1)}$. [...] %K Computer Science, Computational Complexity, cs.CC,Computer Science, Data Structures and Algorithms, cs.DS
[15]
K. Bringmann and M. Künnemann, “Multivariate Fine-Grained Complexity of Longest Common Subsequence,” in SODA’18, Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 2018.
Export
BibTeX
@inproceedings{Bringmann_SODA18, TITLE = {Multivariate Fine-Grained Complexity of Longest Common Subsequence}, AUTHOR = {Bringmann, Karl and K{\"u}nnemann, Marvin}, LANGUAGE = {eng}, ISBN = {978-1-61197-503-1}, DOI = {10.1137/1.9781611975031.79}, PUBLISHER = {SIAM}, YEAR = {2018}, DATE = {2018}, BOOKTITLE = {SODA'18, Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms}, EDITOR = {Czumaj, Artur}, PAGES = {1216--1235}, ADDRESS = {New Orleans, LA, USA}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A K&#252;nnemann, Marvin %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Multivariate Fine-Grained Complexity of Longest Common Subsequence : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F0E-C %R 10.1137/1.9781611975031.79 %D 2018 %B Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2018-01-07 - 2018-01-10 %C New Orleans, LA, USA %B SODA'18 %E Czumaj, Artur %P 1216 - 1235 %I SIAM %@ 978-1-61197-503-1
[16]
J. Bund, C. Lenzen, and M. Medina, “Optimal Metastability-containing Sorting Networks,” in Proceedings of the 2018 Design, Automation & Test in Europe (DATE 2018), Dresden, Germany, 2018.
Export
BibTeX
@inproceedings{Bund_DATE2018, TITLE = {Optimal Metastability-containing Sorting Networks}, AUTHOR = {Bund, Johannes and Lenzen, Christoph and Medina, Moti}, LANGUAGE = {eng}, ISBN = {978-3-9819263-1-6}, DOI = {10.23919/DATE.2018.8342063}, PUBLISHER = {IEEE}, YEAR = {2018}, DATE = {2018}, BOOKTITLE = {Proceedings of the 2018 Design, Automation \& Test in Europe (DATE 2018)}, PAGES = {521--526}, ADDRESS = {Dresden, Germany}, }
Endnote
%0 Conference Proceedings %A Bund, Johannes %A Lenzen, Christoph %A Medina, Moti %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Optimal Metastability-containing Sorting Networks : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3F69-4 %R 10.23919/DATE.2018.8342063 %D 2018 %B Design, Automation & Test in Europe Conference & Exhibition %Z date of event: 2018-03-19 - 2018-03-23 %C Dresden, Germany %B Proceedings of the 2018 Design, Automation & Test in Europe %P 521 - 526 %I IEEE %@ 978-3-9819263-1-6
[17]
J. Garg, M. Hoefer, and K. Mehlhorn, “Approximating the Nash Social Welfare with Budget-Additive Valuations,” in SODA’18, Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 2018.
Export
BibTeX
@inproceedings{GargHoeferMehlhornSODA18, TITLE = {Approximating the {Nash} Social Welfare with Budget-Additive Valuations}, AUTHOR = {Garg, Jugal and Hoefer, Martin and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISBN = {978-1-61197-503-1}, DOI = {10.1137/1.9781611975031.150}, PUBLISHER = {SIAM}, YEAR = {2018}, DATE = {2018}, BOOKTITLE = {SODA'18, Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms}, EDITOR = {Czumaj, Artur}, PAGES = {2326--2340}, ADDRESS = {New Orleans, LA, USA}, }
Endnote
%0 Conference Proceedings %A Garg, Jugal %A Hoefer, Martin %A Mehlhorn, Kurt %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Approximating the Nash Social Welfare with Budget-Additive Valuations : %G eng %U http://hdl.handle.net/21.11116/0000-0000-37F9-A %R 10.1137/1.9781611975031.150 %D 2018 %B Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2018-01-07 - 2018-01-10 %C New Orleans, LA, USA %B SODA'18 %E Czumaj, Artur %P 2326 - 2340 %I SIAM %@ 978-1-61197-503-1
[18]
T. A. G. Hageman, P. A. Loethman, M. Dirnberger, M. C. Elwenspoek, A. Manz, and L. Abelmann, “Macroscopic Equivalence for Microscopic Motion in a Turbulence Driven Three-dimensional Self-assembly Reactor,” Journal of Applied Physics, vol. 123, no. 2, 2018.
Export
BibTeX
@article{Hageman2018, TITLE = {Macroscopic Equivalence for Microscopic Motion in a Turbulence Driven Three-dimensional Self-assembly Reactor}, AUTHOR = {Hageman, T. A. G. and Loethman, P. A. and Dirnberger, Michael and Elwenspoek, M. C. and Manz, A. and Abelmann, L.}, LANGUAGE = {eng}, ISSN = {0021-8979}, DOI = {10.1063/1.5007029}, PUBLISHER = {AIP Publishing}, ADDRESS = {New York, NY}, YEAR = {2018}, DATE = {2018}, JOURNAL = {Journal of Applied Physics}, VOLUME = {123}, NUMBER = {2}, PAGES = {1--10}, EID = {024901}, }
Endnote
%0 Journal Article %A Hageman, T. A. G. %A Loethman, P. A. %A Dirnberger, Michael %A Elwenspoek, M. C. %A Manz, A. %A Abelmann, L. %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Macroscopic Equivalence for Microscopic Motion in a Turbulence Driven Three-dimensional Self-assembly Reactor : %G eng %U http://hdl.handle.net/21.11116/0000-0000-431A-8 %R 10.1063/1.5007029 %7 2018 %D 2018 %J Journal of Applied Physics %O J. Appl. Phys. %V 123 %N 2 %& 1 %P 1 - 10 %Z sequence number: 024901 %I AIP Publishing %C New York, NY %@ false
[19]
C. Ikenmeyer and S. Mengel, “On the Relative Power of Reduction Notions in Arithmetic Circuit Complexity,” Information Processing Letters, vol. 130, 2018.
Export
BibTeX
@article{Ikenmeyer2018, TITLE = {On the Relative Power of Reduction Notions in Arithmetic Circuit Complexity}, AUTHOR = {Ikenmeyer, Christian and Mengel, Stefan}, LANGUAGE = {eng}, ISSN = {0020-0190}, DOI = {10.1016/j.ipl.2017.09.009}, PUBLISHER = {Elsevier}, YEAR = {2018}, DATE = {2018}, JOURNAL = {Information Processing Letters}, VOLUME = {130}, PAGES = {7--10}, }
Endnote
%0 Journal Article %A Ikenmeyer, Christian %A Mengel, Stefan %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On the Relative Power of Reduction Notions in Arithmetic Circuit Complexity : %G eng %U http://hdl.handle.net/21.11116/0000-0000-0361-F %R 10.1016/j.ipl.2017.09.009 %7 2017 %D 2018 %J Information Processing Letters %V 130 %& 7 %P 7 - 10 %I Elsevier %@ false
[20]
P. Khanchandani and C. Lenzen, “Self-Stabilizing Byzantine Clock Synchronization with Optimal Precision,” Theory of Computing Systems, 2018.
Export
BibTeX
@article{_Khanchandani2018, TITLE = {Self-Stabilizing {B}yzantine Clock Synchronization with Optimal Precision}, AUTHOR = {Khanchandani, Pankaj and Lenzen, Christoph}, LANGUAGE = {eng}, ISSN = {1432-4350}, DOI = {10.1007/s00224-017-9840-3}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2018}, JOURNAL = {Theory of Computing Systems}, }
Endnote
%0 Journal Article %A Khanchandani, Pankaj %A Lenzen, Christoph %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Self-Stabilizing Byzantine Clock Synchronization with Optimal Precision : %G eng %U http://hdl.handle.net/21.11116/0000-0000-73AC-D %R 10.1007/s00224-017-9840-3 %7 2018-01-20 %D 2018 %8 20.01.2018 %J Theory of Computing Systems %I Springer %C New York, NY %@ false
[21]
P. Koprowski, K. Mehlhorn, and S. Ray, “Corrigendum to ‘Faster algorithms for computing Hong’s bound on absolute positiveness’ [J. Symb. Comput. 45 (2010) 677–683],” Journal of Symbolic Computation, vol. 87, 2018.
Export
BibTeX
@article{Koprowski2018, TITLE = {Corrigendum to {\textquotedblleft}Faster algorithms for computing Hong's bound on absolute positiveness{\textquotedblright} [J. Symb. Comput. 45 (2010) 677--683]}, AUTHOR = {Koprowski, Przemys{\l}aw and Mehlhorn, Kurt and Ray, Saurabh}, LANGUAGE = {eng}, ISSN = {0747-7171}, DOI = {10.1016/j.jsc.2017.05.008}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2018}, DATE = {2018}, JOURNAL = {Journal of Symbolic Computation}, VOLUME = {87}, PAGES = {238--241}, }
Endnote
%0 Journal Article %A Koprowski, Przemys&#322;aw %A Mehlhorn, Kurt %A Ray, Saurabh %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Corrigendum to &#8220;Faster algorithms for computing Hong's bound on absolute positiveness&#8221; [J. Symb. Comput. 45 (2010) 677&#8211;683] : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3C55-D %R 10.1016/j.jsc.2017.05.008 %7 2017 %D 2018 %J Journal of Symbolic Computation %V 87 %& 238 %P 238 - 241 %I Elsevier %C Amsterdam %@ false
2017
[22]
A. Abboud, K. Bringmann, D. Hermelin, and D. Shabtay, “SETH-Based Lower Bounds for Subset Sum and Bicriteria Path,” 2017. [Online]. Available: http://arxiv.org/abs/1704.04546. (arXiv: 1704.04546)
Abstract
Subset-Sum and k-SAT are two of the most extensively studied problems in computer science, and conjectures about their hardness are among the cornerstones of fine-grained complexity. One of the most intriguing open problems in this area is to base the hardness of one of these problems on the other. Our main result is a tight reduction from k-SAT to Subset-Sum on dense instances, proving that Bellman's 1962 pseudo-polynomial $O^{*}(T)$-time algorithm for Subset-Sum on $n$ numbers and target $T$ cannot be improved to time $T^{1-\varepsilon}\cdot 2^{o(n)}$ for any $\varepsilon>0$, unless the Strong Exponential Time Hypothesis (SETH) fails. This is one of the strongest known connections between any two of the core problems of fine-grained complexity. As a corollary, we prove a "Direct-OR" theorem for Subset-Sum under SETH, offering a new tool for proving conditional lower bounds: It is now possible to assume that deciding whether one out of $N$ given instances of Subset-Sum is a YES instance requires time $(N T)^{1-o(1)}$. As an application of this corollary, we prove a tight SETH-based lower bound for the classical Bicriteria s,t-Path problem, which is extensively studied in Operations Research. We separate its complexity from that of Subset-Sum: On graphs with $m$ edges and edge lengths bounded by $L$, we show that the $O(Lm)$ pseudo-polynomial time algorithm by Joksch from 1966 cannot be improved to $\tilde{O}(L+m)$, in contrast to a recent improvement for Subset Sum (Bringmann, SODA 2017).
Export
BibTeX
@online{DBLP:journals/corr/AbboudBHS17, TITLE = {{SETH}-Based Lower Bounds for Subset Sum and Bicriteria Path}, AUTHOR = {Abboud, Amir and Bringmann, Karl and Hermelin, Danny and Shabtay, Dvir}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1704.04546}, EPRINT = {1704.04546}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Subset-Sum and k-SAT are two of the most extensively studied problems in computer science, and conjectures about their hardness are among the cornerstones of fine-grained complexity. One of the most intriguing open problems in this area is to base the hardness of one of these problems on the other. Our main result is a tight reduction from k-SAT to Subset-Sum on dense instances, proving that Bellman's 1962 pseudo-polynomial $O^{*}(T)$-time algorithm for Subset-Sum on $n$ numbers and target $T$ cannot be improved to time $T^{1-\varepsilon}\cdot 2^{o(n)}$ for any $\varepsilon>0$, unless the Strong Exponential Time Hypothesis (SETH) fails. This is one of the strongest known connections between any two of the core problems of fine-grained complexity. As a corollary, we prove a "Direct-OR" theorem for Subset-Sum under SETH, offering a new tool for proving conditional lower bounds: It is now possible to assume that deciding whether one out of $N$ given instances of Subset-Sum is a YES instance requires time $(N T)^{1-o(1)}$. As an application of this corollary, we prove a tight SETH-based lower bound for the classical Bicriteria s,t-Path problem, which is extensively studied in Operations Research. We separate its complexity from that of Subset-Sum: On graphs with $m$ edges and edge lengths bounded by $L$, we show that the $O(Lm)$ pseudo-polynomial time algorithm by Joksch from 1966 cannot be improved to $\tilde{O}(L+m)$, in contrast to a recent improvement for Subset Sum (Bringmann, SODA 2017).}, }
Endnote
%0 Report %A Abboud, Amir %A Bringmann, Karl %A Hermelin, Danny %A Shabtay, Dvir %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T SETH-Based Lower Bounds for Subset Sum and Bicriteria Path : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-89E5-3 %U http://arxiv.org/abs/1704.04546 %D 2017 %X Subset-Sum and k-SAT are two of the most extensively studied problems in computer science, and conjectures about their hardness are among the cornerstones of fine-grained complexity. One of the most intriguing open problems in this area is to base the hardness of one of these problems on the other. Our main result is a tight reduction from k-SAT to Subset-Sum on dense instances, proving that Bellman's 1962 pseudo-polynomial $O^{*}(T)$-time algorithm for Subset-Sum on $n$ numbers and target $T$ cannot be improved to time $T^{1-\varepsilon}\cdot 2^{o(n)}$ for any $\varepsilon>0$, unless the Strong Exponential Time Hypothesis (SETH) fails. This is one of the strongest known connections between any two of the core problems of fine-grained complexity. As a corollary, we prove a "Direct-OR" theorem for Subset-Sum under SETH, offering a new tool for proving conditional lower bounds: It is now possible to assume that deciding whether one out of $N$ given instances of Subset-Sum is a YES instance requires time $(N T)^{1-o(1)}$. As an application of this corollary, we prove a tight SETH-based lower bound for the classical Bicriteria s,t-Path problem, which is extensively studied in Operations Research. We separate its complexity from that of Subset-Sum: On graphs with $m$ edges and edge lengths bounded by $L$, we show that the $O(Lm)$ pseudo-polynomial time algorithm by Joksch from 1966 cannot be improved to $\tilde{O}(L+m)$, in contrast to a recent improvement for Subset Sum (Bringmann, SODA 2017). %K Computer Science, Data Structures and Algorithms, cs.DS,Computer Science, Computational Complexity, cs.CC
[23]
A. Abboud, A. Backurs, K. Bringmann, and M. Künnemann, “Fine-Grained Complexity of Analyzing Compressed Data: Quantifying Improvements over Decompress-And-Solve,” in 58th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2017), Berkeley, CA, USA, 2017.
Export
BibTeX
@inproceedings{Abboud_FOCS2017, TITLE = {Fine-Grained Complexity of Analyzing Compressed Data: {Q}uantifying Improvements over Decompress-And-Solve}, AUTHOR = {Abboud, Amir and Backurs, Arturs and Bringmann, Karl and K{\"u}nnemann, Marvin}, LANGUAGE = {eng}, ISBN = {978-1-5386-3464-6}, DOI = {10.1109/FOCS.2017.26}, PUBLISHER = {IEEE}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {58th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2017)}, PAGES = {192--203}, ADDRESS = {Berkeley, CA, USA}, }
Endnote
%0 Conference Proceedings %A Abboud, Amir %A Backurs, Arturs %A Bringmann, Karl %A K&#252;nnemann, Marvin %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fine-Grained Complexity of Analyzing Compressed Data: Quantifying Improvements over Decompress-And-Solve : %G eng %U http://hdl.handle.net/21.11116/0000-0000-0475-8 %R 10.1109/FOCS.2017.26 %D 2017 %B 58th Annual IEEE Symposium on Foundations of Computer Science %Z date of event: 2017-10-15 - 2017-10-17 %C Berkeley, CA, USA %B 58th Annual IEEE Symposium on Foundations of Computer Science %P 192 - 203 %I IEEE %@ 978-1-5386-3464-6
[24]
I. Abraham, S. Chechik, and S. Krinninger, “Fully Dynamic All-pairs Shortest Paths with Worst-case Update-time Revisited,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{AbrahamSODA2017, TITLE = {Fully Dynamic All-pairs Shortest Paths with Worst-case Update-time Revisited}, AUTHOR = {Abraham, Ittai and Chechik, Shiri and Krinninger, Sebastian}, LANGUAGE = {eng}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, PAGES = {440--452}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Abraham, Ittai %A Chechik, Shiri %A Krinninger, Sebastian %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fully Dynamic All-pairs Shortest Paths with Worst-case Update-time Revisited : %G eng %U http://hdl.handle.net/21.11116/0000-0001-4FB7-9 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %P 440 - 452 %I SIAM
[25]
I. Abraham, S. Chechik, and S. Krinninger, “Fully dynamic all-pairs shortest paths with worst-case update-time,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{AbrahamCK17, TITLE = {Fully dynamic all-pairs shortest paths with worst-case update-time}, AUTHOR = {Abraham, Ittai and Chechik, Shiri and Krinninger, Sebastian}, LANGUAGE = {eng}, ISBN = {978-1-61197-478-2}, DOI = {10.1137/1.9781611974782.28}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, EDITOR = {Klein, Philip N.}, PAGES = {440--452}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Abraham, Ittai %A Chechik, Shiri %A Krinninger, Sebastian %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fully dynamic all-pairs shortest paths with worst-case update-time : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-52D0-1 %R 10.1137/1.9781611974782.28 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %E Klein, Philip N. %P 440 - 452 %I SIAM %@ 978-1-61197-478-2
[26]
A. Adamaszek, M. P. Renault, A. Rosen, and R. van Stee, “Reordering Buffer Management with Advice,” Journal of Scheduling, vol. 20, no. 5, 2017.
Export
BibTeX
@article{Adamaszek2017, TITLE = {Reordering Buffer Management with Advice}, AUTHOR = {Adamaszek, Anna and Renault, Marc P. and Rosen, Adi and van Stee, Rob}, LANGUAGE = {eng}, ISSN = {1094-6136}, DOI = {10.1007/s10951-016-0487-8}, PUBLISHER = {Wiley}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Journal of Scheduling}, VOLUME = {20}, NUMBER = {5}, PAGES = {423--442}, }
Endnote
%0 Journal Article %A Adamaszek, Anna %A Renault, Marc P. %A Rosen, Adi %A van Stee, Rob %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Reordering Buffer Management with Advice : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-0E60-C %R 10.1007/s10951-016-0487-8 %7 2016-06-17 %D 2017 %J Journal of Scheduling %V 20 %N 5 %& 423 %P 423 - 442 %I Wiley %C New York, NY %@ false
[27]
N. Alon, S. Moran, and A. Yehudayoff, “Sign Rank versus Vapnik-Chervonenkis Dimension,” Sbornik: Mathematics, vol. 208, no. 12, 2017.
Export
BibTeX
@article{Alon2017, TITLE = {Sign Rank versus {V}apnik-{C}hervonenkis Dimension}, AUTHOR = {Alon, Noga and Moran, Shay and Yehudayoff, Amir}, LANGUAGE = {eng}, ISSN = {1064-5616}, DOI = {10.1070/SM8780}, PUBLISHER = {Mathematical Society, Turpion Ltd.}, ADDRESS = {London}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Sbornik: Mathematics}, VOLUME = {208}, NUMBER = {12}, PAGES = {1724--1757}, }
Endnote
%0 Journal Article %A Alon, Noga %A Moran, Shay %A Yehudayoff, Amir %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Sign Rank versus Vapnik-Chervonenkis Dimension : %G eng %U http://hdl.handle.net/21.11116/0000-0000-C8D2-1 %R 10.1070/SM8780 %7 2017 %D 2017 %J Sbornik: Mathematics %O Sb. Math. %V 208 %N 12 %& 1724 %P 1724 - 1757 %I Mathematical Society, Turpion Ltd. %C London %@ false
[28]
E. Althaus, B. Beber, W. Damm, S. Disch, W. Hagemann, A. Rakow, C. Scholl, U. Waldmann, and B. Wirtz, “Verification of Linear Hybrid Systems with Large Discrete State Spaces Using Counterexample-guided Abstraction Refinement,” Science of Computer Programming, vol. 148, 2017.
Export
BibTeX
@article{Althaus2017, TITLE = {Verification of Linear Hybrid Systems with Large Discrete State Spaces Using Counterexample-guided Abstraction Refinement}, AUTHOR = {Althaus, Ernst and Beber, Bj{\"o}rn and Damm, Werner and Disch, Stefan and Hagemann, Willem and Rakow, Astrid and Scholl, Christoph and Waldmann, Uwe and Wirtz, Boris}, LANGUAGE = {eng}, ISSN = {0167-6423}, DOI = {10.1016/j.scico.2017.04.010}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Science of Computer Programming}, VOLUME = {148}, PAGES = {123--160}, }
Endnote
%0 Journal Article %A Althaus, Ernst %A Beber, Bj&#246;rn %A Damm, Werner %A Disch, Stefan %A Hagemann, Willem %A Rakow, Astrid %A Scholl, Christoph %A Waldmann, Uwe %A Wirtz, Boris %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations Automation of Logic, MPI for Informatics, Max Planck Society External Organizations External Organizations Automation of Logic, MPI for Informatics, Max Planck Society External Organizations %T Verification of Linear Hybrid Systems with Large Discrete State Spaces Using Counterexample-guided Abstraction Refinement : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-1C23-5 %R 10.1016/j.scico.2017.04.010 %7 2017-05-10 %D 2017 %J Science of Computer Programming %V 148 %& 123 %P 123 - 160 %I Elsevier %C Amsterdam %@ false
[29]
S. Anand, K. Bringmann, T. Friedrich, N. Garg, and A. Kumar, “Minimizing Maximum (Weighted) Flow-Time on Related and Unrelated Machines,” Algorithmica, vol. 77, no. 2, 2017.
Export
BibTeX
@article{DBLP:journals/algorithmica/0002B0G017, TITLE = {Minimizing Maximum (Weighted) Flow-Time on Related and Unrelated Machines}, AUTHOR = {Anand, S. and Bringmann, Karl and Friedrich, Tobias and Garg, Naveen and Kumar, Amit}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-015-0082-y}, PUBLISHER = {Springer-Verlag}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Algorithmica}, VOLUME = {77}, NUMBER = {2}, PAGES = {515--536}, }
Endnote
%0 Journal Article %A Anand, S. %A Bringmann, Karl %A Friedrich, Tobias %A Garg, Naveen %A Kumar, Amit %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Discrete Optimization, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Minimizing Maximum (Weighted) Flow-Time on Related and Unrelated Machines : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5527-9 %R 10.1007/s00453-015-0082-y %7 2015 %D 2017 %J Algorithmica %V 77 %N 2 %& 515 %P 515 - 536 %I Springer-Verlag %C New York, NY %@ false
[30]
A. Antoniadis, N. Barcelo, M. Consuegra, P. Kling, M. Nugent, K. Pruhs, and M. Scquizzato, “Efficient Computation of Optimal Energy and Fractional Weighted Flow Trade-Off Schedules,” Algorithmica, vol. 79, no. 2, 2017.
Export
BibTeX
@article{Antoniadis2016, TITLE = {Efficient Computation of Optimal Energy and Fractional Weighted Flow Trade-Off Schedules}, AUTHOR = {Antoniadis, Antonios and Barcelo, Neal and Consuegra, Mario and Kling, Peter and Nugent, Michael and Pruhs, Kirk and Scquizzato, Michele}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-016-0208-x}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Algorithmica}, VOLUME = {79}, NUMBER = {2}, PAGES = {568--597}, }
Endnote
%0 Journal Article %A Antoniadis, Antonios %A Barcelo, Neal %A Consuegra, Mario %A Kling, Peter %A Nugent, Michael %A Pruhs, Kirk %A Scquizzato, Michele %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations External Organizations External Organizations External Organizations %T Efficient Computation of Optimal Energy and Fractional Weighted Flow Trade-Off Schedules : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-58AA-7 %R 10.1007/s00453-016-0208-x %7 2016-08-31 %D 2017 %J Algorithmica %V 79 %N 2 %& 568 %P 568 - 597 %I Springer %C New York, NY %@ false
[31]
A. Antoniadis, P. Kling, S. Ott, and S. Riechers, “Continuous Speed Scaling with Variability: A simple and Direct Approach,” Theoretical Computer Science, vol. 678, 2017.
Export
BibTeX
@article{Antoniadis2017, TITLE = {Continuous Speed Scaling with Variability: {A} simple and Direct Approach}, AUTHOR = {Antoniadis, Antonios and Kling, Peter and Ott, Sebastian and Riechers, S{\"o}ren}, LANGUAGE = {eng}, ISSN = {0304-3975}, DOI = {10.1016/j.tcs.2017.03.021}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Theoretical Computer Science}, VOLUME = {678}, PAGES = {1--13}, }
Endnote
%0 Journal Article %A Antoniadis, Antonios %A Kling, Peter %A Ott, Sebastian %A Riechers, S&#246;ren %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Continuous Speed Scaling with Variability: A simple and Direct Approach : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-7857-F %R 10.1016/j.tcs.2017.03.021 %7 2017 %D 2017 %J Theoretical Computer Science %V 678 %& 1 %P 1 - 13 %I Elsevier %C Amsterdam %@ false
[32]
Y. Azar, M. Hoefer, I. Maor, R. Reiffenhäuser, and B. Vöcking, “Truthful Mechanism Design via Correlated Tree Rounding,” Mathematical Programming / A, vol. 163, no. 1–2, 2017.
Export
BibTeX
@article{Azar2017, TITLE = {Truthful Mechanism Design via Correlated Tree Rounding}, AUTHOR = {Azar, Yossi and Hoefer, Martin and Maor, Idan and Reiffenh{\"a}user, Rebecca and V{\"o}cking, Berthold}, LANGUAGE = {eng}, ISSN = {0025-5610}, DOI = {10.1007/s10107-016-1068-5}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Mathematical Programming / A}, VOLUME = {163}, NUMBER = {1-2}, PAGES = {445--469}, }
Endnote
%0 Journal Article %A Azar, Yossi %A Hoefer, Martin %A Maor, Idan %A Reiffenh&#228;user, Rebecca %A V&#246;cking, Berthold %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Truthful Mechanism Design via Correlated Tree Rounding : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-326B-A %R 10.1007/s10107-016-1068-5 %7 2016-09-10 %D 2017 %J Mathematical Programming / A %V 163 %N 1-2 %& 445 %P 445 - 469 %I Springer %C New York, NY %@ false
[33]
J. Baldus and K. Bringmann, “A Fast Implementation of Near Neighbors Queries for Fréchet Distance (GIS Cup),” in 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems (GIS 2017), Redondo Beach, CA, USA, 2017.
Export
BibTeX
@inproceedings{Baldus_SIGSPATIAL2017, TITLE = {A Fast Implementation of Near Neighbors Queries for {F}r&#233;chet Distance ({GIS Cup})}, AUTHOR = {Baldus, Julian and Bringmann, Karl}, LANGUAGE = {eng}, ISBN = {978-1-4503-5490-5}, DOI = {10.1145/3139958.3140062}, PUBLISHER = {ACM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems (GIS 2017)}, EDITOR = {Hoel, Erik and Newsman, Shawn and Ravada, Siva and Tamassia, Roberto and Trjacevski, Goce}, EID = {99}, ADDRESS = {Redondo Beach, CA, USA}, }
Endnote
%0 Conference Proceedings %A Baldus, Julian %A Bringmann, Karl %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Fast Implementation of Near Neighbors Queries for Fr&#233;chet Distance (GIS Cup) : %G eng %U http://hdl.handle.net/21.11116/0000-0001-3E17-1 %R 10.1145/3139958.3140062 %D 2017 %B 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems %Z date of event: 2017-11-07 - 2017-11-10 %C Redondo Beach, CA, USA %B 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems %E Hoel, Erik; Newsman, Shawn; Ravada, Siva; Tamassia, Roberto; Trjacevski, Goce %Z sequence number: 99 %I ACM %@ 978-1-4503-5490-5
[34]
L. Becchetti, A. Clementi, E. Natale, F. Pasquale, R. Silvestri, and L. Trevisan, “Simple Dynamics for Plurality Consensus,” Distributed Computing, vol. 30, no. 4, 2017.
Export
BibTeX
@article{Becchetti2017, TITLE = {Simple Dynamics for Plurality Consensus}, AUTHOR = {Becchetti, Luca and Clementi, Andrea and Natale, Emanuele and Pasquale, Francesco and Silvestri, Riccardo and Trevisan, Luca}, LANGUAGE = {eng}, ISSN = {0178-2770}, DOI = {10.1007/s00446-016-0289-4}, PUBLISHER = {Springer International}, ADDRESS = {Berlin}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Distributed Computing}, VOLUME = {30}, NUMBER = {4}, PAGES = {293--306}, }
Endnote
%0 Journal Article %A Becchetti, Luca %A Clementi, Andrea %A Natale, Emanuele %A Pasquale, Francesco %A Silvestri, Riccardo %A Trevisan, Luca %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Simple Dynamics for Plurality Consensus : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-F885-F %R 10.1007/s00446-016-0289-4 %7 2016-11-22 %D 2017 %J Distributed Computing %V 30 %N 4 %& 293 %P 293 - 306 %I Springer International %C Berlin %@ false
[35]
L. Becchetti, A. Clementi, E. Natale, F. Pasquale, and L. Trevisan, “Find Your Place: Simple Distributed Algorithms for Community Detection,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{BCNPT17, TITLE = {Find Your Place: {S}imple Distributed Algorithms for Community Detection}, AUTHOR = {Becchetti, Luca and Clementi, Andrea and Natale, Emanuele and Pasquale, Francesco and Trevisan, Luca}, LANGUAGE = {eng}, DOI = {10.1137/1.9781611974782.59}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, PAGES = {940--959}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Becchetti, Luca %A Clementi, Andrea %A Natale, Emanuele %A Pasquale, Francesco %A Trevisan, Luca %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Find Your Place: Simple Distributed Algorithms for Community Detection : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5877-A %R 10.1137/1.9781611974782.59 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %P 940 - 959 %I SIAM
[36]
R. Becker, A. Karrenbauer, S. Krinninger, and C. Lenzen, “Near-Optimal Approximate Shortest Paths and Transshipment in Distributed and Streaming Models,” in 31 International Symposium on Distributed Computing (DISC 2017), Vienna, Austria, 2017.
Export
BibTeX
@inproceedings{Becker_DISC17, TITLE = {Near-Optimal Approximate Shortest Paths and Transshipment in Distributed and Streaming Models}, AUTHOR = {Becker, Ruben and Karrenbauer, Andreas and Krinninger, Sebastian and Lenzen, Christoph}, LANGUAGE = {eng}, ISBN = {978-3-95977-053-8}, URL = {urn:nbn:de:0030-drops-80031}, DOI = {10.4230/LIPIcs.DISC.2017.7}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {31 International Symposium on Distributed Computing (DISC 2017)}, EDITOR = {Richa, Andr{\'e}a W.}, PAGES = {1--16}, EID = {7}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {91}, ADDRESS = {Vienna, Austria}, }
Endnote
%0 Conference Proceedings %A Becker, Ruben %A Karrenbauer, Andreas %A Krinninger, Sebastian %A Lenzen, Christoph %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Near-Optimal Approximate Shortest Paths and Transshipment in Distributed and Streaming Models : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F19-F %U urn:nbn:de:0030-drops-80031 %R 10.4230/LIPIcs.DISC.2017.7 %D 2017 %B 31st International Symposium on Distributed Computing %Z date of event: 2017-10-16 - 2017-10-20 %C Vienna, Austria %B 31 International Symposium on Distributed Computing %E Richa, Andr&#233;a W. %P 1 - 16 %Z sequence number: 7 %I Schloss Dagstuhl %@ 978-3-95977-053-8 %B Leibniz International Proceedings in Informatics %N 91 %U http://drops.dagstuhl.de/opus/volltexte/2017/8003/http://drops.dagstuhl.de/doku/urheberrecht1.html
[37]
R. Becker, V. Bonifaci, A. Karrenbauer, P. Kolev, and K. Mehlhorn, “Two Results on Slime Mold Computations,” 2017. [Online]. Available: http://arxiv.org/abs/1707.06631. (arXiv: 1707.06631)
Abstract
In this paper, we present two results on slime mold computations. The first one treats a biologically-grounded model, originally proposed by biologists analyzing the behavior of the slime mold Physarum polycephalum. This primitive organism was empirically shown by Nakagaki et al. to solve shortest path problems in wet-lab experiments (Nature'00). We show that the proposed simple mathematical model actually generalizes to a much wider class of problems, namely undirected linear programs with a non-negative cost vector. For our second result, we consider the discretization of a biologically-inspired model. This model is a directed variant of the biologically-grounded one and was never claimed to describe the behavior of a biological system. Straszak and Vishnoi showed that it can $\epsilon$-approximately solve flow problems (SODA'16) and even general linear programs with positive cost vector (ITCS'16) within a finite number of steps. We give a refined convergence analysis that improves the dependence on $\epsilon$ from polynomial to logarithmic and simultaneously allows to choose a step size that is independent of $\epsilon$. Furthermore, we show that the dynamics can be initialized with a more general set of (infeasible) starting points.
Export
BibTeX
@online{Becker_arxiv2017, TITLE = {Two Results on Slime Mold Computations}, AUTHOR = {Becker, Ruben and Bonifaci, Vincenzo and Karrenbauer, Andreas and Kolev, Pavel and Mehlhorn, Kurt}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1707.06631}, EPRINT = {1707.06631}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {In this paper, we present two results on slime mold computations. The first one treats a biologically-grounded model, originally proposed by biologists analyzing the behavior of the slime mold Physarum polycephalum. This primitive organism was empirically shown by Nakagaki et al. to solve shortest path problems in wet-lab experiments (Nature'00). We show that the proposed simple mathematical model actually generalizes to a much wider class of problems, namely undirected linear programs with a non-negative cost vector. For our second result, we consider the discretization of a biologically-inspired model. This model is a directed variant of the biologically-grounded one and was never claimed to describe the behavior of a biological system. Straszak and Vishnoi showed that it can $\epsilon$-approximately solve flow problems (SODA'16) and even general linear programs with positive cost vector (ITCS'16) within a finite number of steps. We give a refined convergence analysis that improves the dependence on $\epsilon$ from polynomial to logarithmic and simultaneously allows to choose a step size that is independent of $\epsilon$. Furthermore, we show that the dynamics can be initialized with a more general set of (infeasible) starting points.}, }
Endnote
%0 Report %A Becker, Ruben %A Bonifaci, Vincenzo %A Karrenbauer, Andreas %A Kolev, Pavel %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Two Results on Slime Mold Computations : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-FBA8-F %U http://arxiv.org/abs/1707.06631 %D 2017 %X In this paper, we present two results on slime mold computations. The first one treats a biologically-grounded model, originally proposed by biologists analyzing the behavior of the slime mold Physarum polycephalum. This primitive organism was empirically shown by Nakagaki et al. to solve shortest path problems in wet-lab experiments (Nature'00). We show that the proposed simple mathematical model actually generalizes to a much wider class of problems, namely undirected linear programs with a non-negative cost vector. For our second result, we consider the discretization of a biologically-inspired model. This model is a directed variant of the biologically-grounded one and was never claimed to describe the behavior of a biological system. Straszak and Vishnoi showed that it can $\epsilon$-approximately solve flow problems (SODA'16) and even general linear programs with positive cost vector (ITCS'16) within a finite number of steps. We give a refined convergence analysis that improves the dependence on $\epsilon$ from polynomial to logarithmic and simultaneously allows to choose a step size that is independent of $\epsilon$. Furthermore, we show that the dynamics can be initialized with a more general set of (infeasible) starting points. %K Computer Science, Data Structures and Algorithms, cs.DS,Mathematics, Dynamical Systems, math.DS,Mathematics, Optimization and Control, math.OC, Physics, Biological Physics, physics.bio-ph
[38]
X. Bei, J. Garg, M. Hoefer, and K. Mehlhorn, “Earning Limits in Fisher Markets with Spending-Constraint Utilities,” in Algorithmic Game Theory (SAGT 2017), L’Aquila, Italy, 2017.
Export
BibTeX
@inproceedings{BeiSAGT2017, TITLE = {Earning Limits in {Fisher} Markets with Spending-Constraint Utilities}, AUTHOR = {Bei, Xiaohui and Garg, Jugal and Hoefer, Martin and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISBN = {978-3-319-66699-0}, DOI = {10.1007/978-3-319-66700-3_6}, PUBLISHER = {Springer}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Algorithmic Game Theory (SAGT 2017)}, PAGES = {67--79}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10504}, ADDRESS = {L'Aquila, Italy}, }
Endnote
%0 Conference Proceedings %A Bei, Xiaohui %A Garg, Jugal %A Hoefer, Martin %A Mehlhorn, Kurt %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Earning Limits in Fisher Markets with Spending-Constraint Utilities : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-E7DB-7 %R 10.1007/978-3-319-66700-3_6 %D 2017 %B 10th International Symposium on Algorithmic Game Theory %Z date of event: 2017-09-12 - 2017-09-14 %C L'Aquila, Italy %B Algorithmic Game Theory %P 67 - 79 %I Springer %@ 978-3-319-66699-0 %B Lecture Notes in Computer Science %N 10504
[39]
F. Benhamouda, T. Lepoint, C. Mathieu, and H. Zhou, “Optimization of Bootstrapping in Circuits,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{doi:10.1137/1.9781611974782.160, TITLE = {Optimization of Bootstrapping in Circuits}, AUTHOR = {Benhamouda, Fabrice and Lepoint, Tancr{\`e}de and Mathieu, Claire and Zhou, Hang}, LANGUAGE = {eng}, ISBN = {978-1-61197-478-2}, DOI = {10.1137/1.9781611974782.160}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, EDITOR = {Klein, Philip N.}, PAGES = {2423--2433}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Benhamouda, Fabrice %A Lepoint, Tancr&#232;de %A Mathieu, Claire %A Zhou, Hang %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Optimization of Bootstrapping in Circuits : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4EBE-A %R 10.1137/1.9781611974782.160 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %E Klein, Philip N. %P 2423 - 2433 %I SIAM %@ 978-1-61197-478-2
[40]
P. Berenbrink, A. Clementi, R. Elsässer, P. Kling, F. Mallmann-Trenn, and E. Natale, “Ignore or Comply?: On Breaking Symmetry in Consensus,” in PODC’17, ACM Symposium on Principles of Distributed Computing, Washington, DC, USA, 2017.
Export
BibTeX
@inproceedings{Berenbrink:2017:ICB:3087801.3087817, TITLE = {Ignore or Comply?: {O}n Breaking Symmetry in Consensus}, AUTHOR = {Berenbrink, Petra and Clementi, Andrea and Els{\"a}sser, Robert and Kling, Peter and Mallmann-Trenn, Frederik and Natale, Emanuele}, LANGUAGE = {eng}, ISBN = {978-1-4503-4992-5}, DOI = {10.1145/3087801.3087817}, PUBLISHER = {ACM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {PODC'17, ACM Symposium on Principles of Distributed Computing}, PAGES = {335--344}, ADDRESS = {Washington, DC, USA}, }
Endnote
%0 Conference Proceedings %A Berenbrink, Petra %A Clementi, Andrea %A Els&#228;sser, Robert %A Kling, Peter %A Mallmann-Trenn, Frederik %A Natale, Emanuele %+ External Organizations External Organizations External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Ignore or Comply?: On Breaking Symmetry in Consensus : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-76B2-0 %R 10.1145/3087801.3087817 %D 2017 %B ACM Symposium on Principles of Distributed Computing %Z date of event: 2017-07-25 - 2017-07-27 %C Washington, DC, USA %B PODC'17 %P 335 - 344 %I ACM %@ 978-1-4503-4992-5
[41]
O. Beyersdorff, L. Chew, and K. Sreenivasaiah, “A Game Characterisation of Tree-like Q-Resolution Size,” Journal of Computer and System Sciences, vol. In Press, 2017.
Export
BibTeX
@article{Beyersdorff2017, TITLE = {A Game Characterisation of Tree-like {Q-Resolution} Size}, AUTHOR = {Beyersdorff, Olaf and Chew, Leroy and Sreenivasaiah, Karteek}, LANGUAGE = {eng}, ISSN = {0022-0000}, DOI = {10.1016/j.jcss.2016.11.011}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, JOURNAL = {Journal of Computer and System Sciences}, VOLUME = {In Press}, }
Endnote
%0 Journal Article %A Beyersdorff, Olaf %A Chew, Leroy %A Sreenivasaiah, Karteek %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Game Characterisation of Tree-like Q-Resolution Size : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5F80-F %R 10.1016/j.jcss.2016.11.011 %7 2017 %D 2017 %J Journal of Computer and System Sciences %V In Press %I Elsevier %C Amsterdam %@ false
[42]
L. Boczkowski, A. Korman, and E. Natale, “Minimizing Message Size in Stochastic Communication Patterns: Fast Self-Stabilizing Protocols with 3 bits,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{BKN17, TITLE = {Minimizing Message Size in Stochastic Communication Patterns: {F}ast Self-Stabilizing Protocols with 3 bits}, AUTHOR = {Boczkowski, Lucas and Korman, Amos and Natale, Emanuele}, LANGUAGE = {eng}, DOI = {10.1137/1.9781611974782.168}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, PAGES = {2540--2559}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Boczkowski, Lucas %A Korman, Amos %A Natale, Emanuele %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Minimizing Message Size in Stochastic Communication Patterns: Fast Self-Stabilizing Protocols with 3 bits : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-587B-2 %R 10.1137/1.9781611974782.168 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %P 2540 - 2559 %I SIAM
[43]
J.-D. Boissonnat, R. Dyer, and A. Ghosh, “Delaunay Triangulation of Manifolds,” Foundations of Computational Mathematics, vol. First Online, 2017.
Export
BibTeX
@article{Boissonnat2017, TITLE = {Delaunay Triangulation of Manifolds}, AUTHOR = {Boissonnat, Jean-Daniel and Dyer, Ramsay and Ghosh, Arijit}, LANGUAGE = {eng}, ISSN = {1615-3375}, DOI = {10.1007/s10208-017-9344-1}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, JOURNAL = {Foundations of Computational Mathematics}, VOLUME = {First Online}, PAGES = {1--33}, }
Endnote
%0 Journal Article %A Boissonnat, Jean-Daniel %A Dyer, Ramsay %A Ghosh, Arijit %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Delaunay Triangulation of Manifolds : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-7945-0 %R 10.1007/s10208-017-9344-1 %7 2017-02-01 %D 2017 %8 01.02.2017 %J Foundations of Computational Mathematics %V First Online %& 1 %P 1 - 33 %I Springer %C New York, NY %@ false
[44]
K. Bringmann, “A Near-Linear Pseudopolynomial Time Algorithm for Subset Sum,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{DBLP:conf/soda/Bringmann17, TITLE = {A Near-Linear Pseudopolynomial Time Algorithm for Subset Sum}, AUTHOR = {Bringmann, Karl}, LANGUAGE = {eng}, DOI = {10.1137/1.9781611974782.69}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, PAGES = {1073--1084}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Near-Linear Pseudopolynomial Time Algorithm for Subset Sum : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5522-4 %R 10.1137/1.9781611974782.69 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %P 1073 - 1084 %I SIAM
[45]
K. Bringmann, P. Gawrychowski, S. Mozes, and O. Weimann, “Tree Edit Distance Cannot be Computed in Strongly Subcubic Time (unless APSP can),” 2017. [Online]. Available: http://arxiv.org/abs/1703.08940. (arXiv: 1703.08940)
Abstract
The edit distance between two rooted ordered trees with $n$ nodes labeled from an alphabet~$\Sigma$ is the minimum cost of transforming one tree into the other by a sequence of elementary operations consisting of deleting and relabeling existing nodes, as well as inserting new nodes. Tree edit distance is a well known generalization of string edit distance. The fastest known algorithm for tree edit distance runs in cubic $O(n^3)$ time and is based on a similar dynamic programming solution as string edit distance. In this paper we show that a truly subcubic $O(n^{3-\varepsilon})$ time algorithm for tree edit distance is unlikely: For $|\Sigma| = \Omega(n)$, a truly subcubic algorithm for tree edit distance implies a truly subcubic algorithm for the all pairs shortest paths problem. For $|\Sigma| = O(1)$, a truly subcubic algorithm for tree edit distance implies an $O(n^{k-\varepsilon})$ algorithm for finding a maximum weight $k$-clique. Thus, while in terms of upper bounds string edit distance and tree edit distance are highly related, in terms of lower bounds string edit distance exhibits the hardness of the strong exponential time hypothesis [Backurs, Indyk STOC'15] whereas tree edit distance exhibits the hardness of all pairs shortest paths. Our result provides a matching conditional lower bound for one of the last remaining classic dynamic programming problems.
Export
BibTeX
@online{DBLP:journals/corr/BringmannGMW17, TITLE = {Tree Edit Distance Cannot be Computed in Strongly Subcubic Time (unless {APSP} can)}, AUTHOR = {Bringmann, Karl and Gawrychowski, Pawe{\l} and Mozes, Shay and Weimann, Oren}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1703.08940}, EPRINT = {1703.08940}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {The edit distance between two rooted ordered trees with $n$ nodes labeled from an alphabet~$\Sigma$ is the minimum cost of transforming one tree into the other by a sequence of elementary operations consisting of deleting and relabeling existing nodes, as well as inserting new nodes. Tree edit distance is a well known generalization of string edit distance. The fastest known algorithm for tree edit distance runs in cubic $O(n^3)$ time and is based on a similar dynamic programming solution as string edit distance. In this paper we show that a truly subcubic $O(n^{3-\varepsilon})$ time algorithm for tree edit distance is unlikely: For $|\Sigma| = \Omega(n)$, a truly subcubic algorithm for tree edit distance implies a truly subcubic algorithm for the all pairs shortest paths problem. For $|\Sigma| = O(1)$, a truly subcubic algorithm for tree edit distance implies an $O(n^{k-\varepsilon})$ algorithm for finding a maximum weight $k$-clique. Thus, while in terms of upper bounds string edit distance and tree edit distance are highly related, in terms of lower bounds string edit distance exhibits the hardness of the strong exponential time hypothesis [Backurs, Indyk STOC'15] whereas tree edit distance exhibits the hardness of all pairs shortest paths. Our result provides a matching conditional lower bound for one of the last remaining classic dynamic programming problems.}, }
Endnote
%0 Report %A Bringmann, Karl %A Gawrychowski, Pawe&#322; %A Mozes, Shay %A Weimann, Oren %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Tree Edit Distance Cannot be Computed in Strongly Subcubic Time (unless APSP can) : %O Tree Edit Distance Cannot be Computed in Strongly Subcubic Time (unless {APSP} can) %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-8A70-3 %U http://arxiv.org/abs/1703.08940 %D 2017 %X The edit distance between two rooted ordered trees with $n$ nodes labeled from an alphabet~$\Sigma$ is the minimum cost of transforming one tree into the other by a sequence of elementary operations consisting of deleting and relabeling existing nodes, as well as inserting new nodes. Tree edit distance is a well known generalization of string edit distance. The fastest known algorithm for tree edit distance runs in cubic $O(n^3)$ time and is based on a similar dynamic programming solution as string edit distance. In this paper we show that a truly subcubic $O(n^{3-\varepsilon})$ time algorithm for tree edit distance is unlikely: For $|\Sigma| = \Omega(n)$, a truly subcubic algorithm for tree edit distance implies a truly subcubic algorithm for the all pairs shortest paths problem. For $|\Sigma| = O(1)$, a truly subcubic algorithm for tree edit distance implies an $O(n^{k-\varepsilon})$ algorithm for finding a maximum weight $k$-clique. Thus, while in terms of upper bounds string edit distance and tree edit distance are highly related, in terms of lower bounds string edit distance exhibits the hardness of the strong exponential time hypothesis [Backurs, Indyk STOC'15] whereas tree edit distance exhibits the hardness of all pairs shortest paths. Our result provides a matching conditional lower bound for one of the last remaining classic dynamic programming problems. %K Computer Science, Data Structures and Algorithms, cs.DS
[46]
K. Bringmann, A. Gronlund, and K. G. Larsen, “A Dichotomy for Regular Expression Membership Testing,” in 58th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2017), Berkeley, CA, USA, 2017.
Export
BibTeX
@inproceedings{Bringman_FOCS2017, TITLE = {A Dichotomy for Regular Expression Membership Testing}, AUTHOR = {Bringmann, Karl and Gronlund, Allan and Larsen, Kasper Green}, LANGUAGE = {eng}, ISBN = {978-1-5386-3464-6}, DOI = {10.1109/FOCS.2017.36}, PUBLISHER = {IEEE}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {58th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2017)}, PAGES = {307--318}, ADDRESS = {Berkeley, CA, USA}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Gronlund, Allan %A Larsen, Kasper Green %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T A Dichotomy for Regular Expression Membership Testing : %G eng %U http://hdl.handle.net/21.11116/0000-0000-0471-C %R 10.1109/FOCS.2017.36 %D 2017 %B 58th Annual IEEE Symposium on Foundations of Computer Science %Z date of event: 2017-10-15 - 2017-10-17 %C Berkeley, CA, USA %B 58th Annual IEEE Symposium on Foundations of Computer Science %P 307 - 318 %I IEEE %@ 978-1-5386-3464-6
[47]
K. Bringmann, T. Dueholm Hansen, and S. Krinninger, “Improved Algorithms for Computing the Cycle of Minimum Cost-to-Time Ratio in Directed Graphs,” in 44th International Colloquium on Automata, Languages, and Programming (ICALP 2017), Warsaw, Poland, 2017.
Export
BibTeX
@inproceedings{BringmannICALP2017, TITLE = {Improved Algorithms for Computing the Cycle of Minimum Cost-to-Time Ratio in Directed Graphs}, AUTHOR = {Bringmann, Karl and Dueholm Hansen, Thomas and Krinninger, Sebastian}, LANGUAGE = {eng}, ISBN = {978-3-95977-041-5}, URL = {urn:nbn:de:0030-drops-74398}, DOI = {10.4230/LIPIcs.ICALP.2017.124}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {44th International Colloquium on Automata, Languages, and Programming (ICALP 2017)}, EDITOR = {Chatzigiannakis, Ioannis and Indyk, Piotr and Kuhn, Fabian and Muscholl, Anca}, PAGES = {1--16}, EID = {124}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {80}, ADDRESS = {Warsaw, Poland}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Dueholm Hansen, Thomas %A Krinninger, Sebastian %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Improved Algorithms for Computing the Cycle of Minimum Cost-to-Time Ratio in Directed Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-89C4-F %R 10.4230/LIPIcs.ICALP.2017.124 %U urn:nbn:de:0030-drops-74398 %D 2017 %B 44th International Colloquium on Automata, Languages, and Programming %Z date of event: 2017-07-10 - 2017-07-14 %C Warsaw, Poland %B 44th International Colloquium on Automata, Languages, and Programming %E Chatzigiannakis, Ioannis; Indyk, Piotr; Kuhn, Fabian; Muscholl, Anca %P 1 - 16 %Z sequence number: 124 %I Schloss Dagstuhl %@ 978-3-95977-041-5 %B Leibniz International Proceedings in Informatics %N 80 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2017/7439/
[48]
K. Bringmann, S. Cabello, and M. Emmerich, “Maximum Volume Subset Selection for Anchored Boxes,” in 33rd International Symposium on Computational Geometry (SoCG 2017), Brisbane, Australia, 2017.
Export
BibTeX
@inproceedings{bringmann:scg, TITLE = {Maximum Volume Subset Selection for Anchored Boxes}, AUTHOR = {Bringmann, Karl and Cabello, Sergio and Emmerich, Michael}, LANGUAGE = {eng}, ISBN = {978-3-95977-038-5}, URL = {urn:nbn:de:0030-drops-72011}, DOI = {10.4230/LIPIcs.SoCG.2017.22}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {33rd International Symposium on Computational Geometry (SoCG 2017)}, EDITOR = {Aranov, Boris and Katz, Matthew J.}, PAGES = {1--15}, EID = {22}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {77}, ADDRESS = {Brisbane, Australia}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Cabello, Sergio %A Emmerich, Michael %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Maximum Volume Subset Selection for Anchored Boxes : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-7D6F-9 %U urn:nbn:de:0030-drops-72011 %R 10.4230/LIPIcs.SoCG.2017.22 %D 2017 %B 33rd International Symposium on Computational Geometry %Z date of event: 2017-07-04 - 2017-07-07 %C Brisbane, Australia %B 33rd International Symposium on Computational Geometry %E Aranov, Boris; Katz, Matthew J. %P 1 - 15 %Z sequence number: 22 %I Schloss Dagstuhl %@ 978-3-95977-038-5 %B Leibniz International Proceedings in Informatics %N 77 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2017/7201/
[49]
K. Bringmann, T. Dueholm Hansen, and S. Krinninger, “Improved Algorithms for Computing the Cycle of Minimum Cost-to-Time Ratio in Directed Graphs,” 2017. [Online]. Available: http://arxiv.org/abs/1704.08122. (arXiv: 1704.08122)
Abstract
We study the problem of finding the cycle of minimum cost-to-time ratio in a directed graph with $ n $ nodes and $ m $ edges. This problem has a long history in combinatorial optimization and has recently seen interesting applications in the context of quantitative verification. We focus on strongly polynomial algorithms to cover the use-case where the weights are relatively large compared to the size of the graph. Our main result is an algorithm with running time $ \tilde O (m^{3/4} n^{3/2}) $, which gives the first improvement over Megiddo's $ \tilde O (n^3) $ algorithm [JACM'83] for sparse graphs. We further demonstrate how to obtain both an algorithm with running time $ n^3 / 2^{\Omega{(\sqrt{\log n})}} $ on general graphs and an algorithm with running time $ \tilde O (n) $ on constant treewidth graphs. To obtain our main result, we develop a parallel algorithm for negative cycle detection and single-source shortest paths that might be of independent interest.
Export
BibTeX
@online{DBLP:journals/corr/BringmannHK17, TITLE = {Improved Algorithms for Computing the Cycle of Minimum Cost-to-Time Ratio in Directed Graphs}, AUTHOR = {Bringmann, Karl and Dueholm Hansen, Thomas and Krinninger, Sebastian}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1704.08122}, EPRINT = {1704.08122}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We study the problem of finding the cycle of minimum cost-to-time ratio in a directed graph with $ n $ nodes and $ m $ edges. This problem has a long history in combinatorial optimization and has recently seen interesting applications in the context of quantitative verification. We focus on strongly polynomial algorithms to cover the use-case where the weights are relatively large compared to the size of the graph. Our main result is an algorithm with running time $ \tilde O (m^{3/4} n^{3/2}) $, which gives the first improvement over Megiddo's $ \tilde O (n^3) $ algorithm [JACM'83] for sparse graphs. We further demonstrate how to obtain both an algorithm with running time $ n^3 / 2^{\Omega{(\sqrt{\log n})}} $ on general graphs and an algorithm with running time $ \tilde O (n) $ on constant treewidth graphs. To obtain our main result, we develop a parallel algorithm for negative cycle detection and single-source shortest paths that might be of independent interest.}, }
Endnote
%0 Report %A Bringmann, Karl %A Dueholm Hansen, Thomas %A Krinninger, Sebastian %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Improved Algorithms for Computing the Cycle of Minimum Cost-to-Time Ratio in Directed Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-89BC-3 %U http://arxiv.org/abs/1704.08122 %D 2017 %X We study the problem of finding the cycle of minimum cost-to-time ratio in a directed graph with $ n $ nodes and $ m $ edges. This problem has a long history in combinatorial optimization and has recently seen interesting applications in the context of quantitative verification. We focus on strongly polynomial algorithms to cover the use-case where the weights are relatively large compared to the size of the graph. Our main result is an algorithm with running time $ \tilde O (m^{3/4} n^{3/2}) $, which gives the first improvement over Megiddo's $ \tilde O (n^3) $ algorithm [JACM'83] for sparse graphs. We further demonstrate how to obtain both an algorithm with running time $ n^3 / 2^{\Omega{(\sqrt{\log n})}} $ on general graphs and an algorithm with running time $ \tilde O (n) $ on constant treewidth graphs. To obtain our main result, we develop a parallel algorithm for negative cycle detection and single-source shortest paths that might be of independent interest. %K Computer Science, Data Structures and Algorithms, cs.DS
[50]
K. Bringmann and P. Wellnitz, “Clique-Based Lower Bounds for Parsing Tree-Adjoining Grammars,” in 28th Annual Symposium on Combinatorial Pattern Matching (CPM 2017), Warsaw, Poland, 2017.
Export
BibTeX
@inproceedings{BringmannCPM2017, TITLE = {Clique-Based Lower Bounds for Parsing Tree-Adjoining Grammars}, AUTHOR = {Bringmann, Karl and Wellnitz, Philip}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-039-2}, URL = {urn:nbn:de:0030-drops-73329}, DOI = {10.4230/LIPIcs.CPM.2017.12}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {28th Annual Symposium on Combinatorial Pattern Matching (CPM 2017)}, EDITOR = {K{\"a}rkk{\"a}inen, Juha and Radoszweski, Jakub and Rytter, Wojciech}, PAGES = {1--14}, EID = {12}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {78}, ADDRESS = {Warsaw, Poland}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Wellnitz, Philip %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Clique-Based Lower Bounds for Parsing Tree-Adjoining Grammars : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-FB92-0 %R 10.4230/LIPIcs.CPM.2017.12 %U urn:nbn:de:0030-drops-73329 %D 2017 %B 28th Annual Symposium on Combinatorial Pattern Matching %Z date of event: 2017-07-04 - 2017-07-06 %C Warsaw, Poland %B 28th Annual Symposium on Combinatorial Pattern Matching %E K&#228;rkk&#228;inen, Juha; Radoszweski, Jakub; Rytter, Wojciech %P 1 - 14 %Z sequence number: 12 %I Schloss Dagstuhl %@ 978-3-95977-039-2 %B Leibniz International Proceedings in Informatics %N 78 %@ false %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2017/7332/
[51]
K. Bringmann and K. Panagiotou, “Efficient Sampling Methods for Discrete Distributions,” Algorithmica, vol. 79, no. 2, 2017.
Export
BibTeX
@article{BringmannAlgorithmica2016, TITLE = {Efficient Sampling Methods for Discrete Distributions}, AUTHOR = {Bringmann, Karl and Panagiotou, Konstantinos}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-016-0205-0}, PUBLISHER = {Springer-Verlag}, ADDRESS = {New York}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Algorithmica}, VOLUME = {79}, NUMBER = {2}, PAGES = {484--508}, }
Endnote
%0 Journal Article %A Bringmann, Karl %A Panagiotou, Konstantinos %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Efficient Sampling Methods for Discrete Distributions : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-85D0-8 %R 10.1007/s00453-016-0205-0 %7 2016-08-29 %D 2017 %J Algorithmica %V 79 %N 2 %& 484 %P 484 - 508 %I Springer-Verlag %C New York %@ false
[52]
K. Bringmann and S. Krinninger, “A Note on Hardness of Diameter Approximation,” 2017. [Online]. Available: http://arxiv.org/abs/1705.02127. (arXiv: 1705.02127)
Abstract
We revisit the hardness of approximating the diameter of a network. In the CONGEST model, $ \tilde \Omega (n) $ rounds are necessary to compute the diameter [Frischknecht et al. SODA'12]. Abboud et al. DISC 2016 extended this result to sparse graphs and, at a more fine-grained level, showed that, for any integer $ 1 \leq \ell \leq \operatorname{polylog} (n) $, distinguishing between networks of diameter $ 4 \ell + 2 $ and $ 6 \ell + 1 $ requires $ \tilde \Omega (n) $ rounds. We slightly tighten this result by showing that even distinguishing between diameter $ 2 \ell + 1 $ and $ 3 \ell + 1 $ requires $ \tilde \Omega (n) $ rounds. The reduction of Abboud et al. is inspired by recent conditional lower bounds in the RAM model, where the orthogonal vectors problem plays a pivotal role. In our new lower bound, we make the connection to orthogonal vectors explicit, leading to a conceptually more streamlined exposition. This is suited for teaching both the lower bound in the CONGEST model and the conditional lower bound in the RAM model.
Export
BibTeX
@online{DBLP:journals/corr/BringmannK17, TITLE = {A Note on Hardness of Diameter Approximation}, AUTHOR = {Bringmann, Karl and Krinninger, Sebastian}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1705.02127}, EPRINT = {1705.02127}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We revisit the hardness of approximating the diameter of a network. In the CONGEST model, $ \tilde \Omega (n) $ rounds are necessary to compute the diameter [Frischknecht et al. SODA'12]. Abboud et al. DISC 2016 extended this result to sparse graphs and, at a more fine-grained level, showed that, for any integer $ 1 \leq \ell \leq \operatorname{polylog} (n) $, distinguishing between networks of diameter $ 4 \ell + 2 $ and $ 6 \ell + 1 $ requires $ \tilde \Omega (n) $ rounds. We slightly tighten this result by showing that even distinguishing between diameter $ 2 \ell + 1 $ and $ 3 \ell + 1 $ requires $ \tilde \Omega (n) $ rounds. The reduction of Abboud et al. is inspired by recent conditional lower bounds in the RAM model, where the orthogonal vectors problem plays a pivotal role. In our new lower bound, we make the connection to orthogonal vectors explicit, leading to a conceptually more streamlined exposition. This is suited for teaching both the lower bound in the CONGEST model and the conditional lower bound in the RAM model.}, }
Endnote
%0 Report %A Bringmann, Karl %A Krinninger, Sebastian %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T A Note on Hardness of Diameter Approximation : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-89B7-D %U http://arxiv.org/abs/1705.02127 %D 2017 %X We revisit the hardness of approximating the diameter of a network. In the CONGEST model, $ \tilde \Omega (n) $ rounds are necessary to compute the diameter [Frischknecht et al. SODA'12]. Abboud et al. DISC 2016 extended this result to sparse graphs and, at a more fine-grained level, showed that, for any integer $ 1 \leq \ell \leq \operatorname{polylog} (n) $, distinguishing between networks of diameter $ 4 \ell + 2 $ and $ 6 \ell + 1 $ requires $ \tilde \Omega (n) $ rounds. We slightly tighten this result by showing that even distinguishing between diameter $ 2 \ell + 1 $ and $ 3 \ell + 1 $ requires $ \tilde \Omega (n) $ rounds. The reduction of Abboud et al. is inspired by recent conditional lower bounds in the RAM model, where the orthogonal vectors problem plays a pivotal role. In our new lower bound, we make the connection to orthogonal vectors explicit, leading to a conceptually more streamlined exposition. This is suited for teaching both the lower bound in the CONGEST model and the conditional lower bound in the RAM model. %K Computer Science, Data Structures and Algorithms, cs.DS,Computer Science, Distributed, Parallel, and Cluster Computing, cs.DC
[53]
K. Bringmann, C. Ikenmeyer, and J. Zuiddam, “On Algebraic Branching Programs of Small Width,” Electronic Colloquium on Computational Complexity (ECCC) : Report Series, vol. 34 (Revision 1), 2017.
Export
BibTeX
@article{BringmannECCC2017, TITLE = {On Algebraic Branching Programs of Small Width}, AUTHOR = {Bringmann, Karl and Ikenmeyer, Christian and Zuiddam, Jeroen}, LANGUAGE = {eng}, ISSN = {1433-8092}, PUBLISHER = {Hasso-Plattner-Institut f{\"u}r Softwaretechnik GmbH}, ADDRESS = {Potsdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, JOURNAL = {Electronic Colloquium on Computational Complexity (ECCC) : Report Series}, VOLUME = {34 (Revision 1)}, PAGES = {1--30}, }
Endnote
%0 Journal Article %A Bringmann, Karl %A Ikenmeyer, Christian %A Zuiddam, Jeroen %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On Algebraic Branching Programs of Small Width : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-89B2-8 %7 2017 %D 2017 %J Electronic Colloquium on Computational Complexity (ECCC) : Report Series %V 34 (Revision 1) %& 1 %P 1 - 30 %I Hasso-Plattner-Institut f&#252;r Softwaretechnik GmbH %C Potsdam %@ false %U https://eccc.weizmann.ac.il/report/2017/034/
[54]
K. Bringmann, R. Keusch, and J. Lengler, “Sampling Geometric Inhomogeneous Random Graphs in Linear Time,” in 25th Annual European Symposium on Algorithms (ESA 2017), Vienna, Austria, 2017.
Export
BibTeX
@inproceedings{BringmannESA2017, TITLE = {Sampling Geometric Inhomogeneous Random Graphs in Linear Time}, AUTHOR = {Bringmann, Karl and Keusch, Ralph and Lengler, Johannes}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-049-1}, URL = {urn:nbn:de:0030-drops-78396}, DOI = {10.4230/LIPIcs.ESA.2017.20}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {25th Annual European Symposium on Algorithms (ESA 2017)}, EDITOR = {Pruhs, Kirk and Sohler, Christian}, PAGES = {1--15}, EID = {20}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {87}, ADDRESS = {Vienna, Austria}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Keusch, Ralph %A Lengler, Johannes %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Sampling Geometric Inhomogeneous Random Graphs in Linear Time : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-FB87-A %R 10.4230/LIPIcs.ESA.2017.20 %U urn:nbn:de:0030-drops-78396 %D 2017 %B 25th Annual European Symposium on Algorithms %Z date of event: 2017-09-04 - 2017-09-06 %C Vienna, Austria %B 25th Annual European Symposium on Algorithms %E Pruhs, Kirk; Sohler, Christian %P 1 - 15 %Z sequence number: 20 %I Schloss Dagstuhl %@ 978-3-95977-049-1 %B Leibniz International Proceedings in Informatics %N 87 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2017/7839/http://drops.dagstuhl.de/doku/urheberrecht1.html
[55]
K. Bringmann, C. Ikenmeyer, and J. Zuiddam, “On Algebraic Branching Programs of Small Width,” in 32nd Computational Complexity Conference (CCC 2017), Riga, Latvia, 2017.
Export
BibTeX
@inproceedings{BringmannCCC2017, TITLE = {On Algebraic Branching Programs of Small Width}, AUTHOR = {Bringmann, Karl and Ikenmeyer, Christian and Zuiddam, Jeroen}, LANGUAGE = {eng}, ISBN = {978-3-95977-040-8}, URL = {urn:nbn:de:0030-drops-75217}, DOI = {10.4230/LIPIcs.CCC.2017.20}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {32nd Computational Complexity Conference (CCC 2017)}, EDITOR = {O'Donnell, Ryan}, PAGES = {1--31}, EID = {20}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {79}, ADDRESS = {Riga, Latvia}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Ikenmeyer, Christian %A Zuiddam, Jeroen %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On Algebraic Branching Programs of Small Width : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-FB78-C %R 10.4230/LIPIcs.CCC.2017.20 %U urn:nbn:de:0030-drops-75217 %D 2017 %B 32nd Computational Complexity Conference %Z date of event: 2017-07-06 - 2017-07-09 %C Riga, Latvia %B 32nd Computational Complexity Conference %E O'Donnell, Ryan %P 1 - 31 %Z sequence number: 20 %I Schloss Dagstuhl %@ 978-3-95977-040-8 %B Leibniz International Proceedings in Informatics %N 79 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2017/7521/
[56]
K. Bringmann, C. Ikenmeyer, and J. Zuiddam, “On Algebraic Branching Programs of Small Width,” 2017. [Online]. Available: http://arxiv.org/abs/1702.05328. (arXiv: 1702.05328)
Abstract
In 1979 Valiant showed that the complexity class VP_e of families with polynomially bounded formula size is contained in the class VP_s of families that have algebraic branching programs (ABPs) of polynomially bounded size. Motivated by the problem of separating these classes we study the topological closure VP_e-bar, i.e. the class of polynomials that can be approximated arbitrarily closely by polynomials in VP_e. We describe VP_e-bar with a strikingly simple complete polynomial (in characteristic different from 2) whose recursive definition is similar to the Fibonacci numbers. Further understanding this polynomial seems to be a promising route to new formula lower bounds. Our methods are rooted in the study of ABPs of small constant width. In 1992 Ben-Or and Cleve showed that formula size is polynomially equivalent to width-3 ABP size. We extend their result (in characteristic different from 2) by showing that approximate formula size is polynomially equivalent to approximate width-2 ABP size. This is surprising because in 2011 Allender and Wang gave explicit polynomials that cannot be computed by width-2 ABPs at all! The details of our construction lead to the aforementioned characterization of VP_e-bar. As a natural continuation of this work we prove that the class VNP can be described as the class of families that admit a hypercube summation of polynomially bounded dimension over a product of polynomially many affine linear forms. This gives the first separations of algebraic complexity classes from their nondeterministic analogs.
Export
BibTeX
@online{BringmannArXiv2017, TITLE = {On Algebraic Branching Programs of Small Width}, AUTHOR = {Bringmann, Karl and Ikenmeyer, Christian and Zuiddam, Jeroen}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1702.05328}, EPRINT = {1702.05328}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {In 1979 Valiant showed that the complexity class VP_e of families with polynomially bounded formula size is contained in the class VP_s of families that have algebraic branching programs (ABPs) of polynomially bounded size. Motivated by the problem of separating these classes we study the topological closure VP_e-bar, i.e. the class of polynomials that can be approximated arbitrarily closely by polynomials in VP_e. We describe VP_e-bar with a strikingly simple complete polynomial (in characteristic different from 2) whose recursive definition is similar to the Fibonacci numbers. Further understanding this polynomial seems to be a promising route to new formula lower bounds. Our methods are rooted in the study of ABPs of small constant width. In 1992 Ben-Or and Cleve showed that formula size is polynomially equivalent to width-3 ABP size. We extend their result (in characteristic different from 2) by showing that approximate formula size is polynomially equivalent to approximate width-2 ABP size. This is surprising because in 2011 Allender and Wang gave explicit polynomials that cannot be computed by width-2 ABPs at all! The details of our construction lead to the aforementioned characterization of VP_e-bar. As a natural continuation of this work we prove that the class VNP can be described as the class of families that admit a hypercube summation of polynomially bounded dimension over a product of polynomially many affine linear forms. This gives the first separations of algebraic complexity classes from their nondeterministic analogs.}, }
Endnote
%0 Report %A Bringmann, Karl %A Ikenmeyer, Christian %A Zuiddam, Jeroen %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On Algebraic Branching Programs of Small Width : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-89A4-8 %U http://arxiv.org/abs/1702.05328 %D 2017 %X In 1979 Valiant showed that the complexity class VP_e of families with polynomially bounded formula size is contained in the class VP_s of families that have algebraic branching programs (ABPs) of polynomially bounded size. Motivated by the problem of separating these classes we study the topological closure VP_e-bar, i.e. the class of polynomials that can be approximated arbitrarily closely by polynomials in VP_e. We describe VP_e-bar with a strikingly simple complete polynomial (in characteristic different from 2) whose recursive definition is similar to the Fibonacci numbers. Further understanding this polynomial seems to be a promising route to new formula lower bounds. Our methods are rooted in the study of ABPs of small constant width. In 1992 Ben-Or and Cleve showed that formula size is polynomially equivalent to width-3 ABP size. We extend their result (in characteristic different from 2) by showing that approximate formula size is polynomially equivalent to approximate width-2 ABP size. This is surprising because in 2011 Allender and Wang gave explicit polynomials that cannot be computed by width-2 ABPs at all! The details of our construction lead to the aforementioned characterization of VP_e-bar. As a natural continuation of this work we prove that the class VNP can be described as the class of families that admit a hypercube summation of polynomially bounded dimension over a product of polynomially many affine linear forms. This gives the first separations of algebraic complexity classes from their nondeterministic analogs. %K Computer Science, Computational Complexity, cs.CC,
[57]
J. Bund, C. Lenzen, and M. Medina, “Near-Optimal Metastability-Containing Sorting Networks,” in Proceedings of the 2017 Design, Automation & Test in Europe (DATE 2017), Lausanne, Switzerland, 2017.
Export
BibTeX
@inproceedings{BundDATE2017, TITLE = {Near-Optimal Metastability-Containing Sorting Networks}, AUTHOR = {Bund, Johannes and Lenzen, Christoph and Medina, Moti}, LANGUAGE = {eng}, ISBN = {978-1-5090-5826-6}, DOI = {10.23919/DATE.2017.7926987}, PUBLISHER = {IEEE}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the 2017 Design, Automation \& Test in Europe (DATE 2017)}, PAGES = {226--231}, ADDRESS = {Lausanne, Switzerland}, }
Endnote
%0 Conference Proceedings %A Bund, Johannes %A Lenzen, Christoph %A Medina, Moti %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Near-Optimal Metastability-Containing Sorting Networks : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-571A-2 %R 10.23919/DATE.2017.7926987 %D 2017 %B Design, Automation & Test in Europe Conference & Exhibition %Z date of event: 2017-03-27 - 2017-03-31 %C Lausanne, Switzerland %B Proceedings of the 2017 Design, Automation & Test in Europe %P 226 - 231 %I IEEE %@ 978-1-5090-5826-6
[58]
P. Bürgisser, C. Ikenmeyer, and J. Hüttenhain, “Permanent versus Determinant: Not via Saturations,” Proceedings of the American Mathematical Society, vol. 145, 2017.
Export
BibTeX
@article{BHI:17, TITLE = {Permanent versus Determinant: {N}ot via Saturations}, AUTHOR = {B{\"u}rgisser, Peter and Ikenmeyer, Christian and H{\"u}ttenhain, Jesko}, LANGUAGE = {eng}, ISSN = {0002-9939}, DOI = {10.1090/proc/13310}, PUBLISHER = {American Mathematical Society}, ADDRESS = {Providence, R.I. [etc.]}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Proceedings of the American Mathematical Society}, VOLUME = {145}, PAGES = {1247--1258}, }
Endnote
%0 Journal Article %A B&#252;rgisser, Peter %A Ikenmeyer, Christian %A H&#252;ttenhain, Jesko %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Permanent versus Determinant: Not via Saturations : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4F48-A %R 10.1090/proc/13310 %7 2017 %D 2017 %J Proceedings of the American Mathematical Society %V 145 %& 1247 %P 1247 - 1258 %I American Mathematical Society %C Providence, R.I. [etc.] %@ false
[59]
P. Bürgisser and C. Ikenmeyer, “Fundamental Invariants of Orbit Closures,” Journal of Algebra, vol. 477, 2017.
Export
BibTeX
@article{BI:17, TITLE = {Fundamental Invariants of Orbit Closures}, AUTHOR = {B{\"u}rgisser, Peter and Ikenmeyer, Christian}, LANGUAGE = {eng}, ISSN = {0021-8693}, DOI = {10.1016/j.jalgebra.2016.12.035}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Journal of Algebra}, VOLUME = {477}, PAGES = {390--434}, }
Endnote
%0 Journal Article %A B&#252;rgisser, Peter %A Ikenmeyer, Christian %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fundamental Invariants of Orbit Closures : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4F2E-8 %R 10.1016/j.jalgebra.2016.12.035 %7 2017-01-17 %D 2017 %J Journal of Algebra %V 477 %& 390 %P 390 - 434 %I Elsevier %C Amsterdam %@ false
[60]
P. Chalermsook and A. Schmid, “Finding Triangles for Maximum Planar Subgraphs,” in WALCOM: Algorithms and Computation, Hsinchu, Taiwan, 2017.
Export
BibTeX
@inproceedings{PCAS2017, TITLE = {Finding Triangles for Maximum Planar Subgraphs}, AUTHOR = {Chalermsook, Parinya and Schmid, Andreas}, LANGUAGE = {eng}, ISBN = {978-3-319-53924-9}, DOI = {10.1007/978-3-319-53925-6_29}, PUBLISHER = {Springer}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {WALCOM: Algorithms and Computation}, EDITOR = {Poon, Sheung-Hung and Rahman, Md. Saidur and Yen, Hsu-Chun}, PAGES = {373--384}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10167}, ADDRESS = {Hsinchu, Taiwan}, }
Endnote
%0 Conference Proceedings %A Chalermsook, Parinya %A Schmid, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Finding Triangles for Maximum Planar Subgraphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5DDC-F %R 10.1007/978-3-319-53925-6_29 %D 2017 %B 11th International Conference and Workshops on Algorithms and Computation %Z date of event: 2017-03-29 - 2017-03-31 %C Hsinchu, Taiwan %B WALCOM: Algorithms and Computation %E Poon, Sheung-Hung; Rahman, Md. Saidur; Yen, Hsu-Chun %P 373 - 384 %I Springer %@ 978-3-319-53924-9 %B Lecture Notes in Computer Science %N 10167
[61]
P. Chalermsook and D. Vaz, “New Integrality Gap Results for the Firefighters Problem on Trees,” in Approximation and Online Algorithms (WAOA 2016), Aarhus, Denmark, 2017.
Export
BibTeX
@inproceedings{Chalermsook2017, TITLE = {New Integrality Gap Results for the Firefighters Problem on Trees}, AUTHOR = {Chalermsook, Parinya and Vaz, Daniel}, LANGUAGE = {eng}, ISBN = {978-3-319-51740-7}, DOI = {10.1007/978-3-319-51741-4_6}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Approximation and Online Algorithms (WAOA 2016)}, EDITOR = {Jansen, Klaus and Mastrolilli, Monaldo}, PAGES = {65--77}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10138}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Chalermsook, Parinya %A Vaz, Daniel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T New Integrality Gap Results for the Firefighters Problem on Trees : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-575B-0 %R 10.1007/978-3-319-51741-4_6 %D 2017 %B 14th Workshop on Approximation and Online Algorithms %Z date of event: 2016-08-25 - 2016-08-26 %C Aarhus, Denmark %B Approximation and Online Algorithms %E Jansen, Klaus; Mastrolilli, Monaldo %P 65 - 77 %I Springer %@ 978-3-319-51740-7 %B Lecture Notes in Computer Science %N 10138
[62]
P. Chalermsook, S. Das, B. Laekhanukit, and D. Vaz, “Beyond Metric Embedding: Approximating Group Steiner Trees on Bounded Treewidth Graphs,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{doi:10.1137/1.9781611974782.47, TITLE = {Beyond Metric Embedding: {A}pproximating {Group Steiner Trees} on Bounded Treewidth Graphs}, AUTHOR = {Chalermsook, Parinya and Das, Syamantak and Laekhanukit, Bundit and Vaz, Daniel}, LANGUAGE = {eng}, DOI = {10.1137/1.9781611974782.47}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, PAGES = {737--751}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Chalermsook, Parinya %A Das, Syamantak %A Laekhanukit, Bundit %A Vaz, Daniel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Beyond Metric Embedding: Approximating Group Steiner Trees on Bounded Treewidth Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-573D-3 %R 10.1137/1.9781611974782.47 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %P 737 - 751 %I SIAM
[63]
L. S. Chandran, D. Issac, and A. Karrenbauer, “On the Parameterized Complexity of Biclique Cover and Partition,” in 11th International Symposium on Parameterized and Exact Computation (IPEC 2016), Aarhus, Denmark, 2017.
Export
BibTeX
@inproceedings{BicliqueFPT, TITLE = {On the Parameterized Complexity of Biclique Cover and Partition}, AUTHOR = {Chandran, L. Sunil and Issac, Davis and Karrenbauer, Andreas}, LANGUAGE = {eng}, ISBN = {978-3-95977-023-1}, URL = {urn:nbn:de:0030-drops-69293}, DOI = {10.4230/LIPIcs.IPEC.2016.11}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {11th International Symposium on Parameterized and Exact Computation (IPEC 2016)}, EDITOR = {Guo, Jiong and Hermelin, Danny}, PAGES = {1--13}, EID = {11}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {63}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Chandran, L. Sunil %A Issac, Davis %A Karrenbauer, Andreas %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T On the Parameterized Complexity of Biclique Cover and Partition : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-53DB-3 %R 10.4230/LIPIcs.IPEC.2016.11 %U urn:nbn:de:0030-drops-69293 %D 2017 %B 11th International Symposium on Parameterized and Exact Computation %Z date of event: 2016-08-24 - 2016-08-26 %C Aarhus, Denmark %B 11th International Symposium on Parameterized and Exact Computation %E Guo, Jiong; Hermelin, Danny %P 1 - 13 %Z sequence number: 11 %I Schloss Dagstuhl %@ 978-3-95977-023-1 %B Leibniz International Proceedings in Informatics %N 63 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2017/6929/
[64]
A. Choudhary, “Approximation Algorithms for Vietoris-Rips and Ĉech Filtrations,” Universität des Saarlandes, Saarbrücken, 2017.
Abstract
Persistent Homology is a tool to analyze and visualize the shape of data from a topological viewpoint. It computes persistence, which summarizes the evolution of topological and geometric information about metric spaces over multiple scales of distances. While computing persistence is quite efficient for low-dimensional topological features, it becomes overwhelmingly expensive for medium to high-dimensional features. In this thesis, we attack this computational problem from several different angles. We present efficient techniques to approximate the persistence of metric spaces. Three of our methods are tailored towards general point clouds in Euclidean spaces. We make use of high dimensional lattice geometry to reduce the cost of the approximations. In particular, we discover several properties of the Permutahedral lattice, whose Voronoi cell is well-known for its combinatorial properties. The last method is suitable for point clouds with low intrinsic dimension, where we exploit the structural properties of the point set to tame the complexity. In some cases, we achieve a reduction in size complexity by trading off the quality of the approximation. Two of our methods work particularly well in conjunction with dimension-reduction techniques: we arrive at the first approximation schemes whose complexities are only polynomial in the size of the point cloud, and independent of the ambient dimension. On the other hand, we provide a lower bound result: we construct a point cloud that requires super-polynomial complexity for a high-quality approximation of the persistence. Together with our approximation schemes, we show that polynomial complexity is achievable for rough approximations, but impossible for sufficiently fine approximations. For some metric spaces, the intrinsic dimension is low in small neighborhoods of the input points, but much higher for large scales of distances. We develop a concept of local intrinsic dimension to capture this property. We also present several applications of this concept, including an approximation method for persistence. This thesis is written in English.
Export
BibTeX
@phdthesis{Choudharyphd2017, TITLE = {Approximation Algorithms for {V}ietoris-Rips and \v{C}ech Filtrations}, AUTHOR = {Choudhary, Aruni}, LANGUAGE = {eng}, URL = {urn:nbn:de:bsz:291-scidok-ds-269597}, DOI = {10.22028/D291-26959}, SCHOOL = {Universit{\"a}t des Saarlandes}, ADDRESS = {Saarbr{\"u}cken}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Persistent Homology is a tool to analyze and visualize the shape of data from a topological viewpoint. It computes persistence, which summarizes the evolution of topological and geometric information about metric spaces over multiple scales of distances. While computing persistence is quite efficient for low-dimensional topological features, it becomes overwhelmingly expensive for medium to high-dimensional features. In this thesis, we attack this computational problem from several different angles. We present efficient techniques to approximate the persistence of metric spaces. Three of our methods are tailored towards general point clouds in Euclidean spaces. We make use of high dimensional lattice geometry to reduce the cost of the approximations. In particular, we discover several properties of the Permutahedral lattice, whose Voronoi cell is well-known for its combinatorial properties. The last method is suitable for point clouds with low intrinsic dimension, where we exploit the structural properties of the point set to tame the complexity. In some cases, we achieve a reduction in size complexity by trading off the quality of the approximation. Two of our methods work particularly well in conjunction with dimension-reduction techniques: we arrive at the first approximation schemes whose complexities are only polynomial in the size of the point cloud, and independent of the ambient dimension. On the other hand, we provide a lower bound result: we construct a point cloud that requires super-polynomial complexity for a high-quality approximation of the persistence. Together with our approximation schemes, we show that polynomial complexity is achievable for rough approximations, but impossible for sufficiently fine approximations. For some metric spaces, the intrinsic dimension is low in small neighborhoods of the input points, but much higher for large scales of distances. We develop a concept of local intrinsic dimension to capture this property. We also present several applications of this concept, including an approximation method for persistence. This thesis is written in English.}, }
Endnote
%0 Thesis %A Choudhary, Aruni %A referee: Kerber, Michael %Y Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society International Max Planck Research School, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Approximation Algorithms for Vietoris-Rips and &#264;ech Filtrations : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-8D63-5 %U urn:nbn:de:bsz:291-scidok-ds-269597 %R 10.22028/D291-26959 %I Universit&#228;t des Saarlandes %C Saarbr&#252;cken %D 2017 %P 123 p. %V phd %9 phd %X Persistent Homology is a tool to analyze and visualize the shape of data from a topological viewpoint. It computes persistence, which summarizes the evolution of topological and geometric information about metric spaces over multiple scales of distances. While computing persistence is quite efficient for low-dimensional topological features, it becomes overwhelmingly expensive for medium to high-dimensional features. In this thesis, we attack this computational problem from several different angles. We present efficient techniques to approximate the persistence of metric spaces. Three of our methods are tailored towards general point clouds in Euclidean spaces. We make use of high dimensional lattice geometry to reduce the cost of the approximations. In particular, we discover several properties of the Permutahedral lattice, whose Voronoi cell is well-known for its combinatorial properties. The last method is suitable for point clouds with low intrinsic dimension, where we exploit the structural properties of the point set to tame the complexity. In some cases, we achieve a reduction in size complexity by trading off the quality of the approximation. Two of our methods work particularly well in conjunction with dimension-reduction techniques: we arrive at the first approximation schemes whose complexities are only polynomial in the size of the point cloud, and independent of the ambient dimension. On the other hand, we provide a lower bound result: we construct a point cloud that requires super-polynomial complexity for a high-quality approximation of the persistence. Together with our approximation schemes, we show that polynomial complexity is achievable for rough approximations, but impossible for sufficiently fine approximations. For some metric spaces, the intrinsic dimension is low in small neighborhoods of the input points, but much higher for large scales of distances. We develop a concept of local intrinsic dimension to capture this property. We also present several applications of this concept, including an approximation method for persistence. This thesis is written in English. %U https://publikationen.sulb.uni-saarland.de/handle/20.500.11880/26911
[65]
C. Croitoru, “Graph models for rational social interaction,” Universität des Saarlandes, Saarbrücken, 2017.
Export
BibTeX
@phdthesis{CroitoruPhd2017, TITLE = {Graph models for rational social interaction}, AUTHOR = {Croitoru, Cosmina}, LANGUAGE = {eng}, SCHOOL = {Universit{\"a}t des Saarlandes}, ADDRESS = {Saarbr{\"u}cken}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, }
Endnote
%0 Thesis %A Croitoru, Cosmina %Y Mehlhorn, Kurt %A referee: Amgoud, Leila %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society International Max Planck Research School, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Graph models for rational social interaction : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-87DE-5 %I Universit&#228;t des Saarlandes %C Saarbr&#252;cken %D 2017 %P X, 75 p. %V phd %9 phd
[66]
M. Cygan, M. Pilipczuk, M. Pilipczuk, E. J. van Leeuwen, and M. Wrochna, “Polynomial Kernelization for Removing Induced Claws and Diamonds,” Theory of Computing Systems, vol. 60, no. 4, 2017.
Export
BibTeX
@article{CyganAlgorithmica2016, TITLE = {Polynomial Kernelization for Removing Induced Claws and Diamonds}, AUTHOR = {Cygan, Marek and Pilipczuk, Marcin and Pilipczuk, Micha{\l} and van Leeuwen, Erik Jan and Wrochna, Marcin}, LANGUAGE = {eng}, ISSN = {1432-4350}, DOI = {10.1007/s00224-016-9689-x}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Theory of Computing Systems}, VOLUME = {60}, NUMBER = {4}, PAGES = {615--636}, }
Endnote
%0 Journal Article %A Cygan, Marek %A Pilipczuk, Marcin %A Pilipczuk, Micha&#322; %A van Leeuwen, Erik Jan %A Wrochna, Marcin %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Polynomial Kernelization for Removing Induced Claws and Diamonds : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-8606-6 %R 10.1007/s00224-016-9689-x %7 2016-06-23 %D 2017 %J Theory of Computing Systems %V 60 %N 4 %& 615 %P 615 - 636 %I Springer %C New York, NY %@ false
[67]
J. Diaz, O. Pottonen, M. Serna, and E. J. van Leeuwen, “Complexity of Metric Dimension on Planar Graphs,” Journal of Computer and System Sciences, vol. 83, no. 1, 2017.
Export
BibTeX
@article{Diaz2017, TITLE = {Complexity of Metric Dimension on Planar Graphs}, AUTHOR = {Diaz, Josep and Pottonen, Olli and Serna, Maria and van Leeuwen, Erik Jan}, ISSN = {0022-0000}, DOI = {10.1016/j.jcss.2016.06.006}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Journal of Computer and System Sciences}, VOLUME = {83}, NUMBER = {1}, PAGES = {132--158}, }
Endnote
%0 Journal Article %A Diaz, Josep %A Pottonen, Olli %A Serna, Maria %A van Leeuwen, Erik Jan %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Complexity of Metric Dimension on Planar Graphs : %U http://hdl.handle.net/11858/00-001M-0000-002B-A574-5 %R 10.1016/j.jcss.2016.06.006 %7 2016 %D 2017 %J Journal of Computer and System Sciences %V 83 %N 1 %& 132 %P 132 - 158 %I Elsevier %C Amsterdam %@ false
[68]
M. Dirnberger, K. Mehlhorn, M. Grube, and H.-G. Döbereiner, “Preliminaries for Distributed Natural Computing Inspired by the Slime Mold Physarum Polycephalum,” Universität des Saarlandes, Saarbrücken, 2017.
Abstract
This doctoral thesis aims towards distributed natural computing inspired by the slime mold Physarum polycephalum. The vein networks formed by this organism presumably support efficient transport of protoplasmic fluid. Devising models which capture the natural efficiency of the organism and form a suitable basis for the development of natural computing algorithms is an interesting and challenging goal. We start working towards this goal by designing and executing wet-lab experi- ments geared towards producing a large number of images of the vein networks of P. polycephalum. Next, we turn the depicted vein networks into graphs using our own custom software called Nefi. This enables a detailed numerical study, yielding a catalogue of characterizing observables spanning a wide array of different graph properties. To share our results and data, i.e. raw experimental data, graphs and analysis results, we introduce a dedicated repository revolving around slime mold data, the Smgr. The purpose of this repository is to promote data reuse and to foster a practice of increased data sharing. Finally we present a model based on interacting electronic circuits including current controlled voltage sources, which mimics the emergent flow patterns observed in live P. polycephalum. The model is simple, distributed and robust to changes in the underlying network topology. Thus it constitutes a promising basis for the development of distributed natural computing algorithms.
Export
BibTeX
@phdthesis{dirnbergerphd17, TITLE = {Preliminaries for Distributed Natural Computing Inspired by the Slime Mold Physarum Polycephalum}, AUTHOR = {Dirnberger, Michael and Mehlhorn, Kurt and Grube, Martin and D{\"o}bereiner, Hans-G{\"u}nther}, LANGUAGE = {eng}, URL = {urn:nbn:de:bsz:291-scidok-69424}, SCHOOL = {Universit{\"a}t des Saarlandes}, ADDRESS = {Saarbr{\"u}cken}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, ABSTRACT = {This doctoral thesis aims towards distributed natural computing inspired by the slime mold Physarum polycephalum. The vein networks formed by this organism presumably support efficient transport of protoplasmic fluid. Devising models which capture the natural efficiency of the organism and form a suitable basis for the development of natural computing algorithms is an interesting and challenging goal. We start working towards this goal by designing and executing wet-lab experi- ments geared towards producing a large number of images of the vein networks of P. polycephalum. Next, we turn the depicted vein networks into graphs using our own custom software called Nefi. This enables a detailed numerical study, yielding a catalogue of characterizing observables spanning a wide array of different graph properties. To share our results and data, i.e. raw experimental data, graphs and analysis results, we introduce a dedicated repository revolving around slime mold data, the Smgr. The purpose of this repository is to promote data reuse and to foster a practice of increased data sharing. Finally we present a model based on interacting electronic circuits including current controlled voltage sources, which mimics the emergent flow patterns observed in live P. polycephalum. The model is simple, distributed and robust to changes in the underlying network topology. Thus it constitutes a promising basis for the development of distributed natural computing algorithms.}, }
Endnote
%0 Thesis %A Dirnberger, Michael %A Mehlhorn, Kurt %A Grube, Martin %A D&#246;bereiner, Hans-G&#252;nther %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society International Max Planck Research School, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Preliminaries for Distributed Natural Computing Inspired by the Slime Mold Physarum Polycephalum : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-DE4F-0 %U urn:nbn:de:bsz:291-scidok-69424 %I Universit&#228;t des Saarlandes %C Saarbr&#252;cken %D 2017 %P XV, 193 p. %V phd %9 phd %X This doctoral thesis aims towards distributed natural computing inspired by the slime mold Physarum polycephalum. The vein networks formed by this organism presumably support efficient transport of protoplasmic fluid. Devising models which capture the natural efficiency of the organism and form a suitable basis for the development of natural computing algorithms is an interesting and challenging goal. We start working towards this goal by designing and executing wet-lab experi- ments geared towards producing a large number of images of the vein networks of P. polycephalum. Next, we turn the depicted vein networks into graphs using our own custom software called Nefi. This enables a detailed numerical study, yielding a catalogue of characterizing observables spanning a wide array of different graph properties. To share our results and data, i.e. raw experimental data, graphs and analysis results, we introduce a dedicated repository revolving around slime mold data, the Smgr. The purpose of this repository is to promote data reuse and to foster a practice of increased data sharing. Finally we present a model based on interacting electronic circuits including current controlled voltage sources, which mimics the emergent flow patterns observed in live P. polycephalum. The model is simple, distributed and robust to changes in the underlying network topology. Thus it constitutes a promising basis for the development of distributed natural computing algorithms. %U http://scidok.sulb.uni-saarland.de/doku/lic_ohne_pod.php?la=dehttp://scidok.sulb.uni-saarland.de/volltexte/2017/6942/
[69]
M. Dirnberger, K. Mehlhorn, and T. Mehlhorn, “Introducing the Slime Mold Graph Repository,” Journal of Physics D: Applied Physics, vol. 50, no. 26, 2017.
Export
BibTeX
@article{Dirnberger2017, TITLE = {Introducing the Slime Mold Graph Repository}, AUTHOR = {Dirnberger, Michael and Mehlhorn, Kurt and Mehlhorn, Tim}, LANGUAGE = {eng}, ISSN = {0022-3727}, DOI = {10.1088/1361-6463/aa7326}, PUBLISHER = {IOP Publishing}, ADDRESS = {Bristol}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Journal of Physics D: Applied Physics}, VOLUME = {50}, NUMBER = {26}, EID = {264001}, }
Endnote
%0 Journal Article %A Dirnberger, Michael %A Mehlhorn, Kurt %A Mehlhorn, Tim %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Introducing the Slime Mold Graph Repository : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-8464-B %R 10.1088/1361-6463/aa7326 %7 2017 %D 2017 %J Journal of Physics D: Applied Physics %O J. Phys. D: Appl. Phys. %V 50 %N 26 %Z sequence number: 264001 %I IOP Publishing %C Bristol %@ false
[70]
M. Dirnberger and K. Mehlhorn, “Characterizing Networks Formed by P. Polycephalum,” Journal of Physics D: Applied Physics, vol. 50, no. 22, 2017.
Export
BibTeX
@article{Dirnberg_Mehlhorn2017, TITLE = {Characterizing networks formed by \slshape{P. polycephalum}}, AUTHOR = {Dirnberger, Michael and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISSN = {0022-3727}, DOI = {10.1088/1361-6463/aa6e7b}, PUBLISHER = {IOP Publishing}, ADDRESS = {Bristol}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Journal of Physics D: Applied Physics}, VOLUME = {50}, NUMBER = {22}, EID = {224002}, }
Endnote
%0 Journal Article %A Dirnberger, Michael %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Characterizing Networks Formed by P. Polycephalum : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-56FA-2 %R 10.1088/1361-6463/aa6e7b %7 2017 %D 2017 %J Journal of Physics D: Applied Physics %O J. Phys. D: Appl. Phys. %V 50 %N 22 %Z sequence number: 224002 %I IOP Publishing %C Bristol %@ false
[71]
K. Dutta, A. Ghosh, B. Jartoux, and N. H. Mustafa, “Shallow Packings, Semialgebraic Set Systems, Macbeath Regions and Polynomial Partitioning,” in 33rd International Symposium on Computational Geometry (SoCG 2017), Brisbane, Australia, 2017.
Export
BibTeX
@inproceedings{DuttaGJM-Mnets-17, TITLE = {Shallow Packings, Semialgebraic Set Systems, {Macbeath} Regions and Polynomial Partitioning}, AUTHOR = {Dutta, Kunal and Ghosh, Arijit and Jartoux, Bruno and Mustafa, Nabil H.}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-038-5}, URL = {urn:nbn:de:0030-drops-71991}, DOI = {10.4230/LIPIcs.SoCG.2017.38}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {33rd International Symposium on Computational Geometry (SoCG 2017)}, EDITOR = {Aronov, Boris and Katz, Matthew J.}, PAGES = {1--15}, EID = {38}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {77}, ADDRESS = {Brisbane, Australia}, }
Endnote
%0 Conference Proceedings %A Dutta, Kunal %A Ghosh, Arijit %A Jartoux, Bruno %A Mustafa, Nabil H. %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Shallow Packings, Semialgebraic Set Systems, Macbeath Regions and Polynomial Partitioning : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-7941-7 %R 10.4230/LIPIcs.SoCG.2017.38 %U urn:nbn:de:0030-drops-71991 %D 2017 %B 33rd International Symposium on Computational Geometry %Z date of event: 2017-07-04 - 2017-07-07 %C Brisbane, Australia %B 33rd International Symposium on Computational Geometry %E Aronov, Boris; Katz, Matthew J. %P 1 - 15 %Z sequence number: 38 %I Schloss Dagstuhl %@ 978-3-95977-038-5 %B Leibniz International Proceedings in Informatics %N 77 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2017/7199/http://drops.dagstuhl.de/doku/urheberrecht1.html
[72]
P. Dütting and T. Kesselheim, “Best-Response Dynamics in Combinatorial Auctions with Item Bidding,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{doi:10.1137/1.9781611974782.33, TITLE = {Best-Response Dynamics in Combinatorial Auctions with Item Bidding}, AUTHOR = {D{\"u}tting, Paul and Kesselheim, Thomas}, LANGUAGE = {eng}, ISBN = {978-1-61197-478-2}, DOI = {10.1137/1.9781611974782.33}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, EDITOR = {Klein, Philip N.}, PAGES = {521--533}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A D&#252;tting, Paul %A Kesselheim, Thomas %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Best-Response Dynamics in Combinatorial Auctions with Item Bidding : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E5E-2 %R 10.1137/1.9781611974782.33 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %E Klein, Philip N. %P 521 - 533 %I SIAM %@ 978-1-61197-478-2
[73]
M. Ernestus, S. Friedrichs, M. Hemmer, J. Kokemüller, A. Kröller, M. Moeini, and C. Schmidt, “Algorithms for Art Gallery Illumination,” Journal of Global Optimization, vol. 68, no. 1, 2017.
Export
BibTeX
@article{ErnestusJGO2016, TITLE = {Algorithms for Art Gallery Illumination}, AUTHOR = {Ernestus, Maximilian and Friedrichs, Stephan and Hemmer, Michael and Kokem{\"u}ller, Jan and Kr{\"o}ller, Alexander and Moeini, Mahdi and Schmidt, Christiane}, LANGUAGE = {eng}, ISSN = {0925-5001}, DOI = {10.1007/s10898-016-0452-2}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Journal of Global Optimization}, VOLUME = {68}, NUMBER = {1}, PAGES = {23--45}, }
Endnote
%0 Journal Article %A Ernestus, Maximilian %A Friedrichs, Stephan %A Hemmer, Michael %A Kokem&#252;ller, Jan %A Kr&#246;ller, Alexander %A Moeini, Mahdi %A Schmidt, Christiane %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations External Organizations External Organizations %T Algorithms for Art Gallery Illumination : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-0B3C-1 %R 10.1007/s10898-016-0452-2 %7 2016 %D 2017 %J Journal of Global Optimization %V 68 %N 1 %& 23 %P 23 - 45 %I Springer %C New York, NY %@ false
[74]
G. Even and M. Medina, “Online Packet-Routing in Grids with Bounded Buffers,” Algorithmica, vol. 78, no. 3, 2017.
Export
BibTeX
@article{MedinaAlgorithmica2016, TITLE = {Online Packet-Routing in Grids with Bounded Buffers}, AUTHOR = {Even, Guy and Medina, Moti}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-016-0177-0}, PUBLISHER = {Springer-Verlag}, ADDRESS = {New York}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Algorithmica}, VOLUME = {78}, NUMBER = {3}, PAGES = {819--868}, }
Endnote
%0 Journal Article %A Even, Guy %A Medina, Moti %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Online Packet-Routing in Grids with Bounded Buffers : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-85DD-E %R 10.1007/s00453-016-0177-0 %7 2016-07-11 %D 2017 %J Algorithmica %V 78 %N 3 %& 819 %P 819 - 868 %I Springer-Verlag %C New York %@ false
[75]
K. Fleszar, M. Mnich, and J. Spoerhase, “New Algorithms for Maximum Disjoint Paths Based on Tree-likeness,” Mathematical Programming / A, 2017.
Export
BibTeX
@article{edge-disjoint-paths-mapr-17, TITLE = {New Algorithms for Maximum Disjoint Paths Based on Tree-likeness}, AUTHOR = {Fleszar, Krzysztof and Mnich, Matthias and Spoerhase, Joachim}, LANGUAGE = {eng}, ISSN = {0025-5610}, DOI = {10.1007/s10107-017-1199-3}, PUBLISHER = {North-Holland}, ADDRESS = {Heidelberg}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, JOURNAL = {Mathematical Programming / A}, PAGES = {1--29}, }
Endnote
%0 Journal Article %A Fleszar, Krzysztof %A Mnich, Matthias %A Spoerhase, Joachim %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T New Algorithms for Maximum Disjoint Paths Based on Tree-likeness : %G eng %U http://hdl.handle.net/21.11116/0000-0000-B54C-F %R 10.1007/s10107-017-1199-3 %7 2017 %D 2017 %J Mathematical Programming / A %& 1 %P 1 - 29 %I North-Holland %C Heidelberg %@ false
[76]
S. Friedrichs, C. Lenzen, K. Mehlhorn, and M. Ghaffari, “Metastability-Containing Circuits, Parallel Distance Problems, and Terrain Guarding,” Unversität des Saarlandes, Saarbrücken, 2017.
Abstract
We study three problems. The first is the phenomenon of metastability in digital circuits. This is a state of bistable storage elements, such as registers, that is neither logical 0 nor 1 and breaks the abstraction of Boolean logic. We propose a time- and value-discrete model for metastability in digital circuits and show that it reflects relevant physical properties. Further, we propose the fundamentally new approach of using logical masking to perform meaningful computations despite the presence of metastable upsets and analyze what functions can be computed in our model. Additionally, we show that circuits with masking registers grow computationally more powerful with each available clock cycle. The second topic are parallel algorithms, based on an algebraic abstraction of the Moore-Bellman-Ford algorithm, for solving various distance problems. Our focus are distance approximations that obey the triangle inequality while at the same time achieving polylogarithmic depth and low work. Finally, we study the continuous Terrain Guarding Problem. We show that it has a rational discretization with a quadratic number of guard candidates, establish its membership in NP and the existence of a PTAS, and present an efficient implementation of a solver.
Export
BibTeX
@phdthesis{Friedrichsphd2017, TITLE = {Metastability-Containing Circuits, Parallel Distance Problems, and Terrain Guarding}, AUTHOR = {Friedrichs, Stephan and Lenzen, Christoph and Mehlhorn, Kurt and Ghaffari, Mohsen}, LANGUAGE = {eng}, URL = {urn:nbn:de:bsz:291-scidok-69660}, SCHOOL = {Unversit{\"a}t des Saarlandes}, ADDRESS = {Saarbr{\"u}cken}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, ABSTRACT = {We study three problems. The first is the phenomenon of metastability in digital circuits. This is a state of bistable storage elements, such as registers, that is neither logical 0 nor 1 and breaks the abstraction of Boolean logic. We propose a time- and value-discrete model for metastability in digital circuits and show that it reflects relevant physical properties. Further, we propose the fundamentally new approach of using logical masking to perform meaningful computations despite the presence of metastable upsets and analyze what functions can be computed in our model. Additionally, we show that circuits with masking registers grow computationally more powerful with each available clock cycle. The second topic are parallel algorithms, based on an algebraic abstraction of the Moore-Bellman-Ford algorithm, for solving various distance problems. Our focus are distance approximations that obey the triangle inequality while at the same time achieving polylogarithmic depth and low work. Finally, we study the continuous Terrain Guarding Problem. We show that it has a rational discretization with a quadratic number of guard candidates, establish its membership in NP and the existence of a PTAS, and present an efficient implementation of a solver.}, }
Endnote
%0 Thesis %A Friedrichs, Stephan %A Lenzen, Christoph %A Mehlhorn, Kurt %A Ghaffari, Mohsen %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society International Max Planck Research School, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Metastability-Containing Circuits, Parallel Distance Problems, and Terrain Guarding : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-E9A7-B %U urn:nbn:de:bsz:291-scidok-69660 %I Unversit&#228;t des Saarlandes %C Saarbr&#252;cken %D 2017 %P x, 226 p. %V phd %9 phd %X We study three problems. The first is the phenomenon of metastability in digital circuits. This is a state of bistable storage elements, such as registers, that is neither logical 0 nor 1 and breaks the abstraction of Boolean logic. We propose a time- and value-discrete model for metastability in digital circuits and show that it reflects relevant physical properties. Further, we propose the fundamentally new approach of using logical masking to perform meaningful computations despite the presence of metastable upsets and analyze what functions can be computed in our model. Additionally, we show that circuits with masking registers grow computationally more powerful with each available clock cycle. The second topic are parallel algorithms, based on an algebraic abstraction of the Moore-Bellman-Ford algorithm, for solving various distance problems. Our focus are distance approximations that obey the triangle inequality while at the same time achieving polylogarithmic depth and low work. Finally, we study the continuous Terrain Guarding Problem. We show that it has a rational discretization with a quadratic number of guard candidates, establish its membership in NP and the existence of a PTAS, and present an efficient implementation of a solver. %U http://scidok.sulb.uni-saarland.de/volltexte/2017/6966/http://scidok.sulb.uni-saarland.de/doku/lic_ohne_pod.php?la=de
[77]
M. Függer, C. Lenzen, and T. Polzer, “Metastability-Aware Memory-Efficient Time-to-Digital Converters,” in 23rd IEEE International Symposium on Asynchronous Circuits and Systems, San Diego, CA, USA, 2017.
Export
BibTeX
@inproceedings{fueggerASYNC2017, TITLE = {Metastability-Aware Memory-Efficient Time-to-Digital Converters}, AUTHOR = {F{\"u}gger, Matthias and Lenzen, Christoph and Polzer, Thomas}, LANGUAGE = {eng}, ISBN = {978-1-5386-2749-5}, DOI = {10.1109/ASYNC.2017.12}, PUBLISHER = {IEEE}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {23rd IEEE International Symposium on Asynchronous Circuits and Systems}, PAGES = {49--56}, ADDRESS = {San Diego, CA, USA}, }
Endnote
%0 Conference Proceedings %A F&#252;gger, Matthias %A Lenzen, Christoph %A Polzer, Thomas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Metastability-Aware Memory-Efficient Time-to-Digital Converters : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-80CD-D %R 10.1109/ASYNC.2017.12 %D 2017 %B 23rd IEEE International Symposium on Asynchronous Circuits and Systems %Z date of event: 2017-05-21 - 2017-05-24 %C San Diego, CA, USA %B 23rd IEEE International Symposium on Asynchronous Circuits and Systems %P 49 - 56 %I IEEE %@ 978-1-5386-2749-5
[78]
W. Gálvez, F. Grandoni, S. Heydrich, S. Ingala, A. Khan, and A. Wiese, “Approximating Geometric Knapsack via L-Packings,” in 58th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2017), Berkeley, CA, USA, 2017.
Export
BibTeX
@inproceedings{Galvez_FOCS2017, TITLE = {Approximating Geometric Knapsack via {L}-Packings}, AUTHOR = {G{\'a}lvez, Waldo and Grandoni, Fabrizio and Heydrich, Sandy and Ingala, Salvatore and Khan, Arindam and Wiese, Andreas}, LANGUAGE = {eng}, ISBN = {978-1-5386-3464-6}, DOI = {10.1109/FOCS.2017.32}, PUBLISHER = {IEEE}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {58th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2017)}, PAGES = {260--271}, ADDRESS = {Berkeley, CA, USA}, }
Endnote
%0 Conference Proceedings %A G&#225;lvez, Waldo %A Grandoni, Fabrizio %A Heydrich, Sandy %A Ingala, Salvatore %A Khan, Arindam %A Wiese, Andreas %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Approximating Geometric Knapsack via L-Packings : %G eng %U http://hdl.handle.net/21.11116/0000-0000-0469-6 %R 10.1109/FOCS.2017.32 %D 2017 %B 58th Annual IEEE Symposium on Foundations of Computer Science %Z date of event: 2017-10-15 - 2017-10-17 %C Berkeley, CA, USA %B 58th Annual IEEE Symposium on Foundations of Computer Science %P 260 - 271 %I IEEE %@ 978-1-5386-3464-6
[79]
J. Garg, M. Hoefer, and K. Mehlhorn, “Approximating the Nash Social Welfare with Budget-Additive Valuations,” 2017. [Online]. Available: http://arxiv.org/abs/1707.04428. (arXiv: 1707.04428)
Abstract
We present the first constant-factor approximation algorithm for maximizing the Nash social welfare when allocating indivisible items to agents with budget-additive valuation functions. Budget-additive valuations represent an important class of submodular functions. They have attracted a lot of research interest in recent years due to many interesting applications. For every $\varepsilon > 0$, our algorithm obtains a $(2.404 + \varepsilon)$-approximation in time polynomial in the input size and $1/\varepsilon$. Our algorithm relies on rounding an approximate equilibrium in a linear Fisher market where sellers have earning limits (upper bounds on the amount of money they want to earn) and buyers have utility limits (upper bounds on the amount of utility they want to achieve). In contrast to markets with either earning or utility limits, these markets have not been studied before. They turn out to have fundamentally different properties. Although the existence of equilibria is not guaranteed, we show that the market instances arising from the Nash social welfare problem always have an equilibrium. Further, we show that the set of equilibria is not convex, answering a question of [Cole et al, EC 2017]. We design an FPTAS to compute an approximate equilibrium, a result that may be of independent interest.
Export
BibTeX
@online{GargHoeferMehlhorn2017, TITLE = {Approximating the {Nash} Social Welfare with Budget-Additive Valuations}, AUTHOR = {Garg, Jugal and Hoefer, Martin and Mehlhorn, Kurt}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1707.04428}, EPRINT = {1707.04428}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We present the first constant-factor approximation algorithm for maximizing the Nash social welfare when allocating indivisible items to agents with budget-additive valuation functions. Budget-additive valuations represent an important class of submodular functions. They have attracted a lot of research interest in recent years due to many interesting applications. For every $\varepsilon > 0$, our algorithm obtains a $(2.404 + \varepsilon)$-approximation in time polynomial in the input size and $1/\varepsilon$. Our algorithm relies on rounding an approximate equilibrium in a linear Fisher market where sellers have earning limits (upper bounds on the amount of money they want to earn) and buyers have utility limits (upper bounds on the amount of utility they want to achieve). In contrast to markets with either earning or utility limits, these markets have not been studied before. They turn out to have fundamentally different properties. Although the existence of equilibria is not guaranteed, we show that the market instances arising from the Nash social welfare problem always have an equilibrium. Further, we show that the set of equilibria is not convex, answering a question of [Cole et al, EC 2017]. We design an FPTAS to compute an approximate equilibrium, a result that may be of independent interest.}, }
Endnote
%0 Report %A Garg, Jugal %A Hoefer, Martin %A Mehlhorn, Kurt %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Approximating the Nash Social Welfare with Budget-Additive Valuations : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-E7D6-2 %U http://arxiv.org/abs/1707.04428 %D 2017 %X We present the first constant-factor approximation algorithm for maximizing the Nash social welfare when allocating indivisible items to agents with budget-additive valuation functions. Budget-additive valuations represent an important class of submodular functions. They have attracted a lot of research interest in recent years due to many interesting applications. For every $\varepsilon > 0$, our algorithm obtains a $(2.404 + \varepsilon)$-approximation in time polynomial in the input size and $1/\varepsilon$. Our algorithm relies on rounding an approximate equilibrium in a linear Fisher market where sellers have earning limits (upper bounds on the amount of money they want to earn) and buyers have utility limits (upper bounds on the amount of utility they want to achieve). In contrast to markets with either earning or utility limits, these markets have not been studied before. They turn out to have fundamentally different properties. Although the existence of equilibria is not guaranteed, we show that the market instances arising from the Nash social welfare problem always have an equilibrium. Further, we show that the set of equilibria is not convex, answering a question of [Cole et al, EC 2017]. We design an FPTAS to compute an approximate equilibrium, a result that may be of independent interest. %K Computer Science, Data Structures and Algorithms, cs.DS,Computer Science, Computer Science and Game Theory, cs.GT
[80]
F. Gesmundo, C. Ikenmeyer, and G. Panova, “Geometric Complexity Theory and Matrix Powering,” Differential Geometry and its Applications, vol. 55, 2017.
Export
BibTeX
@article{Gesmundo2017, TITLE = {Geometric Complexity Theory and Matrix Powering}, AUTHOR = {Gesmundo, Fulvio and Ikenmeyer, Christian and Panova, Greta}, LANGUAGE = {eng}, ISSN = {0926-2245}, DOI = {10.1016/j.difgeo.2017.07.001}, PUBLISHER = {North-Holland}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Differential Geometry and its Applications}, VOLUME = {55}, PAGES = {106--127}, }
Endnote
%0 Journal Article %A Gesmundo, Fulvio %A Ikenmeyer, Christian %A Panova, Greta %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Geometric Complexity Theory and Matrix Powering : %G eng %U http://hdl.handle.net/21.11116/0000-0000-0453-E %R 10.1016/j.difgeo.2017.07.001 %7 2017 %D 2017 %K Computer Science, Computational Complexity, cs.CC,Mathematics, Representation Theory, math.RT, %J Differential Geometry and its Applications %V 55 %& 106 %P 106 - 127 %I North-Holland %C Amsterdam %@ false
[81]
M. Goswami, X. Gu, V. P. Pingali, and G. Telang, “Computing Teichmüller Maps Between Polygons,” Foundations of Computational Mathematics, vol. 17, no. 2, 2017.
Export
BibTeX
@article{Goswami2017, TITLE = {Computing {Teichm&#252;ller} Maps Between Polygons}, AUTHOR = {Goswami, Mayank and Gu, Xianfeng and Pingali, Vamsi P. and Telang, Gaurish}, LANGUAGE = {eng}, ISSN = {1615-3375}, DOI = {10.1007/s10208-015-9294-4}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Foundations of Computational Mathematics}, VOLUME = {17}, NUMBER = {2}, PAGES = {497--526}, }
Endnote
%0 Journal Article %A Goswami, Mayank %A Gu, Xianfeng %A Pingali, Vamsi P. %A Telang, Gaurish %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Computing Teichm&#252;ller Maps Between Polygons : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-3E48-3 %R 10.1007/s10208-015-9294-4 %7 2015-11-25 %D 2017 %J Foundations of Computational Mathematics %V 17 %N 2 %& 497 %P 497 - 526 %I Springer %C New York, NY %@ false
[82]
M. Goswami, R. Pagh, F. Silvestri, and J. Sivertsen, “Distance Sensitive Bloom Filters Without False Negatives,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{GoswamiSODA2017, TITLE = {Distance Sensitive Bloom Filters Without False Negatives}, AUTHOR = {Goswami, Mayank and Pagh, Rasmus and Silvestri, Francesco and Sivertsen, Johan}, LANGUAGE = {eng}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, PAGES = {257--269}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Goswami, Mayank %A Pagh, Rasmus %A Silvestri, Francesco %A Sivertsen, Johan %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Distance Sensitive Bloom Filters Without False Negatives : %G eng %U http://hdl.handle.net/21.11116/0000-0001-4FA1-1 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %P 257 - 269 %I SIAM
[83]
F. Grandoni, T. Mömke, A. Wiese, and H. Zhou, “To Augment or Not to Augment: Solving Unsplittable Flow on a Path by Creating Slack,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{doi:10.1137/1.9781611974782.159, TITLE = {To Augment or Not to Augment: {S}olving Unsplittable Flow on a Path by Creating Slack}, AUTHOR = {Grandoni, Fabrizio and M{\"o}mke, Tobias and Wiese, Andreas and Zhou, Hang}, LANGUAGE = {eng}, ISBN = {978-1-61197-478-2}, DOI = {10.1137/1.9781611974782.159}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, EDITOR = {Klein, Philip N.}, PAGES = {2411--2422}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Grandoni, Fabrizio %A M&#246;mke, Tobias %A Wiese, Andreas %A Zhou, Hang %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T To Augment or Not to Augment: Solving Unsplittable Flow on a Path by Creating Slack : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4EB5-B %R 10.1137/1.9781611974782.159 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %E Klein, Philip N. %P 2411 - 2422 %I SIAM %@ 978-1-61197-478-2
[84]
S. Heydrich and A. Wiese, “Faster Approximation Schemes for the Two-dimensional Knapsack Problem,” in Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017), Barcelona, Spain, 2017.
Export
BibTeX
@inproceedings{HeydrichW17, TITLE = {Faster Approximation Schemes for the Two-dimensional Knapsack Problem}, AUTHOR = {Heydrich, Sandy and Wiese, Andreas}, LANGUAGE = {eng}, ISBN = {978-1-61197-478-2}, DOI = {10.1137/1.9781611974782.6}, PUBLISHER = {SIAM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2017)}, EDITOR = {Klein, Philip N.}, PAGES = {79--98}, ADDRESS = {Barcelona, Spain}, }
Endnote
%0 Conference Proceedings %A Heydrich, Sandy %A Wiese, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Faster Approximation Schemes for the Two-dimensional Knapsack Problem : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-54AD-3 %R 10.1137/1.9781611974782.6 %D 2017 %B Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2017-01-16 - 2017-01-19 %C Barcelona, Spain %B Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms %E Klein, Philip N. %P 79 - 98 %I SIAM %@ 978-1-61197-478-2
[85]
M. Hoefer and L. Wagner, “Locally Stable Marriage with Strict Preferences,” SIAM Journal on Discrete Mathematics, vol. 31, no. 1, 2017.
Export
BibTeX
@article{HoeferWagner2017, TITLE = {Locally Stable Marriage with Strict Preferences}, AUTHOR = {Hoefer, Martin and Wagner, Lisa}, LANGUAGE = {eng}, ISSN = {0895-4801}, DOI = {10.1137/151003854}, PUBLISHER = {SIAM}, ADDRESS = {Philadelphia, Pa.}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {SIAM Journal on Discrete Mathematics}, VOLUME = {31}, NUMBER = {1}, PAGES = {283--316}, }
Endnote
%0 Journal Article %A Hoefer, Martin %A Wagner, Lisa %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Locally Stable Marriage with Strict Preferences : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-26B4-1 %R 10.1137/151003854 %7 2017-02-23 %D 2017 %J SIAM Journal on Discrete Mathematics %V 31 %N 1 %& 283 %P 283 - 316 %I SIAM %C Philadelphia, Pa. %@ false
[86]
M. Hoefer and B. Kodric, “Combinatorial Secretary Problems with Ordinal Information,” 2017. [Online]. Available: http://arxiv.org/abs/1702.01290. (arXiv: 1702.01290)
Abstract
The secretary problem is a classic model for online decision making. Recently, combinatorial extensions such as matroid or matching secretary problems have become an important tool to study algorithmic problems in dynamic markets. Here the decision maker must know the numerical value of each arriving element, which can be a demanding informational assumption. In this paper, we initiate the study of combinatorial secretary problems with ordinal information, in which the decision maker only needs to be aware of a preference order consistent with the values of arrived elements. The goal is to design online algorithms with small competitive ratios. For a variety of combinatorial problems, such as bipartite matching, general packing LPs, and independent set with bounded local independence number, we design new algorithms that obtain constant competitive ratios. For the matroid secretary problem, we observe that many existing algorithms for special matroid structures maintain their competitive ratios even in the ordinal model. In these cases, the restriction to ordinal information does not represent any additional obstacle. Moreover, we show that ordinal variants of the submodular matroid secretary problems can be solved using algorithms for the linear versions by extending [Feldman and Zenklusen, 2015]. In contrast, we provide a lower bound of $\Omega(\sqrt{n}/(\log n))$ for algorithms that are oblivious to the matroid structure, where $n$ is the total number of elements. This contrasts an upper bound of $O(\log n)$ in the cardinal model, and it shows that the technique of thresholding is not sufficient for good algorithms in the ordinal model.
Export
BibTeX
@online{Hoefer_Kodric2017, TITLE = {Combinatorial Secretary Problems with Ordinal Information}, AUTHOR = {Hoefer, Martin and Kodric, Bojana}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1702.01290}, EPRINT = {1702.01290}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {The secretary problem is a classic model for online decision making. Recently, combinatorial extensions such as matroid or matching secretary problems have become an important tool to study algorithmic problems in dynamic markets. Here the decision maker must know the numerical value of each arriving element, which can be a demanding informational assumption. In this paper, we initiate the study of combinatorial secretary problems with ordinal information, in which the decision maker only needs to be aware of a preference order consistent with the values of arrived elements. The goal is to design online algorithms with small competitive ratios. For a variety of combinatorial problems, such as bipartite matching, general packing LPs, and independent set with bounded local independence number, we design new algorithms that obtain constant competitive ratios. For the matroid secretary problem, we observe that many existing algorithms for special matroid structures maintain their competitive ratios even in the ordinal model. In these cases, the restriction to ordinal information does not represent any additional obstacle. Moreover, we show that ordinal variants of the submodular matroid secretary problems can be solved using algorithms for the linear versions by extending [Feldman and Zenklusen, 2015]. In contrast, we provide a lower bound of $\Omega(\sqrt{n}/(\log n))$ for algorithms that are oblivious to the matroid structure, where $n$ is the total number of elements. This contrasts an upper bound of $O(\log n)$ in the cardinal model, and it shows that the technique of thresholding is not sufficient for good algorithms in the ordinal model.}, }
Endnote
%0 Report %A Hoefer, Martin %A Kodric, Bojana %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Combinatorial Secretary Problems with Ordinal Information : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5C63-3 %U http://arxiv.org/abs/1702.01290 %D 2017 %X The secretary problem is a classic model for online decision making. Recently, combinatorial extensions such as matroid or matching secretary problems have become an important tool to study algorithmic problems in dynamic markets. Here the decision maker must know the numerical value of each arriving element, which can be a demanding informational assumption. In this paper, we initiate the study of combinatorial secretary problems with ordinal information, in which the decision maker only needs to be aware of a preference order consistent with the values of arrived elements. The goal is to design online algorithms with small competitive ratios. For a variety of combinatorial problems, such as bipartite matching, general packing LPs, and independent set with bounded local independence number, we design new algorithms that obtain constant competitive ratios. For the matroid secretary problem, we observe that many existing algorithms for special matroid structures maintain their competitive ratios even in the ordinal model. In these cases, the restriction to ordinal information does not represent any additional obstacle. Moreover, we show that ordinal variants of the submodular matroid secretary problems can be solved using algorithms for the linear versions by extending [Feldman and Zenklusen, 2015]. In contrast, we provide a lower bound of $\Omega(\sqrt{n}/(\log n))$ for algorithms that are oblivious to the matroid structure, where $n$ is the total number of elements. This contrasts an upper bound of $O(\log n)$ in the cardinal model, and it shows that the technique of thresholding is not sufficient for good algorithms in the ordinal model. %K Computer Science, Data Structures and Algorithms, cs.DS
[87]
W. Höhn, J. Mestre, and A. Wiese, “How Unsplittable-flow-covering Helps Scheduling with Job-dependent Cost Functions,” Algorithmica, vol. First Online, 2017.
Export
BibTeX
@article{Hoehn2017, TITLE = {How Unsplittable-flow-covering Helps Scheduling with Job-dependent Cost Functions}, AUTHOR = {H{\"o}hn, Wiebke and Mestre, Julian and Wiese, Andreas}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-017-0300-x}, PUBLISHER = {Springer-Verlag}, ADDRESS = {New York}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, JOURNAL = {Algorithmica}, VOLUME = {First Online}, PAGES = {1--23}, }
Endnote
%0 Journal Article %A H&#246;hn, Wiebke %A Mestre, Julian %A Wiese, Andreas %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T How Unsplittable-flow-covering Helps Scheduling with Job-dependent Cost Functions : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-2618-3 %R 10.1007/s00453-017-0300-x %7 2017 %D 2017 %J Algorithmica %V First Online %& 1 %P 1 - 23 %I Springer-Verlag %C New York %@ false
[88]
C. Ikenmeyer, K. D. Mulmuley, and M. Walter, “On Vanishing of Kronecker Coefficients,” Computational Complexity, vol. 26, no. 4, 2017.
Export
BibTeX
@article{Ikenmeyer2017, TITLE = {On vanishing of {Kronecker} coefficients}, AUTHOR = {Ikenmeyer, Christian and Mulmuley, Ketan D. and Walter, Michael}, LANGUAGE = {eng}, DOI = {10.1007/s00037-017-0158-y}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Computational Complexity}, VOLUME = {26}, NUMBER = {4}, PAGES = {949--992}, }
Endnote
%0 Journal Article %A Ikenmeyer, Christian %A Mulmuley, Ketan D. %A Walter, Michael %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T On Vanishing of Kronecker Coefficients : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-6278-F %R 10.1007/s00037-017-0158-y %7 2017 %D 2017 %J Computational Complexity %V 26 %N 4 %& 949 %P 949 - 992 %I Springer %C New York, NY
[89]
C. Ikenmeyer, B. Komarath, C. Lenzen, V. Lysikov, A. Mokhov, and K. Sreenivasaiah, “On the Complexity of Hazard-free Circuits,” 2017. [Online]. Available: http://arxiv.org/abs/1711.01904. (arXiv: 1711.01904)
Abstract
The problem of constructing hazard-free Boolean circuits dates back to the 1940s and is an important problem in circuit design. Our main lower-bound result unconditionally shows the existence of functions whose circuit complexity is polynomially bounded while every hazard-free implementation is provably of exponential size. Previous lower bounds on the hazard-free complexity were only valid for depth 2 circuits. The same proof method yields that every subcubic implementation of Boolean matrix multiplication must have hazards. These results follow from a crucial structural insight: Hazard-free complexity is a natural generalization of monotone complexity to all (not necessarily monotone) Boolean functions. Thus, we can apply known monotone complexity lower bounds to find lower bounds on the hazard-free complexity. We also lift these methods from the monotone setting to prove exponential hazard-free complexity lower bounds for non-monotone functions. As our main upper-bound result we show how to efficiently convert a Boolean circuit into a bounded-bit hazard-free circuit with only a polynomially large blow-up in the number of gates. Previously, the best known method yielded exponentially large circuits in the worst case, so our algorithm gives an exponential improvement. As a side result we establish the NP-completeness of several hazard detection problems.
Export
BibTeX
@online{Ikenmeyer_Komarath2017, TITLE = {On the Complexity of Hazard-free Circuits}, AUTHOR = {Ikenmeyer, Christian and Komarath, Balagopal and Lenzen, Christoph and Lysikov, Vladimir and Mokhov, Andrey and Sreenivasaiah, Karteek}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1711.01904}, EPRINT = {1711.01904}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {The problem of constructing hazard-free Boolean circuits dates back to the 1940s and is an important problem in circuit design. Our main lower-bound result unconditionally shows the existence of functions whose circuit complexity is polynomially bounded while every hazard-free implementation is provably of exponential size. Previous lower bounds on the hazard-free complexity were only valid for depth 2 circuits. The same proof method yields that every subcubic implementation of Boolean matrix multiplication must have hazards. These results follow from a crucial structural insight: Hazard-free complexity is a natural generalization of monotone complexity to all (not necessarily monotone) Boolean functions. Thus, we can apply known monotone complexity lower bounds to find lower bounds on the hazard-free complexity. We also lift these methods from the monotone setting to prove exponential hazard-free complexity lower bounds for non-monotone functions. As our main upper-bound result we show how to efficiently convert a Boolean circuit into a bounded-bit hazard-free circuit with only a polynomially large blow-up in the number of gates. Previously, the best known method yielded exponentially large circuits in the worst case, so our algorithm gives an exponential improvement. As a side result we establish the NP-completeness of several hazard detection problems.}, }
Endnote
%0 Report %A Ikenmeyer, Christian %A Komarath, Balagopal %A Lenzen, Christoph %A Lysikov, Vladimir %A Mokhov, Andrey %A Sreenivasaiah, Karteek %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T On the Complexity of Hazard-free Circuits : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F22-4 %U http://arxiv.org/abs/1711.01904 %D 2017 %X The problem of constructing hazard-free Boolean circuits dates back to the 1940s and is an important problem in circuit design. Our main lower-bound result unconditionally shows the existence of functions whose circuit complexity is polynomially bounded while every hazard-free implementation is provably of exponential size. Previous lower bounds on the hazard-free complexity were only valid for depth 2 circuits. The same proof method yields that every subcubic implementation of Boolean matrix multiplication must have hazards. These results follow from a crucial structural insight: Hazard-free complexity is a natural generalization of monotone complexity to all (not necessarily monotone) Boolean functions. Thus, we can apply known monotone complexity lower bounds to find lower bounds on the hazard-free complexity. We also lift these methods from the monotone setting to prove exponential hazard-free complexity lower bounds for non-monotone functions. As our main upper-bound result we show how to efficiently convert a Boolean circuit into a bounded-bit hazard-free circuit with only a polynomially large blow-up in the number of gates. Previously, the best known method yielded exponentially large circuits in the worst case, so our algorithm gives an exponential improvement. As a side result we establish the NP-completeness of several hazard detection problems. %K Computer Science, Computational Complexity, cs.CC,
[90]
C. Ikenmeyer and V. Lysikov, “Strassen’s 2x2 Matrix Multiplication Algorithm: A Conceptual Perspective,” 2017. [Online]. Available: http://arxiv.org/abs/1708.08083. (arXiv: 1708.08083)
Abstract
Despite its importance, all proofs of the correctness of Strassen's famous 1969 algorithm to multiply two 2x2 matrices with only seven multiplications involve some more or less tedious calculations such as explicitly multiplying specific 2x2 matrices, expanding expressions to cancel terms with opposing signs, or expanding tensors over the standard basis. This is why the proof is nontrivial to memorize and why many presentations of the proof avoid showing all the details and leave a significant amount of verifications to the reader. In this note we give a short, self-contained, easy to memorize, and elegant proof of the existence of Strassen's algorithm that avoids these types of calculations. We achieve this by focusing on symmetries and algebraic properties. Our proof combines the classical theory of M-pairs, which was initiated by B\"uchi and Clausen in 1985, with recent work on the geometry of Strassen's algorithm by Chiantini, Ikenmeyer, Landsberg, and Ottaviani from 2016.
Export
BibTeX
@online{Ikenmeyer_Lysikov2017, TITLE = {Strassen's 2x2 Matrix Multiplication Algorithm: A Conceptual Perspective}, AUTHOR = {Ikenmeyer, Christian and Lysikov, Vladimir}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1708.08083}, EPRINT = {1708.08083}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Despite its importance, all proofs of the correctness of Strassen's famous 1969 algorithm to multiply two 2x2 matrices with only seven multiplications involve some more or less tedious calculations such as explicitly multiplying specific 2x2 matrices, expanding expressions to cancel terms with opposing signs, or expanding tensors over the standard basis. This is why the proof is nontrivial to memorize and why many presentations of the proof avoid showing all the details and leave a significant amount of verifications to the reader. In this note we give a short, self-contained, easy to memorize, and elegant proof of the existence of Strassen's algorithm that avoids these types of calculations. We achieve this by focusing on symmetries and algebraic properties. Our proof combines the classical theory of M-pairs, which was initiated by B\"uchi and Clausen in 1985, with recent work on the geometry of Strassen's algorithm by Chiantini, Ikenmeyer, Landsberg, and Ottaviani from 2016.}, }
Endnote
%0 Report %A Ikenmeyer, Christian %A Lysikov, Vladimir %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Strassen's 2x2 Matrix Multiplication Algorithm: A Conceptual Perspective : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F1F-9 %U http://arxiv.org/abs/1708.08083 %D 2017 %X Despite its importance, all proofs of the correctness of Strassen's famous 1969 algorithm to multiply two 2x2 matrices with only seven multiplications involve some more or less tedious calculations such as explicitly multiplying specific 2x2 matrices, expanding expressions to cancel terms with opposing signs, or expanding tensors over the standard basis. This is why the proof is nontrivial to memorize and why many presentations of the proof avoid showing all the details and leave a significant amount of verifications to the reader. In this note we give a short, self-contained, easy to memorize, and elegant proof of the existence of Strassen's algorithm that avoids these types of calculations. We achieve this by focusing on symmetries and algebraic properties. Our proof combines the classical theory of M-pairs, which was initiated by B\"uchi and Clausen in 1985, with recent work on the geometry of Strassen's algorithm by Chiantini, Ikenmeyer, Landsberg, and Ottaviani from 2016. %K Computer Science, Data Structures and Algorithms, cs.DS,Computer Science, Symbolic Computation, cs.SC
[91]
C. Ikenmeyer and G. Panova, “Rectangular Kronecker Coefficients and Plethysms in Geometric Complexity Theory,” Advances in Mathematics, vol. 319, 2017.
Export
BibTeX
@article{Ikenmeyer2017, TITLE = {Rectangular {Kronecker} Coefficients and Plethysms in Geometric Complexity Theory}, AUTHOR = {Ikenmeyer, Christian and Panova, Greta}, LANGUAGE = {eng}, ISSN = {0001-8708}, DOI = {10.1016/j.aim.2017.08.024}, PUBLISHER = {Academic Press}, ADDRESS = {Orlando, Fla.}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Advances in Mathematics}, VOLUME = {319}, PAGES = {40--66}, }
Endnote
%0 Journal Article %A Ikenmeyer, Christian %A Panova, Greta %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Rectangular Kronecker Coefficients and Plethysms in Geometric Complexity Theory : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-0E59-D %R 10.1016/j.aim.2017.08.024 %7 2017 %D 2017 %J Advances in Mathematics %V 319 %& 40 %P 40 - 66 %I Academic Press %C Orlando, Fla. %@ false
[92]
C. Ikenmeyer and J. M. Landsberg, “On the Complexity of the Permanent in Various Computational Models,” Journal of Pure and Applied Algebra, vol. 221, no. 12, 2017.
Export
BibTeX
@article{IL:17, TITLE = {On the Complexity of the Permanent in Various Computational Models}, AUTHOR = {Ikenmeyer, Christian and Landsberg, J. M.}, LANGUAGE = {eng}, ISSN = {0022-4049}, DOI = {10.1016/j.jpaa.2017.02.008}, PUBLISHER = {North-Holland}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Journal of Pure and Applied Algebra}, VOLUME = {221}, NUMBER = {12}, PAGES = {2911--2927}, }
Endnote
%0 Journal Article %A Ikenmeyer, Christian %A Landsberg, J. M. %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On the Complexity of the Permanent in Various Computational Models : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4F23-D %R 10.1016/j.jpaa.2017.02.008 %7 2017-02-23 %D 2017 %J Journal of Pure and Applied Algebra %O J. Pure Appl. Algebra %V 221 %N 12 %& 2911 %P 2911 - 2927 %I North-Holland %C Amsterdam %@ false
[93]
G. Jindal, P. Kolev, R. Peng, and S. Sawlani, “Density Independent Algorithms for Sparsifying k-Step Random Walks,” 2017. [Online]. Available: http://arxiv.org/abs/1702.06110. (arXiv: 1702.06110)
Abstract
We give faster algorithms for producing sparse approximations of the transition matrices of $k$-step random walks on undirected, weighted graphs. These transition matrices also form graphs, and arise as intermediate objects in a variety of graph algorithms. Our improvements are based on a better understanding of processes that sample such walks, as well as tighter bounds on key weights underlying these sampling processes. On a graph with $n$ vertices and $m$ edges, our algorithm produces a graph with about $n\log{n}$ edges that approximates the $k$-step random walk graph in about $m + n \log^4{n}$ time. In order to obtain this runtime bound, we also revisit "density independent" algorithms for sparsifying graphs whose runtime overhead is expressed only in terms of the number of vertices.
Export
BibTeX
@online{DBLP:journals/corr/JindalKPS17, TITLE = {Density Independent Algorithms for Sparsifying $k$-Step Random Walks}, AUTHOR = {Jindal, Gorav and Kolev, Pavel and Peng, Richard and Sawlani, Saurabh}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1702.06110}, EPRINT = {1702.06110}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We give faster algorithms for producing sparse approximations of the transition matrices of $k$-step random walks on undirected, weighted graphs. These transition matrices also form graphs, and arise as intermediate objects in a variety of graph algorithms. Our improvements are based on a better understanding of processes that sample such walks, as well as tighter bounds on key weights underlying these sampling processes. On a graph with $n$ vertices and $m$ edges, our algorithm produces a graph with about $n\log{n}$ edges that approximates the $k$-step random walk graph in about $m + n \log^4{n}$ time. In order to obtain this runtime bound, we also revisit "density independent" algorithms for sparsifying graphs whose runtime overhead is expressed only in terms of the number of vertices.}, }
Endnote
%0 Report %A Jindal, Gorav %A Kolev, Pavel %A Peng, Richard %A Sawlani, Saurabh %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Density Independent Algorithms for Sparsifying k-Step Random Walks : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-26A6-1 %U http://arxiv.org/abs/1702.06110 %D 2017 %X We give faster algorithms for producing sparse approximations of the transition matrices of $k$-step random walks on undirected, weighted graphs. These transition matrices also form graphs, and arise as intermediate objects in a variety of graph algorithms. Our improvements are based on a better understanding of processes that sample such walks, as well as tighter bounds on key weights underlying these sampling processes. On a graph with $n$ vertices and $m$ edges, our algorithm produces a graph with about $n\log{n}$ edges that approximates the $k$-step random walk graph in about $m + n \log^4{n}$ time. In order to obtain this runtime bound, we also revisit "density independent" algorithms for sparsifying graphs whose runtime overhead is expressed only in terms of the number of vertices. %K Computer Science, Data Structures and Algorithms, cs.DS
[94]
G. Jindal and M. Sagraloff, “Efficiently Computing Real Roots of Sparse Polynomials,” in ISSAC’17, International Symposium on Symbolic and Algebraic Computation, Kaiserslautern, Germany, 2017.
Export
BibTeX
@inproceedings{JindalISSAC2017, TITLE = {Efficiently Computing Real Roots of Sparse Polynomials}, AUTHOR = {Jindal, Gorav and Sagraloff, Michael}, LANGUAGE = {eng}, ISBN = {978-1-4503-5064-8}, DOI = {10.1145/3087604.3087652}, PUBLISHER = {ACM}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {ISSAC{\textquoteright}17, International Symposium on Symbolic and Algebraic Computation}, PAGES = {229--236}, ADDRESS = {Kaiserslautern, Germany}, }
Endnote
%0 Conference Proceedings %A Jindal, Gorav %A Sagraloff, Michael %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Efficiently Computing Real Roots of Sparse Polynomials : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-FBA6-4 %R 10.1145/3087604.3087652 %D 2017 %B International Symposium on Symbolic and Algebraic Computation %Z date of event: 2017-07-25 - 2017-07-28 %C Kaiserslautern, Germany %B ISSAC&#8217;17 %P 229 - 236 %I ACM %@ 978-1-4503-5064-8
[95]
G. Jindal and M. Sagraloff, “Efficiently Computing Real Roots of Sparse Polynomials,” 2017. [Online]. Available: http://arxiv.org/abs/1704.06979. (arXiv: 1704.06979)
Abstract
We propose an efficient algorithm to compute the real roots of a sparse polynomial $f\in\mathbb{R}[x]$ having $k$ non-zero real-valued coefficients. It is assumed that arbitrarily good approximations of the non-zero coefficients are given by means of a coefficient oracle. For a given positive integer $L$, our algorithm returns disjoint disks $\Delta_{1},\ldots,\Delta_{s}\subset\mathbb{C}$, with $s<2k$, centered at the real axis and of radius less than $2^{-L}$ together with positive integers $\mu_{1},\ldots,\mu_{s}$ such that each disk $\Delta_{i}$ contains exactly $\mu_{i}$ roots of $f$ counted with multiplicity. In addition, it is ensured that each real root of $f$ is contained in one of the disks. If $f$ has only simple real roots, our algorithm can also be used to isolate all real roots. The bit complexity of our algorithm is polynomial in $k$ and $\log n$, and near-linear in $L$ and $\tau$, where $2^{-\tau}$ and $2^{\tau}$ constitute lower and upper bounds on the absolute values of the non-zero coefficients of $f$, and $n$ is the degree of $f$. For root isolation, the bit complexity is polynomial in $k$ and $\log n$, and near-linear in $\tau$ and $\log\sigma^{-1}$, where $\sigma$ denotes the separation of the real roots.
Export
BibTeX
@online{DBLP:journals/corr/JindalS17, TITLE = {Efficiently Computing Real Roots of Sparse Polynomials}, AUTHOR = {Jindal, Gorav and Sagraloff, Michael}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1704.06979}, EPRINT = {1704.06979}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We propose an efficient algorithm to compute the real roots of a sparse polynomial $f\in\mathbb{R}[x]$ having $k$ non-zero real-valued coefficients. It is assumed that arbitrarily good approximations of the non-zero coefficients are given by means of a coefficient oracle. For a given positive integer $L$, our algorithm returns disjoint disks $\Delta_{1},\ldots,\Delta_{s}\subset\mathbb{C}$, with $s<2k$, centered at the real axis and of radius less than $2^{-L}$ together with positive integers $\mu_{1},\ldots,\mu_{s}$ such that each disk $\Delta_{i}$ contains exactly $\mu_{i}$ roots of $f$ counted with multiplicity. In addition, it is ensured that each real root of $f$ is contained in one of the disks. If $f$ has only simple real roots, our algorithm can also be used to isolate all real roots. The bit complexity of our algorithm is polynomial in $k$ and $\log n$, and near-linear in $L$ and $\tau$, where $2^{-\tau}$ and $2^{\tau}$ constitute lower and upper bounds on the absolute values of the non-zero coefficients of $f$, and $n$ is the degree of $f$. For root isolation, the bit complexity is polynomial in $k$ and $\log n$, and near-linear in $\tau$ and $\log\sigma^{-1}$, where $\sigma$ denotes the separation of the real roots.}, }
Endnote
%0 Report %A Jindal, Gorav %A Sagraloff, Michael %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Efficiently Computing Real Roots of Sparse Polynomials : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-8AD1-7 %U http://arxiv.org/abs/1704.06979 %D 2017 %X We propose an efficient algorithm to compute the real roots of a sparse polynomial $f\in\mathbb{R}[x]$ having $k$ non-zero real-valued coefficients. It is assumed that arbitrarily good approximations of the non-zero coefficients are given by means of a coefficient oracle. For a given positive integer $L$, our algorithm returns disjoint disks $\Delta_{1},\ldots,\Delta_{s}\subset\mathbb{C}$, with $s<2k$, centered at the real axis and of radius less than $2^{-L}$ together with positive integers $\mu_{1},\ldots,\mu_{s}$ such that each disk $\Delta_{i}$ contains exactly $\mu_{i}$ roots of $f$ counted with multiplicity. In addition, it is ensured that each real root of $f$ is contained in one of the disks. If $f$ has only simple real roots, our algorithm can also be used to isolate all real roots. The bit complexity of our algorithm is polynomial in $k$ and $\log n$, and near-linear in $L$ and $\tau$, where $2^{-\tau}$ and $2^{\tau}$ constitute lower and upper bounds on the absolute values of the non-zero coefficients of $f$, and $n$ is the degree of $f$. For root isolation, the bit complexity is polynomial in $k$ and $\log n$, and near-linear in $\tau$ and $\log\sigma^{-1}$, where $\sigma$ denotes the separation of the real roots. %K Computer Science, Symbolic Computation, cs.SC
[96]
A. Karrenbauer, R. Becker, C. Scholl, and B. Becker, “From DQBF to QBF by Dependency Elimination,” in Theory and Applications of Satisfiability Testing -- SAT 2017, Melbourne, Australia, 2017.
Export
BibTeX
@inproceedings{KarrenbauerSAT2017, TITLE = {From {DQBF} to {QBF} by Dependency Elimination}, AUTHOR = {Karrenbauer, Andreas and Becker, Ruben and Scholl, Christoph and Becker, Bernd}, LANGUAGE = {eng}, ISBN = {978-3-319-66262-6}, DOI = {10.1007/978-3-319-66263-3_21}, PUBLISHER = {Springer}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Theory and Applications of Satisfiability Testing -- SAT 2017}, EDITOR = {Gaspers, Serge and Walsh, Toby}, PAGES = {326--343}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10491}, ADDRESS = {Melbourne, Australia}, }
Endnote
%0 Conference Proceedings %A Karrenbauer, Andreas %A Becker, Ruben %A Scholl, Christoph %A Becker, Bernd %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T From DQBF to QBF by Dependency Elimination : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-FBAC-7 %R 10.1007/978-3-319-66263-3_21 %D 2017 %B 20th International Conference on Theory and Applications of Satisfiability Testing %Z date of event: 2017-08-28 - 2017-09-01 %C Melbourne, Australia %B Theory and Applications of Satisfiability Testing -- SAT 2017 %E Gaspers, Serge; Walsh, Toby %P 326 - 343 %I Springer %@ 978-3-319-66262-6 %B Lecture Notes in Computer Science %N 10491
[97]
C. Lenzen, N. A. Lynch, C. Newport, and T. Radeva, “Searching without Communicating: Tradeoffs between Performance and Selection Complexity,” Distributed Computing, vol. 30, no. 3, 2017.
Export
BibTeX
@article{DBLP:journals/dc/LenzenLNR17, TITLE = {Searching without Communicating: {T}radeoffs between Performance and Selection Complexity}, AUTHOR = {Lenzen, Christoph and Lynch, Nancy A. and Newport, Calvin and Radeva, Tsvetomira}, LANGUAGE = {eng}, ISSN = {0178-2770}, DOI = {10.1007/s00446-016-0283-x}, PUBLISHER = {Springer International}, ADDRESS = {Berlin}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Distributed Computing}, VOLUME = {30}, NUMBER = {3}, PAGES = {169--191}, }
Endnote
%0 Journal Article %A Lenzen, Christoph %A Lynch, Nancy A. %A Newport, Calvin %A Radeva, Tsvetomira %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Searching without Communicating: Tradeoffs between Performance and Selection Complexity : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-8AA7-9 %R 10.1007/s00446-016-0283-x %7 2016-09-19 %D 2017 %J Distributed Computing %V 30 %N 3 %& 169 %P 169 - 191 %I Springer International %C Berlin %@ false
[98]
C. Lenzen and J. Rybicki, “Self-Stabilising Byzantine Clock Synchronisation is Almost as Easy as Consensus,” in 31 International Symposium on Distributed Computing (DISC 2017), Vienna, Austria, 2017.
Export
BibTeX
@inproceedings{Lenzen_DISC17b, TITLE = {Self-Stabilising {B}yzantine Clock Synchronisation is Almost as Easy as Consensus}, AUTHOR = {Lenzen, Christoph and Rybicki, Joel}, LANGUAGE = {eng}, ISBN = {978-3-95977-053-8}, URL = {urn:nbn:de:0030-drops-79914}, DOI = {10.4230/LIPIcs.DISC.2017.32}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {31 International Symposium on Distributed Computing (DISC 2017)}, EDITOR = {Richa, Andr{\'e}a W.}, PAGES = {1--15}, EID = {32}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {91}, ADDRESS = {Vienna, Austria}, }
Endnote
%0 Conference Proceedings %A Lenzen, Christoph %A Rybicki, Joel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Self-Stabilising Byzantine Clock Synchronisation is Almost as Easy as Consensus : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F31-3 %U urn:nbn:de:0030-drops-79914 %R 10.4230/LIPIcs.DISC.2017.32 %D 2017 %B 31st International Symposium on Distributed Computing %Z date of event: 2017-10-16 - 2017-10-20 %C Vienna, Austria %B 31 International Symposium on Distributed Computing %E Richa, Andr&#233;a W. %P 1 - 15 %Z sequence number: 32 %I Schloss Dagstuhl %@ 978-3-95977-053-8 %B Leibniz International Proceedings in Informatics %N 91 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2017/7991/
[99]
C. Lenzen and J. Rybicki, “Efficient Counting with Optimal Resilience,” SIAM Journal on Computing, vol. 46, no. 4, 2017.
Export
BibTeX
@article{LenzenRybicki2017, TITLE = {Efficient Counting with Optimal Resilience}, AUTHOR = {Lenzen, Christoph and Rybicki, Joel}, LANGUAGE = {eng}, ISSN = {0097-5397}, DOI = {10.1137/16M107877X}, PUBLISHER = {Society for Industrial and Applied Mathematics.}, ADDRESS = {Philadelphia, PA}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {SIAM Journal on Computing}, VOLUME = {46}, NUMBER = {4}, PAGES = {1473--1500}, }
Endnote
%0 Journal Article %A Lenzen, Christoph %A Rybicki, Joel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Efficient Counting with Optimal Resilience : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-EE93-6 %R 10.1137/16M107877X %7 2017 %D 2017 %J SIAM Journal on Computing %V 46 %N 4 %& 1473 %P 1473 - 1500 %I Society for Industrial and Applied Mathematics. %C Philadelphia, PA %@ false
[100]
C. Lenzen and J. Rybicki, “Self-stabilising Byzantine Clock Synchronisation is Almost as Easy as Consensus,” 2017. [Online]. Available: http://arxiv.org/abs/1705.06173. (arXiv: 1705.06173)
Abstract
We give fault-tolerant algorithms for establishing synchrony in distributed systems in which each of the $n$ nodes has its own clock. Our algorithms operate in a very strong fault model: we require self-stabilisation, i.e., the initial state of the system may be arbitrary, and there can be up to $f<n/3$ ongoing Byzantine faults, i.e., nodes that deviate from the protocol in an arbitrary manner. Furthermore, we assume that the local clocks of the nodes may progress at different speeds (clock drift) and communication has bounded delay. In this model, we study the pulse synchronisation problem, where the task is to guarantee that eventually all correct nodes generate well-separated local pulse events (i.e., unlabelled logical clock ticks) in a synchronised manner. Compared to prior work, we achieve exponential improvements in stabilisation time and the number of communicated bits, and give the first sublinear-time algorithm for the problem: - In the deterministic setting, the state-of-the-art solutions stabilise in time $\Theta(f)$ and have each node broadcast $\Theta(f \log f)$ bits per time unit. We exponentially reduce the number of bits broadcasted per time unit to $\Theta(\log f)$ while retaining the same stabilisation time. - In the randomised setting, the state-of-the-art solutions stabilise in time $\Theta(f)$ and have each node broadcast $O(1)$ bits per time unit. We exponentially reduce the stabilisation time to $\log^{O(1)} f$ while each node broadcasts $\log^{O(1)} f$ bits per time unit. These results are obtained by means of a recursive approach reducing the above task of self-stabilising pulse synchronisation in the bounded-delay model to non-self-stabilising binary consensus in the synchronous model. In general, our approach introduces at most logarithmic overheads in terms of stabilisation time and broadcasted bits over the underlying consensus routine.
Export
BibTeX
@online{DBLP:journals/corr/LenzenR17, TITLE = {Self-stabilising Byzantine Clock Synchronisation is Almost as Easy as Consensus}, AUTHOR = {Lenzen, Christoph and Rybicki, Joel}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1705.06173}, EPRINT = {1705.06173}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We give fault-tolerant algorithms for establishing synchrony in distributed systems in which each of the $n$ nodes has its own clock. Our algorithms operate in a very strong fault model: we require self-stabilisation, i.e., the initial state of the system may be arbitrary, and there can be up to $f<n/3$ ongoing Byzantine faults, i.e., nodes that deviate from the protocol in an arbitrary manner. Furthermore, we assume that the local clocks of the nodes may progress at different speeds (clock drift) and communication has bounded delay. In this model, we study the pulse synchronisation problem, where the task is to guarantee that eventually all correct nodes generate well-separated local pulse events (i.e., unlabelled logical clock ticks) in a synchronised manner. Compared to prior work, we achieve exponential improvements in stabilisation time and the number of communicated bits, and give the first sublinear-time algorithm for the problem: -- In the deterministic setting, the state-of-the-art solutions stabilise in time $\Theta(f)$ and have each node broadcast $\Theta(f \log f)$ bits per time unit. We exponentially reduce the number of bits broadcasted per time unit to $\Theta(\log f)$ while retaining the same stabilisation time. -- In the randomised setting, the state-of-the-art solutions stabilise in time $\Theta(f)$ and have each node broadcast $O(1)$ bits per time unit. We exponentially reduce the stabilisation time to $\log^{O(1)} f$ while each node broadcasts $\log^{O(1)} f$ bits per time unit. These results are obtained by means of a recursive approach reducing the above task of self-stabilising pulse synchronisation in the bounded-delay model to non-self-stabilising binary consensus in the synchronous model. In general, our approach introduces at most logarithmic overheads in terms of stabilisation time and broadcasted bits over the underlying consensus routine.}, }
Endnote
%0 Report %A Lenzen, Christoph %A Rybicki, Joel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Self-stabilising Byzantine Clock Synchronisation is Almost as Easy as Consensus : %O Self-stabilising {B}yzantine Clock Synchronisation is Almost as Easy as Consensus %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-8AAA-3 %U http://arxiv.org/abs/1705.06173 %D 2017 %X We give fault-tolerant algorithms for establishing synchrony in distributed systems in which each of the $n$ nodes has its own clock. Our algorithms operate in a very strong fault model: we require self-stabilisation, i.e., the initial state of the system may be arbitrary, and there can be up to $f<n/3$ ongoing Byzantine faults, i.e., nodes that deviate from the protocol in an arbitrary manner. Furthermore, we assume that the local clocks of the nodes may progress at different speeds (clock drift) and communication has bounded delay. In this model, we study the pulse synchronisation problem, where the task is to guarantee that eventually all correct nodes generate well-separated local pulse events (i.e., unlabelled logical clock ticks) in a synchronised manner. Compared to prior work, we achieve exponential improvements in stabilisation time and the number of communicated bits, and give the first sublinear-time algorithm for the problem: - In the deterministic setting, the state-of-the-art solutions stabilise in time $\Theta(f)$ and have each node broadcast $\Theta(f \log f)$ bits per time unit. We exponentially reduce the number of bits broadcasted per time unit to $\Theta(\log f)$ while retaining the same stabilisation time. - In the randomised setting, the state-of-the-art solutions stabilise in time $\Theta(f)$ and have each node broadcast $O(1)$ bits per time unit. We exponentially reduce the stabilisation time to $\log^{O(1)} f$ while each node broadcasts $\log^{O(1)} f$ bits per time unit. These results are obtained by means of a recursive approach reducing the above task of self-stabilising pulse synchronisation in the bounded-delay model to non-self-stabilising binary consensus in the synchronous model. In general, our approach introduces at most logarithmic overheads in terms of stabilisation time and broadcasted bits over the underlying consensus routine. %K Computer Science, Distributed, Parallel, and Cluster Computing, cs.DC
[101]
C. Lenzen and M. Medina, “Robust Routing Made Easy,” in Stabilization, Safety, and Security of Distributed Systems (SSS 2017), Boston, MA, USA, 2017.
Export
BibTeX
@inproceedings{LenzenSSS2017, TITLE = {Robust Routing Made Easy}, AUTHOR = {Lenzen, Christoph and Medina, Moti}, LANGUAGE = {eng}, ISBN = {978-3-319-69083-4}, DOI = {10.1007/978-3-319-69084-1_13}, PUBLISHER = {Springer}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {Stabilization, Safety, and Security of Distributed Systems (SSS 2017)}, EDITOR = {Spirakis, Paul and Tsigas, Philippas}, PAGES = {187--202}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10616}, ADDRESS = {Boston, MA, USA}, }
Endnote
%0 Conference Proceedings %A Lenzen, Christoph %A Medina, Moti %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Robust Routing Made Easy : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F35-F %R 10.1007/978-3-319-69084-1_13 %D 2017 %B 19th International Symposium on Stabilization, Safety, and Security of Distributed System %Z date of event: 2017-11-05 - 2017-11-08 %C Boston, MA, USA %B Stabilization, Safety, and Security of Distributed Systems %E Spirakis, Paul; Tsigas, Philippas %P 187 - 202 %I Springer %@ 978-3-319-69083-4 %B Lecture Notes in Computer Science %N 10616
[102]
C. Lenzen and R. Levi, “Brief Announcement: A Centralized Local Algorithm for the Sparse Spanning Graph Problem,” in 31 International Symposium on Distributed Computing (DISC 2017), Vienna, Austria, 2017.
Export
BibTeX
@inproceedings{Lenzen_DISC17, TITLE = {Brief Announcement: {A} Centralized Local Algorithm for the Sparse Spanning Graph Problem}, AUTHOR = {Lenzen, Christoph and Levi, Reut}, LANGUAGE = {eng}, ISBN = {978-3-95977-053-8}, URL = {urn:nbn:de:0030-drops-80064}, DOI = {10.4230/LIPIcs.DISC.2017.57}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {31 International Symposium on Distributed Computing (DISC 2017)}, EDITOR = {Richa, Andr{\'e}a W.}, PAGES = {1--3}, EID = {57}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {91}, ADDRESS = {Vienna, Austria}, }
Endnote
%0 Conference Proceedings %A Lenzen, Christoph %A Levi, Reut %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Brief Announcement: A Centralized Local Algorithm for the Sparse Spanning Graph Problem : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3F28-E %U urn:nbn:de:0030-drops-80064 %R 10.4230/LIPIcs.DISC.2017.57 %D 2017 %B 31st International Symposium on Distributed Computing %Z date of event: 2017-10-16 - 2017-10-20 %C Vienna, Austria %B 31 International Symposium on Distributed Computing %E Richa, Andr&#233;a W. %P 1 - 3 %Z sequence number: 57 %I Schloss Dagstuhl %@ 978-3-95977-053-8 %B Leibniz International Proceedings in Informatics %N 91 %U http://drops.dagstuhl.de/opus/volltexte/2017/8006/http://drops.dagstuhl.de/doku/urheberrecht1.html
[103]
C. Lenzen and R. Levi, “A Local Algorithm for the Sparse Spanning Graph Problem,” 2017. [Online]. Available: http://arxiv.org/abs/1703.05418. (arXiv: 1703.05418)
Abstract
Constructing a sparse \emph{spanning subgraph} is a fundamental primitive in graph theory. In this paper, we study this problem in the Centralized Local model, where the goal is to decide whether an edge is part of the spanning subgraph by examining only a small part of the input; yet, answers must be globally consistent and independent of prior queries. Unfortunately, maximally sparse spanning subgraphs, i.e., spanning trees, cannot be constructed efficiently in this model. Therefore, we settle for a spanning subgraph containing at most $(1+\varepsilon)n$ edges (where $n$ is the number of vertices and $\varepsilon$ is a given approximation/sparsity parameter). We achieve query complexity of $\tilde{O}(poly(\Delta/\varepsilon)n^{2/3})$,\footnote{$\tilde{O}$-notation hides polylogarithmic factors in $n$.} where $\Delta$ is the maximum degree of the input graph. Our algorithm is the first to do so on arbitrary graphs. Moreover, we achieve the additional property that our algorithm outputs a \emph{spanner,} i.e., distances are approximately preserved. With high probability, for each deleted edge there is a path of $O(poly(\Delta/\varepsilon)\log^2 n)$ hops in the output that connects its endpoints.
Export
BibTeX
@online{DBLP:journals/corr/LenzenL17, TITLE = {A Local Algorithm for the Sparse Spanning Graph Problem}, AUTHOR = {Lenzen, Christoph and Levi, Reut}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1703.05418}, EPRINT = {1703.05418}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Constructing a sparse \emph{spanning subgraph} is a fundamental primitive in graph theory. In this paper, we study this problem in the Centralized Local model, where the goal is to decide whether an edge is part of the spanning subgraph by examining only a small part of the input; yet, answers must be globally consistent and independent of prior queries. Unfortunately, maximally sparse spanning subgraphs, i.e., spanning trees, cannot be constructed efficiently in this model. Therefore, we settle for a spanning subgraph containing at most $(1+\varepsilon)n$ edges (where $n$ is the number of vertices and $\varepsilon$ is a given approximation/sparsity parameter). We achieve query complexity of $\tilde{O}(poly(\Delta/\varepsilon)n^{2/3})$,\footnote{$\tilde{O}$-notation hides polylogarithmic factors in $n$.} where $\Delta$ is the maximum degree of the input graph. Our algorithm is the first to do so on arbitrary graphs. Moreover, we achieve the additional property that our algorithm outputs a \emph{spanner,} i.e., distances are approximately preserved. With high probability, for each deleted edge there is a path of $O(poly(\Delta/\varepsilon)\log^2 n)$ hops in the output that connects its endpoints.}, }
Endnote
%0 Report %A Lenzen, Christoph %A Levi, Reut %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Local Algorithm for the Sparse Spanning Graph Problem : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-8AB0-4 %U http://arxiv.org/abs/1703.05418 %D 2017 %X Constructing a sparse \emph{spanning subgraph} is a fundamental primitive in graph theory. In this paper, we study this problem in the Centralized Local model, where the goal is to decide whether an edge is part of the spanning subgraph by examining only a small part of the input; yet, answers must be globally consistent and independent of prior queries. Unfortunately, maximally sparse spanning subgraphs, i.e., spanning trees, cannot be constructed efficiently in this model. Therefore, we settle for a spanning subgraph containing at most $(1+\varepsilon)n$ edges (where $n$ is the number of vertices and $\varepsilon$ is a given approximation/sparsity parameter). We achieve query complexity of $\tilde{O}(poly(\Delta/\varepsilon)n^{2/3})$,\footnote{$\tilde{O}$-notation hides polylogarithmic factors in $n$.} where $\Delta$ is the maximum degree of the input graph. Our algorithm is the first to do so on arbitrary graphs. Moreover, we achieve the additional property that our algorithm outputs a \emph{spanner,} i.e., distances are approximately preserved. With high probability, for each deleted edge there is a path of $O(poly(\Delta/\varepsilon)\log^2 n)$ hops in the output that connects its endpoints. %K Computer Science, Data Structures and Algorithms, cs.DS
[104]
C. Lenzen and M. Medina, “Robust Routing Made Easy,” 2017. [Online]. Available: http://arxiv.org/abs/1705.04042. (arXiv: 1705.04042)
Abstract
Designing routing schemes is a multidimensional and complex task that depends on the objective function, the computational model (centralized vs. distributed), and the amount of uncertainty (online vs. offline). Nevertheless, there are quite a few well-studied general techniques, for a large variety of network problems. In contrast, in our view, practical techniques for designing robust routing schemes are scarce; while fault-tolerance has been studied from a number of angles, existing approaches are concerned with dealing with faults after the fact by rerouting, self-healing, or similar techniques. We argue that this comes at a high burden for the designer, as in such a system any algorithm must account for the effects of faults on communication. With the goal of initiating efforts towards addressing this issue, we showcase simple and generic transformations that can be used as a blackbox to increase resilience against (independently distributed) faults. Given a network and a routing scheme, we determine a reinforced network and corresponding routing scheme that faithfully preserves the specification and behavior of the original scheme. We show that reasonably small constant overheads in terms of size of the new network compared to the old are sufficient for substantially relaxing the reliability requirements on individual components. The main message in this paper is that the task of designing a robust routing scheme can be decoupled into (i) designing a routing scheme that meets the specification in a fault-free environment, (ii) ensuring that nodes correspond to fault-containment regions, i.e., fail (approximately) independently, and (iii) applying our transformation to obtain a reinforced network and a robust routing scheme that is fault-tolerant.
Export
BibTeX
@online{DBLP:journals/corr/LenzenM17, TITLE = {Robust Routing Made Easy}, AUTHOR = {Lenzen, Christoph and Medina, Moti}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1705.04042}, EPRINT = {1705.04042}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Designing routing schemes is a multidimensional and complex task that depends on the objective function, the computational model (centralized vs. distributed), and the amount of uncertainty (online vs. offline). Nevertheless, there are quite a few well-studied general techniques, for a large variety of network problems. In contrast, in our view, practical techniques for designing robust routing schemes are scarce; while fault-tolerance has been studied from a number of angles, existing approaches are concerned with dealing with faults after the fact by rerouting, self-healing, or similar techniques. We argue that this comes at a high burden for the designer, as in such a system any algorithm must account for the effects of faults on communication. With the goal of initiating efforts towards addressing this issue, we showcase simple and generic transformations that can be used as a blackbox to increase resilience against (independently distributed) faults. Given a network and a routing scheme, we determine a reinforced network and corresponding routing scheme that faithfully preserves the specification and behavior of the original scheme. We show that reasonably small constant overheads in terms of size of the new network compared to the old are sufficient for substantially relaxing the reliability requirements on individual components. The main message in this paper is that the task of designing a robust routing scheme can be decoupled into (i) designing a routing scheme that meets the specification in a fault-free environment, (ii) ensuring that nodes correspond to fault-containment regions, i.e., fail (approximately) independently, and (iii) applying our transformation to obtain a reinforced network and a robust routing scheme that is fault-tolerant.}, }
Endnote
%0 Report %A Lenzen, Christoph %A Medina, Moti %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Robust Routing Made Easy : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-8AAD-E %U http://arxiv.org/abs/1705.04042 %D 2017 %X Designing routing schemes is a multidimensional and complex task that depends on the objective function, the computational model (centralized vs. distributed), and the amount of uncertainty (online vs. offline). Nevertheless, there are quite a few well-studied general techniques, for a large variety of network problems. In contrast, in our view, practical techniques for designing robust routing schemes are scarce; while fault-tolerance has been studied from a number of angles, existing approaches are concerned with dealing with faults after the fact by rerouting, self-healing, or similar techniques. We argue that this comes at a high burden for the designer, as in such a system any algorithm must account for the effects of faults on communication. With the goal of initiating efforts towards addressing this issue, we showcase simple and generic transformations that can be used as a blackbox to increase resilience against (independently distributed) faults. Given a network and a routing scheme, we determine a reinforced network and corresponding routing scheme that faithfully preserves the specification and behavior of the original scheme. We show that reasonably small constant overheads in terms of size of the new network compared to the old are sufficient for substantially relaxing the reliability requirements on individual components. The main message in this paper is that the task of designing a robust routing scheme can be decoupled into (i) designing a routing scheme that meets the specification in a fault-free environment, (ii) ensuring that nodes correspond to fault-containment regions, i.e., fail (approximately) independently, and (iii) applying our transformation to obtain a reinforced network and a robust routing scheme that is fault-tolerant. %K Computer Science, Distributed, Parallel, and Cluster Computing, cs.DC
[105]
R. Levi, G. Moshkovitz, D. Ron, R. Rubinfeld, and A. Shapira, “Constructing Near Spanning Trees with Few Local Inspections,” Random Structures and Algorithms, vol. 50, 2017.
Export
BibTeX
@article{LeviMRRS15, TITLE = {Constructing Near Spanning Trees with Few Local Inspections}, AUTHOR = {Levi, Reut and Moshkovitz, Guy and Ron, Dana and Rubinfeld, Ronitt and Shapira, Asaf}, LANGUAGE = {eng}, ISSN = {1042-9832}, DOI = {10.1002/rsa.20652}, PUBLISHER = {Wiley}, ADDRESS = {New York, N.Y.}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Random Structures and Algorithms}, VOLUME = {50}, PAGES = {183--200}, }
Endnote
%0 Journal Article %A Levi, Reut %A Moshkovitz, Guy %A Ron, Dana %A Rubinfeld, Ronitt %A Shapira, Asaf %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations External Organizations %T Constructing Near Spanning Trees with Few Local Inspections : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-601B-C %R 10.1002/rsa.20652 %7 2016 %D 2017 %J Random Structures and Algorithms %V 50 %& 183 %P 183 - 200 %I Wiley %C New York, N.Y. %@ false
[106]
K. Mehlhorn, S. Näher, and P. Sanders, “Engineering DFS-Based Graph Algorithms,” 2017. [Online]. Available: http://arxiv.org/abs/1703.10023. (arXiv: 1703.10023)
Export
BibTeX
@online{MehlhornDFSarXiv2017, TITLE = {Engineering {DFS}-Based Graph Algorithms}, AUTHOR = {Mehlhorn, Kurt and N{\"a}her, Stefan and Sanders, Peter}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1703.10023}, EPRINT = {1703.10023}, EPRINTTYPE = {arXiv}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, }
Endnote
%0 Report %A Mehlhorn, Kurt %A N&#228;her, Stefan %A Sanders, Peter %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Engineering DFS-Based Graph Algorithms : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-4DAE-7 %U http://arxiv.org/abs/1703.10023 %D 2017
[107]
K. Mehlhorn, A. Neumann, and J. M. Schmidt, “Certifying 3-Edge-Connectivity,” Algorithmica, vol. 77, no. 2, 2017.
Export
BibTeX
@article{Mehlhorn_Neumann_Schmidt2017, TITLE = {Certifying 3-Edge-Connectivity}, AUTHOR = {Mehlhorn, Kurt and Neumann, Adrian and Schmidt, Jens M.}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-015-0075-x}, PUBLISHER = {Springer}, ADDRESS = {New York, NX}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Algorithmica}, VOLUME = {77}, NUMBER = {2}, PAGES = {309--335}, }
Endnote
%0 Journal Article %A Mehlhorn, Kurt %A Neumann, Adrian %A Schmidt, Jens M. %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Certifying 3-Edge-Connectivity : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-6971-B %R 10.1007/s00453-015-0075-x %7 2015-09-22 %D 2017 %J Algorithmica %V 77 %N 2 %& 309 %P 309 - 335 %I Springer %C New York, NX %@ false
[108]
M. Mnich and E. J. van Leeuwen, “Polynomial Kernels for Deletion to Classes of Acyclic Digraphs,” Discrete Optimization, vol. 25, 2017.
Export
BibTeX
@article{MnichLeeuwen2017, TITLE = {Polynomial Kernels for Deletion to Classes of Acyclic Digraphs}, AUTHOR = {Mnich, Matthias and van Leeuwen, Erik Jan}, LANGUAGE = {eng}, ISSN = {1572-5286}, DOI = {10.1016/j.disopt.2017.02.002}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Discrete Optimization}, VOLUME = {25}, PAGES = {48--76}, }
Endnote
%0 Journal Article %A Mnich, Matthias %A van Leeuwen, Erik Jan %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Polynomial Kernels for Deletion to Classes of Acyclic Digraphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002D-DDCA-F %R 10.1016/j.disopt.2017.02.002 %7 2017 %D 2017 %J Discrete Optimization %V 25 %& 48 %P 48 - 76 %I Elsevier %C Amsterdam %@ false
[109]
N. Mustafa, K. Dutta, and A. Ghosh, “Simple Proof of Optimal Epsilon Nets,” Combinatorica, vol. First Online, 2017.
Export
BibTeX
@article{mustafa:hal-01360452, TITLE = {Simple Proof of Optimal Epsilon Nets}, AUTHOR = {Mustafa, Nabil and Dutta, Kunal and Ghosh, Arijit}, LANGUAGE = {eng}, ISSN = {0209-9683}, DOI = {10.1007/s00493-017-3564-5}, PUBLISHER = {Springer}, ADDRESS = {Heidelberg}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, JOURNAL = {Combinatorica}, VOLUME = {First Online}, PAGES = {1--9}, }
Endnote
%0 Journal Article %A Mustafa, Nabil %A Dutta, Kunal %A Ghosh, Arijit %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Simple Proof of Optimal Epsilon Nets : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-65CA-7 %R 10.1007/s00493-017-3564-5 %7 2017 %D 2017 %J Combinatorica %V First Online %& 1 %P 1 - 9 %I Springer %C Heidelberg %@ false
[110]
A. Oulasvirta, A. Feit, P. Lahteenlahti, and A. Karrenbauer, “Computational Support for Functionality Selection in Interaction Design,” ACM Transactions on Computer-Human Interaction, vol. 24, no. 5, 2017.
Export
BibTeX
@article{Oulasvirta2017, TITLE = {Computational Support for Functionality Selection in Interaction Design}, AUTHOR = {Oulasvirta, Antti and Feit, Anna and Lahteenlahti, Perttu and Karrenbauer, Andreas}, LANGUAGE = {eng}, DOI = {10.1145/3131608}, PUBLISHER = {ACM}, ADDRESS = {New York, NY}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {ACM Transactions on Computer-Human Interaction}, VOLUME = {24}, NUMBER = {5}, EID = {34}, }
Endnote
%0 Journal Article %A Oulasvirta, Antti %A Feit, Anna %A Lahteenlahti, Perttu %A Karrenbauer, Andreas %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Computational Support for Functionality Selection in Interaction Design : %G eng %U http://hdl.handle.net/21.11116/0000-0000-2DE0-1 %R 10.1145/3131608 %7 2017 %D 2017 %J ACM Transactions on Computer-Human Interaction %O TOCHI %V 24 %N 5 %Z sequence number: 34 %I ACM %C New York, NY
[111]
R. B. Tan, E. J. van Leeuwen, and J. van Leeuwen, “Shortcutting Directed and Undirected Networks with a Degree Constraint,” Discrete Applied Mathematics, vol. 220, 2017.
Export
BibTeX
@article{TanDAM2017, TITLE = {Shortcutting Directed and Undirected Networks with a Degree Constraint}, AUTHOR = {Tan, Richard B. and van Leeuwen, Erik Jan and van Leeuwen, Jan}, LANGUAGE = {eng}, ISSN = {0166-218X}, DOI = {10.1016/j.dam.2016.12.016}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, JOURNAL = {Discrete Applied Mathematics}, VOLUME = {220}, PAGES = {91--117}, }
Endnote
%0 Journal Article %A Tan, Richard B. %A van Leeuwen, Erik Jan %A van Leeuwen, Jan %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Shortcutting Directed and Undirected Networks with a Degree Constraint : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-539D-F %R 10.1016/j.dam.2016.12.016 %7 2017 %D 2017 %J Discrete Applied Mathematics %V 220 %& 91 %P 91 - 117 %I Elsevier %C Amsterdam %@ false
[112]
G. Tarawneh, M. Függer, and C. Lenzen, “Metastability Tolerant Computing,” in 23rd IEEE International Symposium on Asynchronous Circuits and Systems, San Diego, CA, USA, 2017.
Export
BibTeX
@inproceedings{TarawnehASYNC2017, TITLE = {Metastability Tolerant Computing}, AUTHOR = {Tarawneh, Ghaith and F{\"u}gger, Matthias and Lenzen, Christoph}, LANGUAGE = {eng}, ISBN = {978-1-5386-2749-5}, DOI = {10.1109/ASYNC.2017.9}, PUBLISHER = {IEEE}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, DATE = {2017}, BOOKTITLE = {23rd IEEE International Symposium on Asynchronous Circuits and Systems}, PAGES = {25--32}, ADDRESS = {San Diego, CA, USA}, }
Endnote
%0 Conference Proceedings %A Tarawneh, Ghaith %A F&#252;gger, Matthias %A Lenzen, Christoph %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Metastability Tolerant Computing : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-80CB-2 %R 10.1109/ASYNC.2017.9 %D 2017 %B 23rd IEEE International Symposium on Asynchronous Circuits and Systems %Z date of event: 2017-05-21 - 2017-05-24 %C San Diego, CA, USA %B 23rd IEEE International Symposium on Asynchronous Circuits and Systems %P 25 - 32 %I IEEE %@ 978-1-5386-2749-5
[113]
A. Wiese, “Independent Set of Convex Polygons: From nϵ to 1+ϵ via Shrinking,” Algorithmica, vol. First Online, 2017.
Export
BibTeX
@article{Wiese2017, TITLE = {Independent Set of Convex Polygons: From $n^{\epsilon}$ to 1+$\epsilon$ via Shrinking}, AUTHOR = {Wiese, Andreas}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-017-0347-8}, PUBLISHER = {Springer-Verlag}, ADDRESS = {New York}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, JOURNAL = {Algorithmica}, VOLUME = {First Online}, PAGES = {1--17}, }
Endnote
%0 Journal Article %A Wiese, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Independent Set of Convex Polygons: From n&#1013; to 1+&#1013; via Shrinking : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002E-2602-4 %R 10.1007/s00453-017-0347-8 %7 2017 %D 2017 %J Algorithmica %V First Online %& 1 %P 1 - 17 %I Springer-Verlag %C New York %@ false
[114]
D. Ziegler, A. Abujabal, R. S. Roy, and G. Weikum, “Efficiency-aware Answering of Compositional Questions using Answer Type Prediction,” in The 8th International Joint Conference on Natural Language Processing (IJCNLP 2017), Taipei, Taiwan, 2017.
Export
BibTeX
@inproceedings{ZieglerIJCNLP2017, TITLE = {Efficiency-aware Answering of Compositional Questions using Answer Type Prediction}, AUTHOR = {Ziegler, David and Abujabal, Abdalghani and Roy, Rishiraj Saha and Weikum, Gerhard}, LANGUAGE = {eng}, ISBN = {978-1-948087-01-8}, URL = {http://aclweb.org/anthology/I17-2038}, PUBLISHER = {Asian Federation of Natural Language Processing}, YEAR = {2017}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {The 8th International Joint Conference on Natural Language Processing (IJCNLP 2017)}, PAGES = {222--227}, ADDRESS = {Taipei, Taiwan}, }
Endnote
%0 Conference Proceedings %A Ziegler, David %A Abujabal, Abdalghani %A Roy, Rishiraj Saha %A Weikum, Gerhard %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Databases and Information Systems, MPI for Informatics, Max Planck Society Databases and Information Systems, MPI for Informatics, Max Planck Society Databases and Information Systems, MPI for Informatics, Max Planck Society %T Efficiency-aware Answering of Compositional Questions using Answer Type Prediction : %G eng %U http://hdl.handle.net/21.11116/0000-0000-3B5F-5 %U http://aclweb.org/anthology/I17-2038 %D 2017 %B 8th International Joint Conference on Natural Language Processing %Z date of event: 2017-11-27 - 2017-12-01 %C Taipei, Taiwan %B The 8th International Joint Conference on Natural Language Processing %P 222 - 227 %I Asian Federation of Natural Language Processing %@ 978-1-948087-01-8
2016
[115]
I. Abraham, D. Durfee, I. Koutis, S. Krinninger, and R. Peng, “On Fully Dynamic Graph Sparsifiers,” 2016. [Online]. Available: http://arxiv.org/abs/1604.02094. (arXiv: 1604.02094)
Abstract
We initiate the study of dynamic algorithms for graph sparsification problems and obtain fully dynamic algorithms, allowing both edge insertions and edge deletions, that take polylogarithmic time after each update in the graph. Our three main results are as follows. First, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-spectral sparsifier with amortized update time $poly(\log{n}, \epsilon^{-1})$. Second, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-cut sparsifier with \emph{worst-case} update time $poly(\log{n}, \epsilon^{-1})$. Both sparsifiers have size $ n \cdot poly(\log{n}, \epsilon^{-1})$. Third, we apply our dynamic sparsifier algorithm to obtain a fully dynamic algorithm for maintaining a $(1 + \epsilon)$-approximation to the value of the maximum flow in an unweighted, undirected, bipartite graph with amortized update time $poly(\log{n}, \epsilon^{-1})$.
Export
BibTeX
@online{Abrahamdkkp16, TITLE = {On Fully Dynamic Graph Sparsifiers}, AUTHOR = {Abraham, Ittai and Durfee, David and Koutis, Ioannis and Krinninger, Sebastian and Peng, Richard}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1604.02094}, EPRINT = {1604.02094}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We initiate the study of dynamic algorithms for graph sparsification problems and obtain fully dynamic algorithms, allowing both edge insertions and edge deletions, that take polylogarithmic time after each update in the graph. Our three main results are as follows. First, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-spectral sparsifier with amortized update time $poly(\log{n}, \epsilon^{-1})$. Second, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-cut sparsifier with \emph{worst-case} update time $poly(\log{n}, \epsilon^{-1})$. Both sparsifiers have size $ n \cdot poly(\log{n}, \epsilon^{-1})$. Third, we apply our dynamic sparsifier algorithm to obtain a fully dynamic algorithm for maintaining a $(1 + \epsilon)$-approximation to the value of the maximum flow in an unweighted, undirected, bipartite graph with amortized update time $poly(\log{n}, \epsilon^{-1})$.}, }
Endnote
%0 Report %A Abraham, Ittai %A Durfee, David %A Koutis, Ioannis %A Krinninger, Sebastian %A Peng, Richard %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On Fully Dynamic Graph Sparsifiers : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-510E-1 %U http://arxiv.org/abs/1604.02094 %D 2016 %X We initiate the study of dynamic algorithms for graph sparsification problems and obtain fully dynamic algorithms, allowing both edge insertions and edge deletions, that take polylogarithmic time after each update in the graph. Our three main results are as follows. First, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-spectral sparsifier with amortized update time $poly(\log{n}, \epsilon^{-1})$. Second, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-cut sparsifier with \emph{worst-case} update time $poly(\log{n}, \epsilon^{-1})$. Both sparsifiers have size $ n \cdot poly(\log{n}, \epsilon^{-1})$. Third, we apply our dynamic sparsifier algorithm to obtain a fully dynamic algorithm for maintaining a $(1 + \epsilon)$-approximation to the value of the maximum flow in an unweighted, undirected, bipartite graph with amortized update time $poly(\log{n}, \epsilon^{-1})$. %K Computer Science, Data Structures and Algorithms, cs.DS
[116]
I. Abraham, D. Durfee, I. Koutis, S. Krinninger, and R. Peng, “On Fully Dynamic Graph Sparsifiers,” in FOCS 2016, New Brunswick, NJ, USA, 2016.
Abstract
We initiate the study of dynamic algorithms for graph sparsification problems and obtain fully dynamic algorithms, allowing both edge insertions and edge deletions, that take polylogarithmic time after each update in the graph. Our three main results are as follows. First, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-spectral sparsifier with amortized update time $poly(\log{n}, \epsilon^{-1})$. Second, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-cut sparsifier with \emph{worst-case} update time $poly(\log{n}, \epsilon^{-1})$. Both sparsifiers have size $ n \cdot poly(\log{n}, \epsilon^{-1})$. Third, we apply our dynamic sparsifier algorithm to obtain a fully dynamic algorithm for maintaining a $(1 + \epsilon)$-approximation to the value of the maximum flow in an unweighted, undirected, bipartite graph with amortized update time $poly(\log{n}, \epsilon^{-1})$.
Export
BibTeX
@inproceedings{Abrahamdkkp2016, TITLE = {On Fully Dynamic Graph Sparsifiers}, AUTHOR = {Abraham, Ittai and Durfee, David and Koutis, Ioannis and Krinninger, Sebastian and Peng, Richard}, LANGUAGE = {eng}, ISBN = {978-1-5090-3933-3}, DOI = {10.1109/FOCS.2016.44}, PUBLISHER = {IEEE}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, ABSTRACT = {We initiate the study of dynamic algorithms for graph sparsification problems and obtain fully dynamic algorithms, allowing both edge insertions and edge deletions, that take polylogarithmic time after each update in the graph. Our three main results are as follows. First, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-spectral sparsifier with amortized update time $poly(\log{n}, \epsilon^{-1})$. Second, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-cut sparsifier with \emph{worst-case} update time $poly(\log{n}, \epsilon^{-1})$. Both sparsifiers have size $ n \cdot poly(\log{n}, \epsilon^{-1})$. Third, we apply our dynamic sparsifier algorithm to obtain a fully dynamic algorithm for maintaining a $(1 + \epsilon)$-approximation to the value of the maximum flow in an unweighted, undirected, bipartite graph with amortized update time $poly(\log{n}, \epsilon^{-1})$.}, BOOKTITLE = {FOCS 2016}, PAGES = {396--405}, ADDRESS = {New Brunswick, NJ, USA}, }
Endnote
%0 Conference Proceedings %A Abraham, Ittai %A Durfee, David %A Koutis, Ioannis %A Krinninger, Sebastian %A Peng, Richard %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On Fully Dynamic Graph Sparsifiers : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-52C6-A %R 10.1109/FOCS.2016.44 %D 2016 %B 57th Annual IEEE Symposium on Foundations of Computer Science %Z date of event: 2016-10-09 - 2016-10-11 %C New Brunswick, NJ, USA %X We initiate the study of dynamic algorithms for graph sparsification problems and obtain fully dynamic algorithms, allowing both edge insertions and edge deletions, that take polylogarithmic time after each update in the graph. Our three main results are as follows. First, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-spectral sparsifier with amortized update time $poly(\log{n}, \epsilon^{-1})$. Second, we give a fully dynamic algorithm for maintaining a $ (1 \pm \epsilon) $-cut sparsifier with \emph{worst-case} update time $poly(\log{n}, \epsilon^{-1})$. Both sparsifiers have size $ n \cdot poly(\log{n}, \epsilon^{-1})$. Third, we apply our dynamic sparsifier algorithm to obtain a fully dynamic algorithm for maintaining a $(1 + \epsilon)$-approximation to the value of the maximum flow in an unweighted, undirected, bipartite graph with amortized update time $poly(\log{n}, \epsilon^{-1})$. %K Computer Science, Data Structures and Algorithms, cs.DS %B FOCS 2016 %P 396 - 405 %I IEEE %@ 978-1-5090-3933-3
[117]
I. Abraham, S. Chechik, and S. Krinninger, “Fully Dynamic All-pairs Shortest Paths with Worst-case Update-time revisited,” 2016. [Online]. Available: http://arxiv.org/abs/1607.05132. (arXiv: 1607.05132)
Abstract
We revisit the classic problem of dynamically maintaining shortest paths between all pairs of nodes of a directed weighted graph. The allowed updates are insertions and deletions of nodes and their incident edges. We give worst-case guarantees on the time needed to process a single update (in contrast to related results, the update time is not amortized over a sequence of updates). Our main result is a simple randomized algorithm that for any parameter $c>1$ has a worst-case update time of $O(cn^{2+2/3} \log^{4/3}{n})$ and answers distance queries correctly with probability $1-1/n^c$, against an adaptive online adversary if the graph contains no negative cycle. The best deterministic algorithm is by Thorup [STOC 2005] with a worst-case update time of $\tilde O(n^{2+3/4})$ and assumes non-negative weights. This is the first improvement for this problem for more than a decade. Conceptually, our algorithm shows that randomization along with a more direct approach can provide better bounds.
Export
BibTeX
@online{Krinningerarxiv16, TITLE = {Fully Dynamic All-pairs Shortest Paths with Worst-case Update-time revisited}, AUTHOR = {Abraham, Ittai and Chechik, Shiri and Krinninger, Sebastian}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1607.05132}, EPRINT = {1607.05132}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We revisit the classic problem of dynamically maintaining shortest paths between all pairs of nodes of a directed weighted graph. The allowed updates are insertions and deletions of nodes and their incident edges. We give worst-case guarantees on the time needed to process a single update (in contrast to related results, the update time is not amortized over a sequence of updates). Our main result is a simple randomized algorithm that for any parameter $c>1$ has a worst-case update time of $O(cn^{2+2/3} \log^{4/3}{n})$ and answers distance queries correctly with probability $1-1/n^c$, against an adaptive online adversary if the graph contains no negative cycle. The best deterministic algorithm is by Thorup [STOC 2005] with a worst-case update time of $\tilde O(n^{2+3/4})$ and assumes non-negative weights. This is the first improvement for this problem for more than a decade. Conceptually, our algorithm shows that randomization along with a more direct approach can provide better bounds.}, }
Endnote
%0 Report %A Abraham, Ittai %A Chechik, Shiri %A Krinninger, Sebastian %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fully Dynamic All-pairs Shortest Paths with Worst-case Update-time revisited : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-50F8-A %U http://arxiv.org/abs/1607.05132 %D 2016 %X We revisit the classic problem of dynamically maintaining shortest paths between all pairs of nodes of a directed weighted graph. The allowed updates are insertions and deletions of nodes and their incident edges. We give worst-case guarantees on the time needed to process a single update (in contrast to related results, the update time is not amortized over a sequence of updates). Our main result is a simple randomized algorithm that for any parameter $c>1$ has a worst-case update time of $O(cn^{2+2/3} \log^{4/3}{n})$ and answers distance queries correctly with probability $1-1/n^c$, against an adaptive online adversary if the graph contains no negative cycle. The best deterministic algorithm is by Thorup [STOC 2005] with a worst-case update time of $\tilde O(n^{2+3/4})$ and assumes non-negative weights. This is the first improvement for this problem for more than a decade. Conceptually, our algorithm shows that randomization along with a more direct approach can provide better bounds. %K Computer Science, Data Structures and Algorithms, cs.DS
[118]
H. Ackermann, P. Berenbrink, S. Fischer, and M. Hoefer, “Concurrent Imitation Dynamics in Congestion Games,” Distributed Computing, vol. 29, no. 2, 2016.
Export
BibTeX
@article{Ackermann2016, TITLE = {Concurrent Imitation Dynamics in Congestion Games}, AUTHOR = {Ackermann, Heiner and Berenbrink, Petra and Fischer, Simon and Hoefer, Martin}, LANGUAGE = {eng}, ISSN = {0178-2770}, DOI = {10.1007/s00446-014-0223-6}, PUBLISHER = {Springer International}, ADDRESS = {Berlin}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Distributed Computing}, VOLUME = {29}, NUMBER = {2}, PAGES = {105--125}, }
Endnote
%0 Journal Article %A Ackermann, Heiner %A Berenbrink, Petra %A Fischer, Simon %A Hoefer, Martin %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Concurrent Imitation Dynamics in Congestion Games : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-C479-5 %R 10.1007/s00446-014-0223-6 %7 2014 %D 2016 %J Distributed Computing %V 29 %N 2 %& 105 %P 105 - 125 %I Springer International %C Berlin %@ false
[119]
A. Adamaszek, A. Antoniadis, and T. Mömke, “Airports and Railways: Facility Location Meets Network Design,” in 33rd International Symposium on Theoretical Aspects of Computer Science (STACS 2016), Orléans, France, 2016.
Export
BibTeX
@inproceedings{AdamaszekSTACS2016, TITLE = {Airports and Railways: {F}acility Location Meets Network Design}, AUTHOR = {Adamaszek, Anna and Antoniadis, Antonios and M{\"o}mke, Tobias}, LANGUAGE = {eng}, ISSN = {1868-896}, ISBN = {978-3-95977-001-9}, URL = {urn:nbn:de:0030-drops-57074}, DOI = {10.4230/LIPIcs.STACS.2016.6}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {33rd International Symposium on Theoretical Aspects of Computer Science (STACS 2016)}, EDITOR = {Ollinger, Nicolas and Vollmer, Heribert}, PAGES = {1--14}, EID = {6}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {47}, ADDRESS = {Orl{\'e}ans, France}, }
Endnote
%0 Conference Proceedings %A Adamaszek, Anna %A Antoniadis, Antonios %A M&#246;mke, Tobias %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Airports and Railways: Facility Location Meets Network Design : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-4312-A %R 10.4230/LIPIcs.STACS.2016.6 %U urn:nbn:de:0030-drops-57074 %D 2016 %B 33rd International Symposium on Theoretical Aspects of Computer Science %Z date of event: 2016-02-17 - 2016-02-20 %C Orl&#233;ans, France %B 33rd International Symposium on Theoretical Aspects of Computer Science %E Ollinger, Nicolas; Vollmer, Heribert %P 1 - 14 %Z sequence number: 6 %I Schloss Dagstuhl %@ 978-3-95977-001-9 %B Leibniz International Proceedings in Informatics %N 47 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2016/5707/http://drops.dagstuhl.de/doku/urheberrecht1.html
[120]
A. Adamaszek, P. Chalermsook, A. Ene, and A. Wiese, “Submodular Unsplittable Flow on Trees,” in Integer Programming and Combinatorial Optimization (IPCO 2016), Liège, Belgium, 2016.
Export
BibTeX
@inproceedings{AdamaszekIPCO2016, TITLE = {Submodular Unsplittable Flow on Trees}, AUTHOR = {Adamaszek, Anna and Chalermsook, Parinya and Ene, Alina and Wiese, Andreas}, LANGUAGE = {eng}, ISBN = {978-3-319-33460-8}, DOI = {10.1007/978-3-319-33461-5_28}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Integer Programming and Combinatorial Optimization (IPCO 2016)}, EDITOR = {Louveaux, Quentin and Skutella, Martin}, PAGES = {337--349}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9682}, ADDRESS = {Li{\`e}ge, Belgium}, }
Endnote
%0 Conference Proceedings %A Adamaszek, Anna %A Chalermsook, Parinya %A Ene, Alina %A Wiese, Andreas %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Submodular Unsplittable Flow on Trees : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0244-8 %R 10.1007/978-3-319-33461-5_28 %D 2016 %B 18th Conference on Integer Programming and Combinatorial Optimization %Z date of event: 2016-06-01 - 2016-06-03 %C Li&#232;ge, Belgium %B Integer Programming and Combinatorial Optimization %E Louveaux, Quentin; Skutella, Martin %P 337 - 349 %I Springer %@ 978-3-319-33460-8 %B Lecture Notes in Computer Science %N 9682
[121]
N. Alon, S. Moran, and A. Yehudayoff, “Sign Rank Versus VC Dimension,” in 29th Annual Conference on Learning Theory (COLT 2016), New York, NY, USA, 2016.
Export
BibTeX
@inproceedings{MoranCOLT2016, TITLE = {Sign Rank Versus {VC} Dimension}, AUTHOR = {Alon, Noga and Moran, Shay and Yehudayoff, Amir}, LANGUAGE = {eng}, ISSN = {1938-7228}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {29th Annual Conference on Learning Theory (COLT 2016)}, EDITOR = {Feldman, Vitaly and Rakhlin, Alexander and Shamir, Ohad}, SERIES = {JMLR Workshop \& Conference Proceedings}, VOLUME = {49}, ADDRESS = {New York, NY, USA}, }
Endnote
%0 Conference Proceedings %A Alon, Noga %A Moran, Shay %A Yehudayoff, Amir %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Sign Rank Versus VC Dimension : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-51AE-C %D 2016 %B 29th Annual Conference on Learning Theory %Z date of event: 2016-06-23 - 2016-06-26 %C New York, NY, USA %B 29th Annual Conference on Learning Theory %E Feldman, Vitaly; Rakhlin, Alexander; Shamir, Ohad %B JMLR Workshop & Conference Proceedings %N 49 %@ false
[122]
E. Althaus, B. Beber, W. Damm, S. Disch, W. Hagemann, A. Rakow, C. Scholl, U. Waldmann, and B. Wirtz, “Verification of Linear Hybrid Systems with Large Discrete State Spaces: Exploring the Design Space for Optimization,” SFB/TR 14 AVACS, ATR103, 2016.
Abstract
This paper provides a suite of optimization techniques for the verification of safety properties of linear hybrid automata with large discrete state spaces, such as naturally arising when incorporating health state monitoring and degradation levels into the controller design. Such models can -- in contrast to purely functional controller models -- not analyzed with hybrid verification engines relying on explicit representations of modes, but require fully symbolic representations for both the continuous and discrete part of the state space. The optimization techniques shown yield consistently a speedup of about 20 against previously published results for a similar benchmark suite, and complement these with new results on counterexample guided abstraction refinement. In combination with the methods guaranteeing preciseness of abstractions, this allows to significantly extend the class of models for which safety can be established, covering in particular models with 23 continuous variables and 2 to the 71 discrete states, 20 continuous variables and 2 to the 199 discrete states, and 9 continuous variables and 2 to the 271 discrete states.
Export
BibTeX
@techreport{AlthausBeberDammEtAl2016ATR, TITLE = {Verification of Linear Hybrid Systems with Large Discrete State Spaces: Exploring the Design Space for Optimization}, AUTHOR = {Althaus, Ernst and Beber, Bj{\"o}rn and Damm, Werner and Disch, Stefan and Hagemann, Willem and Rakow, Astrid and Scholl, Christoph and Waldmann, Uwe and Wirtz, Boris}, LANGUAGE = {eng}, ISSN = {1860-9821}, NUMBER = {ATR103}, INSTITUTION = {SFB/TR 14 AVACS}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, ABSTRACT = {This paper provides a suite of optimization techniques for the verification of safety properties of linear hybrid automata with large discrete state spaces, such as naturally arising when incorporating health state monitoring and degradation levels into the controller design. Such models can -- in contrast to purely functional controller models -- not analyzed with hybrid verification engines relying on explicit representations of modes, but require fully symbolic representations for both the continuous and discrete part of the state space. The optimization techniques shown yield consistently a speedup of about 20 against previously published results for a similar benchmark suite, and complement these with new results on counterexample guided abstraction refinement. In combination with the methods guaranteeing preciseness of abstractions, this allows to significantly extend the class of models for which safety can be established, covering in particular models with 23 continuous variables and 2 to the 71 discrete states, 20 continuous variables and 2 to the 199 discrete states, and 9 continuous variables and 2 to the 271 discrete states.}, TYPE = {AVACS Technical Report}, VOLUME = {103}, }
Endnote
%0 Report %A Althaus, Ernst %A Beber, Bj&#246;rn %A Damm, Werner %A Disch, Stefan %A Hagemann, Willem %A Rakow, Astrid %A Scholl, Christoph %A Waldmann, Uwe %A Wirtz, Boris %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations Automation of Logic, MPI for Informatics, Max Planck Society International Max Planck Research School, MPI for Informatics, Max Planck Society External Organizations External Organizations Automation of Logic, MPI for Informatics, Max Planck Society External Organizations %T Verification of Linear Hybrid Systems with Large Discrete State Spaces: Exploring the Design Space for Optimization : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4540-0 %Y SFB/TR 14 AVACS %D 2016 %P 93 p. %X This paper provides a suite of optimization techniques for the verification of safety properties of linear hybrid automata with large discrete state spaces, such as naturally arising when incorporating health state monitoring and degradation levels into the controller design. Such models can -- in contrast to purely functional controller models -- not analyzed with hybrid verification engines relying on explicit representations of modes, but require fully symbolic representations for both the continuous and discrete part of the state space. The optimization techniques shown yield consistently a speedup of about 20 against previously published results for a similar benchmark suite, and complement these with new results on counterexample guided abstraction refinement. In combination with the methods guaranteeing preciseness of abstractions, this allows to significantly extend the class of models for which safety can be established, covering in particular models with 23 continuous variables and 2 to the 71 discrete states, 20 continuous variables and 2 to the 199 discrete states, and 9 continuous variables and 2 to the 271 discrete states. %B AVACS Technical Report %N 103 %@ false %U http://www.avacs.org/fileadmin/Publikationen/Open/avacs_technical_report_103.pdf
[123]
A. Antoniadis, N. Barcelo, M. Nugent, K. Pruhs, K. Schewior, and M. Scquizzato, “Chasing Convex Bodies and Functions,” in LATIN 2016: Theoretical Informatics, Ensenada, Mexico, 2016.
Export
BibTeX
@inproceedings{AntoniadisLATIN2016, TITLE = {Chasing Convex Bodies and Functions}, AUTHOR = {Antoniadis, Antonios and Barcelo, Neal and Nugent, Michael and Pruhs, Kirk and Schewior, Kevin and Scquizzato, Michele}, LANGUAGE = {eng}, ISBN = {978-3-662-49528-5}, DOI = {10.1007/978-3-662-49529-2_6}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {LATIN 2016: Theoretical Informatics}, EDITOR = {Kranakis, Evangelos and Navarro, Gonzalo and Ch{\'a}vez, Edgar}, PAGES = {68--81}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9644}, ADDRESS = {Ensenada, Mexico}, }
Endnote
%0 Conference Proceedings %A Antoniadis, Antonios %A Barcelo, Neal %A Nugent, Michael %A Pruhs, Kirk %A Schewior, Kevin %A Scquizzato, Michele %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations External Organizations External Organizations %T Chasing Convex Bodies and Functions : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-48D5-E %R 10.1007/978-3-662-49529-2_6 %D 2016 %B 12th Latin American Theoretical Informatics Symposium %Z date of event: 2016-04-11 - 2016-04-15 %C Ensenada, Mexico %B LATIN 2016: Theoretical Informatics %E Kranakis, Evangelos; Navarro, Gonzalo; Ch&#225;vez, Edgar %P 68 - 81 %I Springer %@ 978-3-662-49528-5 %B Lecture Notes in Computer Science %N 9644
[124]
J. Babu, M. Basavaraju, L. S. Chandran, and M. C. Francis, “On Induced Colourful Paths in Triangle-free Graphs,” 2016. [Online]. Available: http://arxiv.org/abs/1604.06070. (arXiv: 1604.06070)
Abstract
Given a graph $G=(V,E)$ whose vertices have been properly coloured, we say that a path in $G$ is "colourful" if no two vertices in the path have the same colour. It is a corollary of the Gallai-Roy Theorem that every properly coloured graph contains a colourful path on $\chi(G)$ vertices. It is interesting to think of what analogous result one could obtain if one considers induced colourful paths instead of just colourful paths. We explore a conjecture that states that every properly coloured triangle-free graph $G$ contains an induced colourful path on $\chi(G)$ vertices. As proving this conjecture in its fullest generality seems to be difficult, we study a special case of the conjecture. We show that the conjecture is true when the girth of $G$ is equal to $\chi(G)$. Even this special case of the conjecture does not seem to have an easy proof: our method involves a detailed analysis of a special kind of greedy colouring algorithm. This result settles the conjecture for every properly coloured triangle-free graph $G$ with girth at least $\chi(G)$.
Export
BibTeX
@online{BBCF2016, TITLE = {On Induced Colourful Paths in Triangle-free Graphs}, AUTHOR = {Babu, Jasine and Basavaraju, Manu and Chandran, L. Sunil and Francis, Mathew C.}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1604.06070}, EPRINT = {1604.06070}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Given a graph $G=(V,E)$ whose vertices have been properly coloured, we say that a path in $G$ is "colourful" if no two vertices in the path have the same colour. It is a corollary of the Gallai-Roy Theorem that every properly coloured graph contains a colourful path on $\chi(G)$ vertices. It is interesting to think of what analogous result one could obtain if one considers induced colourful paths instead of just colourful paths. We explore a conjecture that states that every properly coloured triangle-free graph $G$ contains an induced colourful path on $\chi(G)$ vertices. As proving this conjecture in its fullest generality seems to be difficult, we study a special case of the conjecture. We show that the conjecture is true when the girth of $G$ is equal to $\chi(G)$. Even this special case of the conjecture does not seem to have an easy proof: our method involves a detailed analysis of a special kind of greedy colouring algorithm. This result settles the conjecture for every properly coloured triangle-free graph $G$ with girth at least $\chi(G)$.}, }
Endnote
%0 Report %A Babu, Jasine %A Basavaraju, Manu %A Chandran, L. Sunil %A Francis, Mathew C. %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On Induced Colourful Paths in Triangle-free Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-6134-C %U http://arxiv.org/abs/1604.06070 %D 2016 %X Given a graph $G=(V,E)$ whose vertices have been properly coloured, we say that a path in $G$ is "colourful" if no two vertices in the path have the same colour. It is a corollary of the Gallai-Roy Theorem that every properly coloured graph contains a colourful path on $\chi(G)$ vertices. It is interesting to think of what analogous result one could obtain if one considers induced colourful paths instead of just colourful paths. We explore a conjecture that states that every properly coloured triangle-free graph $G$ contains an induced colourful path on $\chi(G)$ vertices. As proving this conjecture in its fullest generality seems to be difficult, we study a special case of the conjecture. We show that the conjecture is true when the girth of $G$ is equal to $\chi(G)$. Even this special case of the conjecture does not seem to have an easy proof: our method involves a detailed analysis of a special kind of greedy colouring algorithm. This result settles the conjecture for every properly coloured triangle-free graph $G$ with girth at least $\chi(G)$. %K Mathematics, Combinatorics, math.CO,
[125]
R. Becker, M. Sagraloff, V. Sharma, J. Xu, and C. Yap, “Complexity Analysis of Root Clustering for a Complex Polynomial,” in ISSAC 2016, 41st International Symposium on Symbolic and Algebraic Computation, Waterloo, Canada, 2016.
Export
BibTeX
@inproceedings{BeckerISSAC2016, TITLE = {Complexity Analysis of Root Clustering for a Complex Polynomial}, AUTHOR = {Becker, Ruben and Sagraloff, Michael and Sharma, Vikram and Xu, Juan and Yap, Chee}, LANGUAGE = {eng}, ISBN = {978-1-4503-4380-0}, DOI = {10.1145/2930889.2930939}, PUBLISHER = {ACM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {ISSAC 2016, 41st International Symposium on Symbolic and Algebraic Computation}, EDITOR = {Rosenkranz, Markus}, PAGES = {71--78}, ADDRESS = {Waterloo, Canada}, }
Endnote
%0 Conference Proceedings %A Becker, Ruben %A Sagraloff, Michael %A Sharma, Vikram %A Xu, Juan %A Yap, Chee %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Complexity Analysis of Root Clustering for a Complex Polynomial : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-02C0-E %R 10.1145/2930889.2930939 %D 2016 %B 41st International Symposium on Symbolic and Algebraic Computation %Z date of event: 2016-06-19 - 2016-06-22 %C Waterloo, Canada %B ISSAC 2016 %E Rosenkranz, Markus %P 71 - 78 %I ACM %@ 978-1-4503-4380-0
[126]
R. Becker, M. Sagraloff, V. Sharma, and C. Yap, “A Simple Near-Optimal Subdivision Algorithm for Complex Root Isolation based on the Pellet Test and Newton Iteration,” 2016. [Online]. Available: http://arxiv.org/abs/1509.06231. (arXiv: 1509.06231)
Abstract
We describe a subdivision algorithm for isolating the complex roots of a polynomial $F\in\mathbb{C}[x]$. Our model assumes that each coefficient of $F$ has an oracle to return an approximation to any absolute error bound. Given any box $\mathcal{B}$ in the complex plane containing only simple roots of $F$, our algorithm returns disjoint isolating disks for the roots in $\mathcal{B}$. Our complexity analysis bounds the absolute error to which the coefficients of $F$ have to be provided, the total number of iterations, and the overall bit complexity. This analysis shows that the complexity of our algorithm is controlled by the geometry of the roots in a near neighborhood of the input box $\mathcal{B}$, namely, the number of roots and their pairwise distances. The number of subdivision steps is near-optimal. For the \emph{benchmark problem}, namely, to isolate all the roots of an integer polynomial of degree $n$ with coefficients of bitsize less than $\tau$, our algorithm needs $\tilde{O}(n^3+n^2\tau)$ bit operations, which is comparable to the record bound of Pan (2002). It is the first time that such a bound has been achieved using subdivision methods, and independent of divide-and-conquer techniques such as Sch\"onhage's splitting circle technique. Our algorithm uses the quadtree construction of Weyl (1924) with two key ingredients: using Pellet's Theorem (1881) combined with Graeffe iteration, we derive a soft test to count the number of roots in a disk. Using Newton iteration combined with bisection, in a form inspired by the quadratic interval method from Abbot (2006), we achieve quadratic convergence towards root clusters. Relative to the divide-conquer algorithms, our algorithm is simple with the potential of being practical. This paper is self-contained: we provide pseudo-code for all subroutines used by our algorithm.
Export
BibTeX
@online{BeckerarXiv2016, TITLE = {A Simple Near-Optimal Subdivision Algorithm for Complex Root Isolation based on the Pellet Test and Newton Iteration}, AUTHOR = {Becker, Ruben and Sagraloff, Michael and Sharma, Vikram and Yap, Chee}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1509.06231}, EPRINT = {1509.06231}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We describe a subdivision algorithm for isolating the complex roots of a polynomial $F\in\mathbb{C}[x]$. Our model assumes that each coefficient of $F$ has an oracle to return an approximation to any absolute error bound. Given any box $\mathcal{B}$ in the complex plane containing only simple roots of $F$, our algorithm returns disjoint isolating disks for the roots in $\mathcal{B}$. Our complexity analysis bounds the absolute error to which the coefficients of $F$ have to be provided, the total number of iterations, and the overall bit complexity. This analysis shows that the complexity of our algorithm is controlled by the geometry of the roots in a near neighborhood of the input box $\mathcal{B}$, namely, the number of roots and their pairwise distances. The number of subdivision steps is near-optimal. For the \emph{benchmark problem}, namely, to isolate all the roots of an integer polynomial of degree $n$ with coefficients of bitsize less than $\tau$, our algorithm needs $\tilde{O}(n^3+n^2\tau)$ bit operations, which is comparable to the record bound of Pan (2002). It is the first time that such a bound has been achieved using subdivision methods, and independent of divide-and-conquer techniques such as Sch\"onhage's splitting circle technique. Our algorithm uses the quadtree construction of Weyl (1924) with two key ingredients: using Pellet's Theorem (1881) combined with Graeffe iteration, we derive a soft test to count the number of roots in a disk. Using Newton iteration combined with bisection, in a form inspired by the quadratic interval method from Abbot (2006), we achieve quadratic convergence towards root clusters. Relative to the divide-conquer algorithms, our algorithm is simple with the potential of being practical. This paper is self-contained: we provide pseudo-code for all subroutines used by our algorithm.}, }
Endnote
%0 Report %A Becker, Ruben %A Sagraloff, Michael %A Sharma, Vikram %A Yap, Chee %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T A Simple Near-Optimal Subdivision Algorithm for Complex Root Isolation based on the Pellet Test and Newton Iteration : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-02B8-2 %U http://arxiv.org/abs/1509.06231 %D 2016 %X We describe a subdivision algorithm for isolating the complex roots of a polynomial $F\in\mathbb{C}[x]$. Our model assumes that each coefficient of $F$ has an oracle to return an approximation to any absolute error bound. Given any box $\mathcal{B}$ in the complex plane containing only simple roots of $F$, our algorithm returns disjoint isolating disks for the roots in $\mathcal{B}$. Our complexity analysis bounds the absolute error to which the coefficients of $F$ have to be provided, the total number of iterations, and the overall bit complexity. This analysis shows that the complexity of our algorithm is controlled by the geometry of the roots in a near neighborhood of the input box $\mathcal{B}$, namely, the number of roots and their pairwise distances. The number of subdivision steps is near-optimal. For the \emph{benchmark problem}, namely, to isolate all the roots of an integer polynomial of degree $n$ with coefficients of bitsize less than $\tau$, our algorithm needs $\tilde{O}(n^3+n^2\tau)$ bit operations, which is comparable to the record bound of Pan (2002). It is the first time that such a bound has been achieved using subdivision methods, and independent of divide-and-conquer techniques such as Sch\"onhage's splitting circle technique. Our algorithm uses the quadtree construction of Weyl (1924) with two key ingredients: using Pellet's Theorem (1881) combined with Graeffe iteration, we derive a soft test to count the number of roots in a disk. Using Newton iteration combined with bisection, in a form inspired by the quadratic interval method from Abbot (2006), we achieve quadratic convergence towards root clusters. Relative to the divide-conquer algorithms, our algorithm is simple with the potential of being practical. This paper is self-contained: we provide pseudo-code for all subroutines used by our algorithm. %K Computer Science, Numerical Analysis, cs.NA,Computer Science, Symbolic Computation, cs.SC,Mathematics, Numerical Analysis, math.NA
[127]
R. Becker, A. Karrenbauer, S. Krinninger, and C. Lenzen, “Near-Optimal Approximate Shortest Paths and Transshipment in Distributed and Streaming Models,” 2016. [Online]. Available: http://arxiv.org/abs/1607.05127. (arXiv: 1607.05127)
Abstract
We present a method for solving the transshipment problem - also known as uncapacitated minimum cost flow - up to a multiplicative error of $1 + \epsilon$ in undirected graphs with polynomially bounded integer edge weights using a tailored gradient descent algorithm. An important special case of the transshipment problem is the single-source shortest paths (SSSP) problem. Our gradient descent algorithm takes $O(\epsilon^{-3} \mathrm{polylog} n)$ iterations and in each iteration it needs to solve a variant of the transshipment problem up to a multiplicative error of $\mathrm{polylog} n$. In particular, this allows us to perform a single iteration by computing a solution on a sparse spanner of logarithmic stretch. As a consequence, we improve prior work by obtaining the following results: (1) RAM model: $(1+\epsilon)$-approximate transshipment in $\tilde{O}(\epsilon^{-3}(m + n^{1 + o(1)}))$ computational steps (leveraging a recent $O(m^{1+o(1)})$-step $O(1)$-approximation due to Sherman [2016]). (2) Multipass Streaming model: $(1 + \epsilon)$-approximate transshipment and SSSP using $\tilde{O}(n) $ space and $\tilde{O}(\epsilon^{-O(1)})$ passes. (3) Broadcast Congested Clique model: $(1 + \epsilon)$-approximate transshipment and SSSP using $\tilde{O}(\epsilon^{-O(1)})$ rounds. (4) Broadcast Congest model: $(1 + \epsilon)$-approximate SSSP using $\tilde{O}(\epsilon^{-O(1)}(\sqrt{n} + D))$ rounds, where $ D $ is the (hop) diameter of the network. The previous fastest algorithms for the last three models above leverage sparse hop sets. We bypass the hop set computation; using a spanner is sufficient in our method. The above bounds assume non-negative integer edge weights that are polynomially bounded in $n$; for general non-negative weights, running times scale with the logarithm of the maximum ratio between non-zero weights.
Export
BibTeX
@online{Becker_arXiv1607.05127, TITLE = {Near-Optimal Approximate Shortest Paths and Transshipment in Distributed and Streaming Models}, AUTHOR = {Becker, Ruben and Karrenbauer, Andreas and Krinninger, Sebastian and Lenzen, Christoph}, URL = {http://arxiv.org/abs/1607.05127}, EPRINT = {1607.05127}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We present a method for solving the transshipment problem -- also known as uncapacitated minimum cost flow -- up to a multiplicative error of $1 + \epsilon$ in undirected graphs with polynomially bounded integer edge weights using a tailored gradient descent algorithm. An important special case of the transshipment problem is the single-source shortest paths (SSSP) problem. Our gradient descent algorithm takes $O(\epsilon^{-3} \mathrm{polylog} n)$ iterations and in each iteration it needs to solve a variant of the transshipment problem up to a multiplicative error of $\mathrm{polylog} n$. In particular, this allows us to perform a single iteration by computing a solution on a sparse spanner of logarithmic stretch. As a consequence, we improve prior work by obtaining the following results: (1) RAM model: $(1+\epsilon)$-approximate transshipment in $\tilde{O}(\epsilon^{-3}(m + n^{1 + o(1)}))$ computational steps (leveraging a recent $O(m^{1+o(1)})$-step $O(1)$-approximation due to Sherman [2016]). (2) Multipass Streaming model: $(1 + \epsilon)$-approximate transshipment and SSSP using $\tilde{O}(n) $ space and $\tilde{O}(\epsilon^{-O(1)})$ passes. (3) Broadcast Congested Clique model: $(1 + \epsilon)$-approximate transshipment and SSSP using $\tilde{O}(\epsilon^{-O(1)})$ rounds. (4) Broadcast Congest model: $(1 + \epsilon)$-approximate SSSP using $\tilde{O}(\epsilon^{-O(1)}(\sqrt{n} + D))$ rounds, where $ D $ is the (hop) diameter of the network. The previous fastest algorithms for the last three models above leverage sparse hop sets. We bypass the hop set computation; using a spanner is sufficient in our method. The above bounds assume non-negative integer edge weights that are polynomially bounded in $n$; for general non-negative weights, running times scale with the logarithm of the maximum ratio between non-zero weights.}, }
Endnote
%0 Report %A Becker, Ruben %A Karrenbauer, Andreas %A Krinninger, Sebastian %A Lenzen, Christoph %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Near-Optimal Approximate Shortest Paths and Transshipment in Distributed and Streaming Models : %U http://hdl.handle.net/11858/00-001M-0000-002B-8419-1 %U http://arxiv.org/abs/1607.05127 %D 2016 %X We present a method for solving the transshipment problem - also known as uncapacitated minimum cost flow - up to a multiplicative error of $1 + \epsilon$ in undirected graphs with polynomially bounded integer edge weights using a tailored gradient descent algorithm. An important special case of the transshipment problem is the single-source shortest paths (SSSP) problem. Our gradient descent algorithm takes $O(\epsilon^{-3} \mathrm{polylog} n)$ iterations and in each iteration it needs to solve a variant of the transshipment problem up to a multiplicative error of $\mathrm{polylog} n$. In particular, this allows us to perform a single iteration by computing a solution on a sparse spanner of logarithmic stretch. As a consequence, we improve prior work by obtaining the following results: (1) RAM model: $(1+\epsilon)$-approximate transshipment in $\tilde{O}(\epsilon^{-3}(m + n^{1 + o(1)}))$ computational steps (leveraging a recent $O(m^{1+o(1)})$-step $O(1)$-approximation due to Sherman [2016]). (2) Multipass Streaming model: $(1 + \epsilon)$-approximate transshipment and SSSP using $\tilde{O}(n) $ space and $\tilde{O}(\epsilon^{-O(1)})$ passes. (3) Broadcast Congested Clique model: $(1 + \epsilon)$-approximate transshipment and SSSP using $\tilde{O}(\epsilon^{-O(1)})$ rounds. (4) Broadcast Congest model: $(1 + \epsilon)$-approximate SSSP using $\tilde{O}(\epsilon^{-O(1)}(\sqrt{n} + D))$ rounds, where $ D $ is the (hop) diameter of the network. The previous fastest algorithms for the last three models above leverage sparse hop sets. We bypass the hop set computation; using a spanner is sufficient in our method. The above bounds assume non-negative integer edge weights that are polynomially bounded in $n$; for general non-negative weights, running times scale with the logarithm of the maximum ratio between non-zero weights. %K Computer Science, Data Structures and Algorithms, cs.DS
[128]
R. Becker, M. Fickert, and A. Karrenbauer, “A Novel Dual Ascent Algorithm for Solving the Min-Cost Flow Problem,” in Proceedings of the Eighteenth Workshop on Algorithm Engineering and Experiments (ALENEX 2016), Arlington, VA, USA, 2016.
Export
BibTeX
@inproceedings{BeckerALENEX2016, TITLE = {A Novel Dual Ascent Algorithm for Solving the Min-Cost Flow Problem}, AUTHOR = {Becker, Ruben and Fickert, Maximilian and Karrenbauer, Andreas}, LANGUAGE = {eng}, ISBN = {978-1-61197-431-7}, DOI = {10.1137/1.9781611974317.13}, PUBLISHER = {SIAM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {Proceedings of the Eighteenth Workshop on Algorithm Engineering and Experiments (ALENEX 2016)}, EDITOR = {Goodrich, Michael and Mitzenmacher, Michael}, PAGES = {151--159}, ADDRESS = {Arlington, VA, USA}, }
Endnote
%0 Conference Proceedings %A Becker, Ruben %A Fickert, Maximilian %A Karrenbauer, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Novel Dual Ascent Algorithm for Solving the Min-Cost Flow Problem : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0029-4AD0-6 %R 10.1137/1.9781611974317.13 %D 2016 %B Eighteenth Workshop on Algorithm Engineering and Experiments %Z date of event: 2016-01-10 - 2016-01-10 %C Arlington, VA, USA %B Proceedings of the Eighteenth Workshop on Algorithm Engineering and Experiments %E Goodrich, Michael; Mitzenmacher, Michael %P 151 - 159 %I SIAM %@ 978-1-61197-431-7
[129]
R. Becker, A. Karrenbauer, and K. Mehlhorn, “An Integer Interior Point Method for Min-Cost Flow Using Arc Contractions and Deletions,” 2016. [Online]. Available: http://arxiv.org/abs/1612.04689. (arXiv: 1612.04689)
Abstract
We present an interior point method for the min-cost flow problem that uses arc contractions and deletions to steer clear from the boundary of the polytope when path-following methods come too close. We obtain a randomized algorithm running in expected $\tilde O( m^{3/2} )$ time that only visits integer lattice points in the vicinity of the central path of the polytope. This enables us to use integer arithmetic like classical combinatorial algorithms typically do. We provide explicit bounds on the size of the numbers that appear during all computations. By presenting an integer arithmetic interior point algorithm we avoid the tediousness of floating point error analysis and achieve a method that is guaranteed to be free of any numerical issues. We thereby eliminate one of the drawbacks of numerical methods in contrast to combinatorial min-cost flow algorithms that still yield the most efficient implementations in practice, despite their inferior worst-case time complexity.
Export
BibTeX
@online{DBLP:journals/corr/BeckerKM16, TITLE = {An Integer Interior Point Method for Min-Cost Flow Using Arc Contractions and Deletions}, AUTHOR = {Becker, Ruben and Karrenbauer, Andreas and Mehlhorn, Kurt}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1612.04689}, EPRINT = {1612.04689}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We present an interior point method for the min-cost flow problem that uses arc contractions and deletions to steer clear from the boundary of the polytope when path-following methods come too close. We obtain a randomized algorithm running in expected $\tilde O( m^{3/2} )$ time that only visits integer lattice points in the vicinity of the central path of the polytope. This enables us to use integer arithmetic like classical combinatorial algorithms typically do. We provide explicit bounds on the size of the numbers that appear during all computations. By presenting an integer arithmetic interior point algorithm we avoid the tediousness of floating point error analysis and achieve a method that is guaranteed to be free of any numerical issues. We thereby eliminate one of the drawbacks of numerical methods in contrast to combinatorial min-cost flow algorithms that still yield the most efficient implementations in practice, despite their inferior worst-case time complexity.}, }
Endnote
%0 Report %A Becker, Ruben %A Karrenbauer, Andreas %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T An Integer Interior Point Method for Min-Cost Flow Using Arc Contractions and Deletions : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5714-E %U http://arxiv.org/abs/1612.04689 %D 2016 %X We present an interior point method for the min-cost flow problem that uses arc contractions and deletions to steer clear from the boundary of the polytope when path-following methods come too close. We obtain a randomized algorithm running in expected $\tilde O( m^{3/2} )$ time that only visits integer lattice points in the vicinity of the central path of the polytope. This enables us to use integer arithmetic like classical combinatorial algorithms typically do. We provide explicit bounds on the size of the numbers that appear during all computations. By presenting an integer arithmetic interior point algorithm we avoid the tediousness of floating point error analysis and achieve a method that is guaranteed to be free of any numerical issues. We thereby eliminate one of the drawbacks of numerical methods in contrast to combinatorial min-cost flow algorithms that still yield the most efficient implementations in practice, despite their inferior worst-case time complexity. %K Computer Science, Data Structures and Algorithms, cs.DS,Mathematics, Numerical Analysis, math.NA,Mathematics, Optimization and Control, math.OC
[130]
X. Bei, J. Garg, and M. Hoefer, “Ascending-Price Algorithms for Unknown Markets,” in EC’16, ACM Conference on Economics and Computation, Maastricht, The Netherlands, 2016.
Export
BibTeX
@inproceedings{BeiEC2016a, TITLE = {Ascending-Price Algorithms for Unknown Markets}, AUTHOR = {Bei, Xiaohui and Garg, Jugal and Hoefer, Martin}, LANGUAGE = {eng}, ISBN = {978-1-4503-3936-0}, DOI = {10.1145/2940716.2940765}, PUBLISHER = {ACM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {EC'16, ACM Conference on Economics and Computation}, PAGES = {699--699}, ADDRESS = {Maastricht, The Netherlands}, }
Endnote
%0 Conference Proceedings %A Bei, Xiaohui %A Garg, Jugal %A Hoefer, Martin %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Ascending-Price Algorithms for Unknown Markets : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-841F-6 %R 10.1145/2940716.2940765 %D 2016 %B ACM Conference on Economics and Computation %Z date of event: 2016-07-24 - 2016-07-28 %C Maastricht, The Netherlands %B EC'16 %P 699 - 699 %I ACM %@ 978-1-4503-3936-0
[131]
X. Bei, W. Chen, J. Garg, M. Hoefer, and X. Sun, “Learning Market Parameters Using Aggregate Demand Queries,” in Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 2016.
Export
BibTeX
@inproceedings{BeiAAAI2016, TITLE = {Learning Market Parameters Using Aggregate Demand Queries}, AUTHOR = {Bei, Xiaohui and Chen, Wei and Garg, Jugal and Hoefer, Martin and Sun, Xiaoming}, LANGUAGE = {eng}, ISBN = {978-1-57735-760-5}, PUBLISHER = {AAAI}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence}, PAGES = {404--410}, ADDRESS = {Phoenix, AZ, USA}, }
Endnote
%0 Conference Proceedings %A Bei, Xiaohui %A Chen, Wei %A Garg, Jugal %A Hoefer, Martin %A Sun, Xiaoming %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Learning Market Parameters Using Aggregate Demand Queries : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0029-AC36-C %D 2016 %B Thirtieth AAAI Conference on Artificial Intelligence %Z date of event: 2016-02-12 - 2016-02-17 %C Phoenix, AZ, USA %B Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence %P 404 - 410 %I AAAI %@ 978-1-57735-760-5 %U http://www.aaai.org/ocs/index.php/AAAI/AAAI16/paper/view/12052/11612
[132]
X. Bei, J. Garg, M. Hoefer, and K. Mehlhorn, “Computing Equilibria in Markets with Budget-Additive Utilities,” in 24th Annual European Symposium on Algorithms (ESA 2016), Aarhus, Denmark, 2016.
Export
BibTeX
@inproceedings{BeiESA2016, TITLE = {Computing Equilibria in Markets with Budget-Additive Utilities}, AUTHOR = {Bei, Xiaohui and Garg, Jugal and Hoefer, Martin and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISBN = {978-3-95977-015-6}, URL = {urn:nbn:de:0030-drops-63504}, DOI = {10.4230/LIPIcs.ESA.2016.8}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {24th Annual European Symposium on Algorithms (ESA 2016)}, EDITOR = {Sankowski, Piotr and Zaroliagis, Christos}, PAGES = {1--14}, EID = {8}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {57}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Bei, Xiaohui %A Garg, Jugal %A Hoefer, Martin %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Computing Equilibria in Markets with Budget-Additive Utilities : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-479B-5 %R 10.4230/LIPIcs.ESA.2016.8 %U urn:nbn:de:0030-drops-63504 %D 2016 %B 24th Annual European Symposium on Algorithms %Z date of event: 2016-08-22 - 2016-08-26 %C Aarhus, Denmark %B 24th Annual European Symposium on Algorithms %E Sankowski, Piotr; Zaroliagis, Christos %P 1 - 14 %Z sequence number: 8 %I Schloss Dagstuhl %@ 978-3-95977-015-6 %B Leibniz International Proceedings in Informatics %N 57 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2016/6350/
[133]
X. Bei, J. Garg, M. Hoefer, and K. Mehlhorn, “Computing Equilibria in Markets with Budget-Additive Utilities,” 2016. [Online]. Available: http://arxiv.org/abs/1603.07210. (arXiv: 1603.07210)
Abstract
We present the first analysis of Fisher markets with buyers that have budget-additive utility functions. Budget-additive utilities are elementary concave functions with numerous applications in online adword markets and revenue optimization problems. They extend the standard case of linear utilities and have been studied in a variety of other market models. In contrast to the frequently studied CES utilities, they have a global satiation point which can imply multiple market equilibria with quite different characteristics. Our main result is an efficient combinatorial algorithm to compute a market equilibrium with a Pareto-optimal allocation of goods. It relies on a new descending-price approach and, as a special case, also implies a novel combinatorial algorithm for computing a market equilibrium in linear Fisher markets. We complement these positive results with a number of hardness results for related computational questions. We prove that it is NP-hard to compute a market equilibrium that maximizes social welfare, and it is PPAD-hard to find any market equilibrium with utility functions with separate satiation points for each buyer and each good.
Export
BibTeX
@online{BeiGargHoeferMehlhorn2016, TITLE = {Computing Equilibria in Markets with Budget-Additive Utilities}, AUTHOR = {Bei, Xiaohui and Garg, Jugal and Hoefer, Martin and Mehlhorn, Kurt}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1603.07210}, EPRINT = {1603.07210}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We present the first analysis of Fisher markets with buyers that have budget-additive utility functions. Budget-additive utilities are elementary concave functions with numerous applications in online adword markets and revenue optimization problems. They extend the standard case of linear utilities and have been studied in a variety of other market models. In contrast to the frequently studied CES utilities, they have a global satiation point which can imply multiple market equilibria with quite different characteristics. Our main result is an efficient combinatorial algorithm to compute a market equilibrium with a Pareto-optimal allocation of goods. It relies on a new descending-price approach and, as a special case, also implies a novel combinatorial algorithm for computing a market equilibrium in linear Fisher markets. We complement these positive results with a number of hardness results for related computational questions. We prove that it is NP-hard to compute a market equilibrium that maximizes social welfare, and it is PPAD-hard to find any market equilibrium with utility functions with separate satiation points for each buyer and each good.}, }
Endnote
%0 Report %A Bei, Xiaohui %A Garg, Jugal %A Hoefer, Martin %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Computing Equilibria in Markets with Budget-Additive Utilities : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-FCC0-C %U http://arxiv.org/abs/1603.07210 %D 2016 %X We present the first analysis of Fisher markets with buyers that have budget-additive utility functions. Budget-additive utilities are elementary concave functions with numerous applications in online adword markets and revenue optimization problems. They extend the standard case of linear utilities and have been studied in a variety of other market models. In contrast to the frequently studied CES utilities, they have a global satiation point which can imply multiple market equilibria with quite different characteristics. Our main result is an efficient combinatorial algorithm to compute a market equilibrium with a Pareto-optimal allocation of goods. It relies on a new descending-price approach and, as a special case, also implies a novel combinatorial algorithm for computing a market equilibrium in linear Fisher markets. We complement these positive results with a number of hardness results for related computational questions. We prove that it is NP-hard to compute a market equilibrium that maximizes social welfare, and it is PPAD-hard to find any market equilibrium with utility functions with separate satiation points for each buyer and each good. %K Computer Science, Computer Science and Game Theory, cs.GT,Computer Science, Data Structures and Algorithms, cs.DS
[134]
A. Bishnu, K. Dutta, A. Ghosh, and S. Paul, “(1,j)-set Problem in Graphs,” Discrete Mathematics, vol. 339, no. 10, 2016.
Export
BibTeX
@article{DBLP:journals/dm/BishnuDGP16, TITLE = {$(1,j)$-Set Problem in Graphs}, AUTHOR = {Bishnu, Arijit and Dutta, Kunal and Ghosh, Arijit and Paul, Subhabrata}, LANGUAGE = {eng}, ISSN = {0012-365X}, DOI = {10.1016/j.disc.2016.04.008}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Discrete Mathematics}, VOLUME = {339}, NUMBER = {10}, PAGES = {2515--2525}, }
Endnote
%0 Journal Article %A Bishnu, Arijit %A Dutta, Kunal %A Ghosh, Arijit %A Paul, Subhabrata %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T (1,j)-set Problem in Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-65B7-2 %R 10.1016/j.disc.2016.04.008 %7 2014-10-12 %D 2016 %J Discrete Mathematics %V 339 %N 10 %& 2515 %P 2515 - 2525 %I Elsevier %C Amsterdam %@ false
[135]
M. Bläser, G. Jindal, and A. Pandey, “Greedy Strikes Again: A Deterministic PTAS for Commutative Rank of Matrix Spaces,” Electronic Colloquium on Computational Complexity (ECCC): Report Series, vol. 145, 2016.
Export
BibTeX
@article{DBLP:journals/eccc/BlaserJP16, TITLE = {Greedy Strikes Again: {A} Deterministic {PTAS} for Commutative Rank of Matrix Spaces}, AUTHOR = {Bl{\"a}ser, Markus and Jindal, Gorav and Pandey, Anurag}, LANGUAGE = {eng}, ISSN = {1433-8092}, PUBLISHER = {Hasso-Plattner-Institut f{\"u}r Softwaretechnik GmbH}, ADDRESS = {Potsdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, JOURNAL = {Electronic Colloquium on Computational Complexity (ECCC): Report Series}, VOLUME = {145}, PAGES = {1--12}, }
Endnote
%0 Journal Article %A Bl&#228;ser, Markus %A Jindal, Gorav %A Pandey, Anurag %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Greedy Strikes Again: A Deterministic PTAS for Commutative Rank of Matrix Spaces : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E5A-A %7 2016 %D 2016 %J Electronic Colloquium on Computational Complexity (ECCC): Report Series %V 145 %& 1 %P 1 - 12 %I Hasso-Plattner-Institut f&#252;r Softwaretechnik GmbH %C Potsdam %@ false %U https://eccc.weizmann.ac.il/report/2016/145/
[136]
G. Bodwin and S. Krinninger, “Fully Dynamic Spanners with Worst-Case Update Time,” 2016. [Online]. Available: http://arxiv.org/abs/1606.07864. (arXiv: 1606.07864)
Abstract
An $\alpha$-spanner of a graph $ G $ is a subgraph $ H $ such that $ H $ preserves all distances of $ G $ within a factor of $ \alpha $. In this paper, we give fully dynamic algorithms for maintaining a spanner $ H $ of a graph $ G $ undergoing edge insertions and deletions with worst-case guarantees on the running time after each update. In particular, our algorithms maintain: (1) a $3$-spanner with $ \tilde O (n^{1+1/2}) $ edges with worst-case update time $ \tilde O (n^{3/4}) $, or (2) a $5$-spanner with $ \tilde O (n^{1+1/3}) $ edges with worst-case update time $ \tilde O (n^{5/9}) $. These size/stretch tradeoffs are best possible (up to logarithmic factors). They can be extended to the weighted setting at very minor cost. Our algorithms are randomized and correct with high probability against an oblivious adversary. We also further extend our techniques to construct a $5$-spanner with suboptimal size/stretch tradeoff, but improved worst-case update time. To the best of our knowledge, these are the first dynamic spanner algorithms with sublinear worst-case update time guarantees. Since it is known how to maintain a spanner using small amortized but large worst-case update time [Baswana et al. SODA'08], obtaining algorithms with strong worst-case bounds, as presented in this paper, seems to be the next natural step for this problem.
Export
BibTeX
@online{BodwinK2016, TITLE = {Fully Dynamic Spanners with Worst-Case Update Time}, AUTHOR = {Bodwin, Greg and Krinninger, Sebastian}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1606.07864}, EPRINT = {1606.07864}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {An $\alpha$-spanner of a graph $ G $ is a subgraph $ H $ such that $ H $ preserves all distances of $ G $ within a factor of $ \alpha $. In this paper, we give fully dynamic algorithms for maintaining a spanner $ H $ of a graph $ G $ undergoing edge insertions and deletions with worst-case guarantees on the running time after each update. In particular, our algorithms maintain: (1) a $3$-spanner with $ \tilde O (n^{1+1/2}) $ edges with worst-case update time $ \tilde O (n^{3/4}) $, or (2) a $5$-spanner with $ \tilde O (n^{1+1/3}) $ edges with worst-case update time $ \tilde O (n^{5/9}) $. These size/stretch tradeoffs are best possible (up to logarithmic factors). They can be extended to the weighted setting at very minor cost. Our algorithms are randomized and correct with high probability against an oblivious adversary. We also further extend our techniques to construct a $5$-spanner with suboptimal size/stretch tradeoff, but improved worst-case update time. To the best of our knowledge, these are the first dynamic spanner algorithms with sublinear worst-case update time guarantees. Since it is known how to maintain a spanner using small amortized but large worst-case update time [Baswana et al. SODA'08], obtaining algorithms with strong worst-case bounds, as presented in this paper, seems to be the next natural step for this problem.}, }
Endnote
%0 Report %A Bodwin, Greg %A Krinninger, Sebastian %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fully Dynamic Spanners with Worst-Case Update Time : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-50FF-B %U http://arxiv.org/abs/1606.07864 %D 2016 %X An $\alpha$-spanner of a graph $ G $ is a subgraph $ H $ such that $ H $ preserves all distances of $ G $ within a factor of $ \alpha $. In this paper, we give fully dynamic algorithms for maintaining a spanner $ H $ of a graph $ G $ undergoing edge insertions and deletions with worst-case guarantees on the running time after each update. In particular, our algorithms maintain: (1) a $3$-spanner with $ \tilde O (n^{1+1/2}) $ edges with worst-case update time $ \tilde O (n^{3/4}) $, or (2) a $5$-spanner with $ \tilde O (n^{1+1/3}) $ edges with worst-case update time $ \tilde O (n^{5/9}) $. These size/stretch tradeoffs are best possible (up to logarithmic factors). They can be extended to the weighted setting at very minor cost. Our algorithms are randomized and correct with high probability against an oblivious adversary. We also further extend our techniques to construct a $5$-spanner with suboptimal size/stretch tradeoff, but improved worst-case update time. To the best of our knowledge, these are the first dynamic spanner algorithms with sublinear worst-case update time guarantees. Since it is known how to maintain a spanner using small amortized but large worst-case update time [Baswana et al. SODA'08], obtaining algorithms with strong worst-case bounds, as presented in this paper, seems to be the next natural step for this problem. %K Computer Science, Data Structures and Algorithms, cs.DS
[137]
G. Bodwin and S. Krinninger, “Fully Dynamic Spanners with Worst-Case Update Time,” in 24th Annual European Symposium on Algorithms (ESA 2016), Aarhus, Denmark, 2016.
Export
BibTeX
@inproceedings{BodwinK16, TITLE = {Fully Dynamic Spanners with Worst-Case Update Time}, AUTHOR = {Bodwin, Greg and Krinninger, Sebastian}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-015-6}, DOI = {10.4230/LIPIcs.ESA.2016.17}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {24th Annual European Symposium on Algorithms (ESA 2016)}, EDITOR = {Sankowski, Piotr and Zaroliagis, Christos}, PAGES = {1--18}, EID = {17}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {57}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Bodwin, Greg %A Krinninger, Sebastian %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fully Dynamic Spanners with Worst-Case Update Time : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-52CC-D %R 10.4230/LIPIcs.ESA.2016.17 %D 2016 %B 24th Annual European Symposium on Algorithms %Z date of event: 2016-08-22 - 2016-08-26 %C Aarhus, Denmark %B 24th Annual European Symposium on Algorithms %E Sankowski, Piotr; Zaroliagis, Christos %P 1 - 18 %Z sequence number: 17 %I Schloss Dagstuhl %@ 978-3-95977-015-6 %B Leibniz International Proceedings in Informatics %N 57 %@ false %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2016/6368/
[138]
Y. Bouzidi, S. Lazard, G. Moroz, M. Pouget, F. Rouillier, and M. Sagraloff, “Solving Bivariate Systems Using Rational Univariate Representations,” Journal of Complexity, vol. 37, 2016.
Export
BibTeX
@article{Bouzidi2016, TITLE = {Solving bivariate systems using {Rational Univariate Representations}}, AUTHOR = {Bouzidi, Yacine and Lazard, Sylvain and Moroz, Guillaume and Pouget, Marc and Rouillier, Fabrice and Sagraloff, Michael}, LANGUAGE = {eng}, ISSN = {0885-064X}, DOI = {10.1016/j.jco.2016.07.002}, PUBLISHER = {Academic Press}, ADDRESS = {Orlando, Fla.}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Journal of Complexity}, VOLUME = {37}, PAGES = {34--75}, }
Endnote
%0 Journal Article %A Bouzidi, Yacine %A Lazard, Sylvain %A Moroz, Guillaume %A Pouget, Marc %A Rouillier, Fabrice %A Sagraloff, Michael %+ External Organizations External Organizations External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Solving Bivariate Systems Using Rational Univariate Representations : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-841C-C %R 10.1016/j.jco.2016.07.002 %7 2016-07-12 %D 2016 %J Journal of Complexity %V 37 %& 34 %P 34 - 75 %I Academic Press %C Orlando, Fla. %@ false
[139]
C. Brand and M. Sagraloff, “On the Complexity of Solving Zero-Dimensional Polynomial Systems via Projection,” in ISSAC 2016, 41st International Symposium on Symbolic and Algebraic Computation, Waterloo, Canada, 2016.
Export
BibTeX
@inproceedings{BrandISSAC2016, TITLE = {On the Complexity of Solving Zero-Dimensional Polynomial Systems via Projection}, AUTHOR = {Brand, Cornelius and Sagraloff, Michael}, LANGUAGE = {eng}, ISBN = {978-1-4503-4380-0}, DOI = {10.1145/2930889.2930934}, PUBLISHER = {ACM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {ISSAC 2016, 41st International Symposium on Symbolic and Algebraic Computation}, EDITOR = {Rosenkranz, Markus}, PAGES = {151--158}, ADDRESS = {Waterloo, Canada}, }
Endnote
%0 Conference Proceedings %A Brand, Cornelius %A Sagraloff, Michael %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T On the Complexity of Solving Zero-Dimensional Polynomial Systems via Projection : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-02B2-E %R 10.1145/2930889.2930934 %D 2016 %B 41st International Symposium on Symbolic and Algebraic Computation %Z date of event: 2016-06-19 - 2016-06-22 %C Waterloo, Canada %B ISSAC 2016 %E Rosenkranz, Markus %P 151 - 158 %I ACM %@ 978-1-4503-4380-0
[140]
C. Brand and M. Sagraloff, “On the Complexity of Solving Zero-Dimensional Polynomial Systems via Projection,” 2016. [Online]. Available: http://arxiv.org/abs/1604.08944. (arXiv: 1604.08944)
Abstract
Given a zero-dimensional polynomial system consisting of n integer polynomials in n variables, we propose a certified and complete method to compute all complex solutions of the system as well as a corresponding separating linear form l with coefficients of small bit size. For computing l, we need to project the solutions into one dimension along O(n) distinct directions but no further algebraic manipulations. The solutions are then directly reconstructed from the considered projections. The first step is deterministic, whereas the second step uses randomization, thus being Las-Vegas. The theoretical analysis of our approach shows that the overall cost for the two problems considered above is dominated by the cost of carrying out the projections. We also give bounds on the bit complexity of our algorithms that are exclusively stated in terms of the number of variables, the total degree and the bitsize of the input polynomials.
Export
BibTeX
@online{BrandarXiv2016, TITLE = {On the Complexity of Solving Zero-Dimensional Polynomial Systems via Projection}, AUTHOR = {Brand, Cornelius and Sagraloff, Michael}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1604.08944}, EPRINT = {1604.08944}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Given a zero-dimensional polynomial system consisting of n integer polynomials in n variables, we propose a certified and complete method to compute all complex solutions of the system as well as a corresponding separating linear form l with coefficients of small bit size. For computing l, we need to project the solutions into one dimension along O(n) distinct directions but no further algebraic manipulations. The solutions are then directly reconstructed from the considered projections. The first step is deterministic, whereas the second step uses randomization, thus being Las-Vegas. The theoretical analysis of our approach shows that the overall cost for the two problems considered above is dominated by the cost of carrying out the projections. We also give bounds on the bit complexity of our algorithms that are exclusively stated in terms of the number of variables, the total degree and the bitsize of the input polynomials.}, }
Endnote
%0 Report %A Brand, Cornelius %A Sagraloff, Michael %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T On the Complexity of Solving Zero-Dimensional Polynomial Systems via Projection : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-02AF-7 %U http://arxiv.org/abs/1604.08944 %D 2016 %X Given a zero-dimensional polynomial system consisting of n integer polynomials in n variables, we propose a certified and complete method to compute all complex solutions of the system as well as a corresponding separating linear form l with coefficients of small bit size. For computing l, we need to project the solutions into one dimension along O(n) distinct directions but no further algebraic manipulations. The solutions are then directly reconstructed from the considered projections. The first step is deterministic, whereas the second step uses randomization, thus being Las-Vegas. The theoretical analysis of our approach shows that the overall cost for the two problems considered above is dominated by the cost of carrying out the projections. We also give bounds on the bit complexity of our algorithms that are exclusively stated in terms of the number of variables, the total degree and the bitsize of the input polynomials. %K Computer Science, Symbolic Computation, cs.SC,Computer Science, Computational Complexity, cs.CC
[141]
U. Brandes, E. Holm, and A. Karrenbauer, “Cliques in Regular Graphs and the Core-Periphery Problem in Social Networks,” in Combinatorial Optimization and Applications (COCOA 2016), Hong Kong, China, 2016.
Export
BibTeX
@inproceedings{BHK2016, TITLE = {Cliques in Regular Graphs and the Core-Periphery Problem in Social Networks}, AUTHOR = {Brandes, Ulrik and Holm, Eugenia and Karrenbauer, Andreas}, LANGUAGE = {eng}, ISBN = {978-3-319-48748-9}, DOI = {10.1007/978-3-319-48749-6_13}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Combinatorial Optimization and Applications (COCOA 2016)}, EDITOR = {Chan, T-H. Hubert and Li, Minming and Wang, Lusheng}, PAGES = {175--186}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10043}, ADDRESS = {Hong Kong, China}, }
Endnote
%0 Conference Proceedings %A Brandes, Ulrik %A Holm, Eugenia %A Karrenbauer, Andreas %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Cliques in Regular Graphs and the Core-Periphery Problem in Social Networks : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-832D-8 %R 10.1007/978-3-319-48749-6_13 %D 2016 %B 10th Annual International Conference on Combinatorial Optimization and Applications %Z date of event: 2016-12-16 - 2016-12-18 %C Hong Kong, China %B Combinatorial Optimization and Applications %E Chan, T-H. Hubert; Li, Minming; Wang, Lusheng %P 175 - 186 %I Springer %@ 978-3-319-48748-9 %B Lecture Notes in Computer Science %N 10043
[142]
K. Bringmann, F. Grandoni, B. Saha, and V. V. Williams, “Truly Sub-cubic Algorithms for Language Edit Distance and RNA-Folding via Fast Bounded-Difference Min-Plus Product,” in FOCS 2016, New Brunswick, NJ, USA, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/focs/BringmannGSW16, TITLE = {Truly Sub-cubic Algorithms for Language Edit Distance and {RNA}-Folding via Fast Bounded-Difference Min-Plus Product}, AUTHOR = {Bringmann, Karl and Grandoni, Fabrizio and Saha, Barna and Williams, Virginia Vassilevska}, LANGUAGE = {eng}, DOI = {10.1109/FOCS.2016.48}, PUBLISHER = {IEEE}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {FOCS 2016}, PAGES = {375--384}, ADDRESS = {New Brunswick, NJ, USA}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Grandoni, Fabrizio %A Saha, Barna %A Williams, Virginia Vassilevska %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Discrete Optimization, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Truly Sub-cubic Algorithms for Language Edit Distance and RNA-Folding via Fast Bounded-Difference Min-Plus Product : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-53DE-E %R 10.1109/FOCS.2016.48 %D 2016 %B 57th Annual IEEE Symposium on Foundations of Computer Science %Z date of event: 2016-10-09 - 2016-10-11 %C New Brunswick, NJ, USA %B FOCS 2016 %P 375 - 384 %I IEEE
[143]
K. Bringmann, L. Kozma, S. Moran, and N. S. Narayanaswamy, “Hitting Set for Hypergraphs of Low VC-dimension,” in 24th Annual European Symposium on Algorithms (ESA 2016), Aarhus, Denmark, 2016.
Export
BibTeX
@inproceedings{MoranESA2016, TITLE = {Hitting Set for Hypergraphs of Low {VC}-dimension}, AUTHOR = {Bringmann, Karl and Kozma, L{\'a}szl{\'o} and Moran, Shay and Narayanaswamy, N. S.}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-015-6}, URL = {urn:nbn:de:0030-drops-63749}, DOI = {10.4230/LIPIcs.ESA.2016.23}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {24th Annual European Symposium on Algorithms (ESA 2016)}, EDITOR = {Sankowski, Piotr and Zaroliagis, Christos}, PAGES = {1--18}, EID = {23}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {57}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Bringmann, Karl %A Kozma, L&#225;szl&#243; %A Moran, Shay %A Narayanaswamy, N. S. %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Hitting Set for Hypergraphs of Low VC-dimension : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-51F4-A %U urn:nbn:de:0030-drops-63749 %R 10.4230/LIPIcs.ESA.2016.23 %D 2016 %B 24th Annual European Symposium on Algorithms %Z date of event: 2016-08-22 - 2016-08-26 %C Aarhus, Denmark %B 24th Annual European Symposium on Algorithms %E Sankowski, Piotr; Zaroliagis, Christos %P 1 - 18 %Z sequence number: 23 %I Schloss Dagstuhl %@ 978-3-95977-015-6 %B Leibniz International Proceedings in Informatics %N 57 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2016/6374/http://drops.dagstuhl.de/doku/urheberrecht1.html
[144]
K. Bringmann, T. Sauerwald, A. Stauffer, and H. Sun, “Balls into Bins via Local Search: Cover Time and Maximum Load,” Random Structures and Algorithms, vol. 48, no. 4, 2016.
Export
BibTeX
@article{DBLP:journals/rsa/BringmannSSS16, TITLE = {Balls into Bins via Local Search: Cover Time and Maximum Load}, AUTHOR = {Bringmann, Karl and Sauerwald, Thomas and Stauffer, Alexandre and Sun, He}, LANGUAGE = {eng}, ISSN = {1042-9832}, DOI = {10.1002/rsa.20602}, PUBLISHER = {Wiley}, ADDRESS = {New York, N.Y.}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Random Structures and Algorithms}, VOLUME = {48}, NUMBER = {4}, PAGES = {681--702}, }
Endnote
%0 Journal Article %A Bringmann, Karl %A Sauerwald, Thomas %A Stauffer, Alexandre %A Sun, He %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Balls into Bins via Local Search: Cover Time and Maximum Load : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-54D5-7 %R 10.1002/rsa.20602 %7 2016 %D 2016 %J Random Structures and Algorithms %V 48 %N 4 %& 681 %P 681 - 702 %I Wiley %C New York, N.Y. %@ false
[145]
K. Bringmann, A. Grønlund, and K. G. Larsen, “A Dichotomy for Regular Expression Membership Testing,” 2016. [Online]. Available: http://arxiv.org/abs/1611.00918. (arXiv: 1611.00918)
Abstract
We study regular expression membership testing: Given a regular expression of size $m$ and a string of size $n$, decide whether the string is in the language described by the regular expression. Its classic $O(nm)$ algorithm is one of the big success stories of the 70s, which allowed pattern matching to develop into the standard tool that it is today. Many special cases of pattern matching have been studied that can be solved faster than in quadratic time. However, a systematic study of tractable cases was made possible only recently, with the first conditional lower bounds reported by Backurs and Indyk [FOCS'16]. Restricted to any "type" of homogeneous regular expressions of depth 2 or 3, they either presented a near-linear time algorithm or a quadratic conditional lower bound, with one exception known as the Word Break problem. In this paper we complete their work as follows: 1) We present two almost-linear time algorithms that generalize all known almost-linear time algorithms for special cases of regular expression membership testing. 2) We classify all types, except for the Word Break problem, into almost-linear time or quadratic time assuming the Strong Exponential Time Hypothesis. This extends the classification from depth 2 and 3 to any constant depth. 3) For the Word Break problem we give an improved $\tilde{O}(n m^{1/3} + m)$ algorithm. Surprisingly, we also prove a matching conditional lower bound for combinatorial algorithms. This establishes Word Break as the only intermediate problem. In total, we prove matching upper and lower bounds for any type of bounded-depth homogeneous regular expressions, which yields a full dichotomy for regular expression membership testing.
Export
BibTeX
@online{BringmannGL16, TITLE = {A Dichotomy for Regular Expression Membership Testing}, AUTHOR = {Bringmann, Karl and Gr{\o}nlund, Allan and Larsen, Kasper Green}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1611.00918}, EPRINT = {1611.00918}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We study regular expression membership testing: Given a regular expression of size $m$ and a string of size $n$, decide whether the string is in the language described by the regular expression. Its classic $O(nm)$ algorithm is one of the big success stories of the 70s, which allowed pattern matching to develop into the standard tool that it is today. Many special cases of pattern matching have been studied that can be solved faster than in quadratic time. However, a systematic study of tractable cases was made possible only recently, with the first conditional lower bounds reported by Backurs and Indyk [FOCS'16]. Restricted to any "type" of homogeneous regular expressions of depth 2 or 3, they either presented a near-linear time algorithm or a quadratic conditional lower bound, with one exception known as the Word Break problem. In this paper we complete their work as follows: 1) We present two almost-linear time algorithms that generalize all known almost-linear time algorithms for special cases of regular expression membership testing. 2) We classify all types, except for the Word Break problem, into almost-linear time or quadratic time assuming the Strong Exponential Time Hypothesis. This extends the classification from depth 2 and 3 to any constant depth. 3) For the Word Break problem we give an improved $\tilde{O}(n m^{1/3} + m)$ algorithm. Surprisingly, we also prove a matching conditional lower bound for combinatorial algorithms. This establishes Word Break as the only intermediate problem. In total, we prove matching upper and lower bounds for any type of bounded-depth homogeneous regular expressions, which yields a full dichotomy for regular expression membership testing.}, }
Endnote
%0 Report %A Bringmann, Karl %A Gr&#248;nlund, Allan %A Larsen, Kasper Green %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T A Dichotomy for Regular Expression Membership Testing : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5300-F %U http://arxiv.org/abs/1611.00918 %D 2016 %X We study regular expression membership testing: Given a regular expression of size $m$ and a string of size $n$, decide whether the string is in the language described by the regular expression. Its classic $O(nm)$ algorithm is one of the big success stories of the 70s, which allowed pattern matching to develop into the standard tool that it is today. Many special cases of pattern matching have been studied that can be solved faster than in quadratic time. However, a systematic study of tractable cases was made possible only recently, with the first conditional lower bounds reported by Backurs and Indyk [FOCS'16]. Restricted to any "type" of homogeneous regular expressions of depth 2 or 3, they either presented a near-linear time algorithm or a quadratic conditional lower bound, with one exception known as the Word Break problem. In this paper we complete their work as follows: 1) We present two almost-linear time algorithms that generalize all known almost-linear time algorithms for special cases of regular expression membership testing. 2) We classify all types, except for the Word Break problem, into almost-linear time or quadratic time assuming the Strong Exponential Time Hypothesis. This extends the classification from depth 2 and 3 to any constant depth. 3) For the Word Break problem we give an improved $\tilde{O}(n m^{1/3} + m)$ algorithm. Surprisingly, we also prove a matching conditional lower bound for combinatorial algorithms. This establishes Word Break as the only intermediate problem. In total, we prove matching upper and lower bounds for any type of bounded-depth homogeneous regular expressions, which yields a full dichotomy for regular expression membership testing. %K Computer Science, Data Structures and Algorithms, cs.DS,Computer Science, Computational Complexity, cs.CC
[146]
K. Bringmann, R. Keusch, and J. Lengler, “Geometric Inhomogeneous Random Graphs,” 2016. [Online]. Available: http://arxiv.org/abs/1511.00576. (arXiv: 1511.00576)
Abstract
Real-world networks, like social networks or the internet infrastructure, have structural properties such as their large clustering coefficient that can best be described in terms of an underlying geometry. This is why the focus of the literature on theoretical models for real-world networks shifted from classic models without geometry, such as Chung-Lu random graphs, to modern geometry-based models, such as hyperbolic random graphs. With this paper we contribute to the theoretical analysis of these modern, more realistic random graph models. However, we do not directly study hyperbolic random graphs, but replace them by a more general model that we call \emph{geometric inhomogeneous random graphs} (GIRGs). Since we ignore constant factors in the edge probabilities, our model is technically simpler (specifically, we avoid hyperbolic cosines), while preserving the qualitative behaviour of hyperbolic random graphs, and we suggest to replace hyperbolic random graphs by our new model in future theoretical studies. We prove the following fundamental structural and algorithmic results on GIRGs. (1) We provide a sampling algorithm that generates a random graph from our model in expected linear time, improving the best-known sampling algorithm for hyperbolic random graphs by a factor $O(\sqrt{n})$, (2) we establish that GIRGs have a constant clustering coefficient, (3) we show that GIRGs have small separators, i.e., it suffices to delete a sublinear number of edges to break the giant component into two large pieces, and (4) we show how to compress GIRGs using an expected linear number of bits.
Export
BibTeX
@online{Bringmannarxiv16, TITLE = {Geometric Inhomogeneous Random Graphs}, AUTHOR = {Bringmann, Karl and Keusch, Ralph and Lengler, Johannes}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1511.00576}, EPRINT = {1511.00576}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Real-world networks, like social networks or the internet infrastructure, have structural properties such as their large clustering coefficient that can best be described in terms of an underlying geometry. This is why the focus of the literature on theoretical models for real-world networks shifted from classic models without geometry, such as Chung-Lu random graphs, to modern geometry-based models, such as hyperbolic random graphs. With this paper we contribute to the theoretical analysis of these modern, more realistic random graph models. However, we do not directly study hyperbolic random graphs, but replace them by a more general model that we call \emph{geometric inhomogeneous random graphs} (GIRGs). Since we ignore constant factors in the edge probabilities, our model is technically simpler (specifically, we avoid hyperbolic cosines), while preserving the qualitative behaviour of hyperbolic random graphs, and we suggest to replace hyperbolic random graphs by our new model in future theoretical studies. We prove the following fundamental structural and algorithmic results on GIRGs. (1) We provide a sampling algorithm that generates a random graph from our model in expected linear time, improving the best-known sampling algorithm for hyperbolic random graphs by a factor $O(\sqrt{n})$, (2) we establish that GIRGs have a constant clustering coefficient, (3) we show that GIRGs have small separators, i.e., it suffices to delete a sublinear number of edges to break the giant component into two large pieces, and (4) we show how to compress GIRGs using an expected linear number of bits.}, }
Endnote
%0 Report %A Bringmann, Karl %A Keusch, Ralph %A Lengler, Johannes %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Geometric Inhomogeneous Random Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-52F3-4 %U http://arxiv.org/abs/1511.00576 %D 2016 %X Real-world networks, like social networks or the internet infrastructure, have structural properties such as their large clustering coefficient that can best be described in terms of an underlying geometry. This is why the focus of the literature on theoretical models for real-world networks shifted from classic models without geometry, such as Chung-Lu random graphs, to modern geometry-based models, such as hyperbolic random graphs. With this paper we contribute to the theoretical analysis of these modern, more realistic random graph models. However, we do not directly study hyperbolic random graphs, but replace them by a more general model that we call \emph{geometric inhomogeneous random graphs} (GIRGs). Since we ignore constant factors in the edge probabilities, our model is technically simpler (specifically, we avoid hyperbolic cosines), while preserving the qualitative behaviour of hyperbolic random graphs, and we suggest to replace hyperbolic random graphs by our new model in future theoretical studies. We prove the following fundamental structural and algorithmic results on GIRGs. (1) We provide a sampling algorithm that generates a random graph from our model in expected linear time, improving the best-known sampling algorithm for hyperbolic random graphs by a factor $O(\sqrt{n})$, (2) we establish that GIRGs have a constant clustering coefficient, (3) we show that GIRGs have small separators, i.e., it suffices to delete a sublinear number of edges to break the giant component into two large pieces, and (4) we show how to compress GIRGs using an expected linear number of bits. %K cs.SI,Computer Science, Discrete Mathematics, cs.DM,Computer Science, Networking and Internet Architecture, cs.NI
[147]
K. Bringmann, R. Keusch, J. Lengler, Y. Maus, and A. Molla, “Greedy Routing and the Algorithmic Small-World Phenomenom,” 2016. [Online]. Available: http://arxiv.org/abs/1612.05539. (arXiv: 1612.05539)
Abstract
The algorithmic small-world phenomenon, empirically established by Milgram's letter forwarding experiments from the 60s, was theoretically explained by Kleinberg in 2000. However, from today's perspective his model has several severe shortcomings that limit the applicability to real-world networks. In order to give a more convincing explanation of the algorithmic small-world phenomenon, we study greedy routing in a more realistic random graph model (geometric inhomogeneous random graphs), which overcomes the previous shortcomings. Apart from exhibiting good properties in theory, it has also been extensively experimentally validated that this model reasonably captures real-world networks. In this model, we show that greedy routing succeeds with constant probability, and in case of success almost surely finds a path that is an almost shortest path. Our results are robust to changes in the model parameters and the routing objective. Moreover, since constant success probability is too low for technical applications, we study natural local patching methods augmenting greedy routing by backtracking and we show that such methods can ensure success probability 1 in a number of steps that is close to the shortest path length. These results also address the question of Krioukov et al. whether there are efficient local routing protocols for the internet graph. There were promising experimental studies, but the question remained unsolved theoretically. Our results give for the first time a rigorous and analytical answer, assuming our random graph model.
Export
BibTeX
@online{BringmannKLMM16, TITLE = {Greedy Routing and the Algorithmic Small-World Phenomenom}, AUTHOR = {Bringmann, Karl and Keusch, Ralph and Lengler, Johannes and Maus, Yannic and Molla, Anisur}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1612.05539}, EPRINT = {1612.05539}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {The algorithmic small-world phenomenon, empirically established by Milgram's letter forwarding experiments from the 60s, was theoretically explained by Kleinberg in 2000. However, from today's perspective his model has several severe shortcomings that limit the applicability to real-world networks. In order to give a more convincing explanation of the algorithmic small-world phenomenon, we study greedy routing in a more realistic random graph model (geometric inhomogeneous random graphs), which overcomes the previous shortcomings. Apart from exhibiting good properties in theory, it has also been extensively experimentally validated that this model reasonably captures real-world networks. In this model, we show that greedy routing succeeds with constant probability, and in case of success almost surely finds a path that is an almost shortest path. Our results are robust to changes in the model parameters and the routing objective. Moreover, since constant success probability is too low for technical applications, we study natural local patching methods augmenting greedy routing by backtracking and we show that such methods can ensure success probability 1 in a number of steps that is close to the shortest path length. These results also address the question of Krioukov et al. whether there are efficient local routing protocols for the internet graph. There were promising experimental studies, but the question remained unsolved theoretically. Our results give for the first time a rigorous and analytical answer, assuming our random graph model.}, }
Endnote
%0 Report %A Bringmann, Karl %A Keusch, Ralph %A Lengler, Johannes %A Maus, Yannic %A Molla, Anisur %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Greedy Routing and the Algorithmic Small-World Phenomenom : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5309-E %U http://arxiv.org/abs/1612.05539 %D 2016 %X The algorithmic small-world phenomenon, empirically established by Milgram's letter forwarding experiments from the 60s, was theoretically explained by Kleinberg in 2000. However, from today's perspective his model has several severe shortcomings that limit the applicability to real-world networks. In order to give a more convincing explanation of the algorithmic small-world phenomenon, we study greedy routing in a more realistic random graph model (geometric inhomogeneous random graphs), which overcomes the previous shortcomings. Apart from exhibiting good properties in theory, it has also been extensively experimentally validated that this model reasonably captures real-world networks. In this model, we show that greedy routing succeeds with constant probability, and in case of success almost surely finds a path that is an almost shortest path. Our results are robust to changes in the model parameters and the routing objective. Moreover, since constant success probability is too low for technical applications, we study natural local patching methods augmenting greedy routing by backtracking and we show that such methods can ensure success probability 1 in a number of steps that is close to the shortest path length. These results also address the question of Krioukov et al. whether there are efficient local routing protocols for the internet graph. There were promising experimental studies, but the question remained unsolved theoretically. Our results give for the first time a rigorous and analytical answer, assuming our random graph model. %K cs.SI,Computer Science, Discrete Mathematics, cs.DM,Computer Science, Networking and Internet Architecture, cs.NI,
[148]
K. Bringmann, D. Hermelin, M. Mnich, and E. J. van Leeuwen, “Parameterized Complexity Dichotomy for Steiner Multicut,” Journal of Computer and System Sciences, 2016.
Export
BibTeX
@article{DBLP:journals/jcss/BringmannHML16, TITLE = {Parameterized Complexity Dichotomy for {Steiner} {Multicut}}, AUTHOR = {Bringmann, Karl and Hermelin, Danny and Mnich, Matthias and van Leeuwen, Erik Jan}, LANGUAGE = {eng}, ISSN = {0022-0000}, DOI = {10.1016/j.jcss.2016.03.003}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Journal of Computer and System Sciences}, PAGES = {1020--1043}, }
Endnote
%0 Journal Article %A Bringmann, Karl %A Hermelin, Danny %A Mnich, Matthias %A van Leeuwen, Erik Jan %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Parameterized Complexity Dichotomy for Steiner Multicut : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-535A-7 %R 10.1016/j.jcss.2016.03.003 %7 2016 %D 2016 %J Journal of Computer and System Sciences %& 1020 %P 1020 - 1043 %I Elsevier %C Amsterdam %@ false
[149]
K. Bringmann and W. Mulzer, “Approximability of the Discrete Fréchet Distance,” Journal on Computational Geometry, vol. 7, no. 2, 2016.
Export
BibTeX
@article{DBLP:conf/compgeom/BringmannM16, TITLE = {Approximability of the Discrete {Fr\'echet} Distance}, AUTHOR = {Bringmann, Karl and Mulzer, Wolfgang}, LANGUAGE = {eng}, ISSN = {1920-180X}, DOI = {10.20382/jocg.v7i2a4}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, JOURNAL = {Journal on Computational Geometry}, VOLUME = {7}, NUMBER = {2}, PAGES = {46--76}, }
Endnote
%0 Journal Article %A Bringmann, Karl %A Mulzer, Wolfgang %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Approximability of the Discrete Fr&#233;chet Distance : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-550C-7 %R 10.20382/jocg.v7i2a4 %7 2016 %D 2016 %J Journal on Computational Geometry %O JoCG %V 7 %N 2 %& 46 %P 46 - 76 %@ false
[150]
K. Bringmann, R. Keusch, and J. Lengler, “Average Distance in a General Class of Scale-Free Networks with Underlying Geometry,” 2016. [Online]. Available: http://arxiv.org/abs/1602.05712. (arXiv: 1602.05712)
Abstract
In Chung-Lu random graphs, a classic model for real-world networks, each vertex is equipped with a weight drawn from a power-law distribution (for which we fix an exponent $2 < \beta < 3$), and two vertices form an edge independently with probability proportional to the product of their weights. Modern, more realistic variants of this model also equip each vertex with a random position in a specific underlying geometry, which is typically Euclidean, such as the unit square, circle, or torus. The edge probability of two vertices then depends, say, inversely polynomial on their distance. We show that specific choices, such as the underlying geometry being Euclidean or the dependence on the distance being inversely polynomial, do not significantly influence the average distance, by studying a generic augmented version of Chung-Lu random graphs. Specifically, we analyze a model where the edge probability of two vertices can depend arbitrarily on their positions, as long as the marginal probability of forming an edge (for two vertices with fixed weights, one fixed position, and one random position) is as in Chung-Lu random graphs, i.e., proportional to the product of their weights. The resulting class contains Chung-Lu random graphs, hyperbolic random graphs, and geometric inhomogeneous random graphs as special cases. Our main result is that this general model has the same average distance as Chung-Lu random graphs, up to a factor $1+o(1)$. The proof also yields that our model has a giant component and polylogarithmic diameter with high probability.
Export
BibTeX
@online{Bringarx16, TITLE = {Average Distance in a General Class of Scale-Free Networks with Underlying Geometry}, AUTHOR = {Bringmann, Karl and Keusch, Ralph and Lengler, Johannes}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1602.05712}, EPRINT = {1602.05712}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {In Chung-Lu random graphs, a classic model for real-world networks, each vertex is equipped with a weight drawn from a power-law distribution (for which we fix an exponent $2 < \beta < 3$), and two vertices form an edge independently with probability proportional to the product of their weights. Modern, more realistic variants of this model also equip each vertex with a random position in a specific underlying geometry, which is typically Euclidean, such as the unit square, circle, or torus. The edge probability of two vertices then depends, say, inversely polynomial on their distance. We show that specific choices, such as the underlying geometry being Euclidean or the dependence on the distance being inversely polynomial, do not significantly influence the average distance, by studying a generic augmented version of Chung-Lu random graphs. Specifically, we analyze a model where the edge probability of two vertices can depend arbitrarily on their positions, as long as the marginal probability of forming an edge (for two vertices with fixed weights, one fixed position, and one random position) is as in Chung-Lu random graphs, i.e., proportional to the product of their weights. The resulting class contains Chung-Lu random graphs, hyperbolic random graphs, and geometric inhomogeneous random graphs as special cases. Our main result is that this general model has the same average distance as Chung-Lu random graphs, up to a factor $1+o(1)$. The proof also yields that our model has a giant component and polylogarithmic diameter with high probability.}, }
Endnote
%0 Report %A Bringmann, Karl %A Keusch, Ralph %A Lengler, Johannes %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Average Distance in a General Class of Scale-Free Networks with Underlying Geometry : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-52FA-5 %U http://arxiv.org/abs/1602.05712 %D 2016 %X In Chung-Lu random graphs, a classic model for real-world networks, each vertex is equipped with a weight drawn from a power-law distribution (for which we fix an exponent $2 < \beta < 3$), and two vertices form an edge independently with probability proportional to the product of their weights. Modern, more realistic variants of this model also equip each vertex with a random position in a specific underlying geometry, which is typically Euclidean, such as the unit square, circle, or torus. The edge probability of two vertices then depends, say, inversely polynomial on their distance. We show that specific choices, such as the underlying geometry being Euclidean or the dependence on the distance being inversely polynomial, do not significantly influence the average distance, by studying a generic augmented version of Chung-Lu random graphs. Specifically, we analyze a model where the edge probability of two vertices can depend arbitrarily on their positions, as long as the marginal probability of forming an edge (for two vertices with fixed weights, one fixed position, and one random position) is as in Chung-Lu random graphs, i.e., proportional to the product of their weights. The resulting class contains Chung-Lu random graphs, hyperbolic random graphs, and geometric inhomogeneous random graphs as special cases. Our main result is that this general model has the same average distance as Chung-Lu random graphs, up to a factor $1+o(1)$. The proof also yields that our model has a giant component and polylogarithmic diameter with high probability. %K Computer Science, Discrete Mathematics, cs.DM,cs.SI
[151]
J. Case and T. Kötzing, “Topological Separations in Inductive Inference,” Theoretical Computer Science, vol. 260, 2016.
Export
BibTeX
@article{CaseKotzing2016, TITLE = {Topological Separations in Inductive Inference}, AUTHOR = {Case, John and K{\"o}tzing, Timo}, LANGUAGE = {eng}, ISSN = {0304-3975}, DOI = {10.1016/j.tcs.2015.10.036}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Theoretical Computer Science}, VOLUME = {260}, PAGES = {33--45}, }
Endnote
%0 Journal Article %A Case, John %A K&#246;tzing, Timo %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Topological Separations in Inductive Inference : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-1A2D-8 %R 10.1016/j.tcs.2015.10.036 %7 2015-10-30 %D 2016 %J Theoretical Computer Science %V 260 %& 33 %P 33 - 45 %I Elsevier %C Amsterdam %@ false
[152]
K. Censor-Hillel, P. Kaski, J. H. Korhonen, C. Lenzen, A. Paz, and J. Suomela, “Algebraic Methods in the Congested Clique,” Distributed Computing, vol. First Online, 2016.
Export
BibTeX
@article{Censor-Hillel2016, TITLE = {Algebraic Methods in the Congested Clique}, AUTHOR = {Censor-Hillel, Keren and Kaski, Petteri and Korhonen, Janne H. and Lenzen, Christoph and Paz, Ami and Suomela, Jukka}, LANGUAGE = {eng}, ISSN = {0178-2770}, DOI = {10.1007/s00446-016-0270-2}, PUBLISHER = {Springer International}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, JOURNAL = {Distributed Computing}, VOLUME = {First Online}, PAGES = {1--18}, }
Endnote
%0 Journal Article %A Censor-Hillel, Keren %A Kaski, Petteri %A Korhonen, Janne H. %A Lenzen, Christoph %A Paz, Ami %A Suomela, Jukka %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Algebraic Methods in the Congested Clique : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5733-8 %R 10.1007/s00446-016-0270-2 %7 2016 %D 2016 %J Distributed Computing %V First Online %& 1 %P 1 - 18 %I Springer International %C New York, NY %@ false
[153]
P. Chalermsook, M. Goswami, L. Kozma, K. Mehlhorn, and T. Saranurak, “The Landscape of Bounds for Binary Search Trees,” 2016. [Online]. Available: http://arxiv.org/abs/1603.04892. (arXiv: 1603.04892)
Abstract
Binary search trees (BSTs) with rotations can adapt to various kinds of structure in search sequences, achieving amortized access times substantially better than the Theta(log n) worst-case guarantee. Classical examples of structural properties include static optimality, sequential access, working set, key-independent optimality, and dynamic finger, all of which are now known to be achieved by the two famous online BST algorithms (Splay and Greedy). (...) In this paper, we introduce novel properties that explain the efficiency of sequences not captured by any of the previously known properties, and which provide new barriers to the dynamic optimality conjecture. We also establish connections between various properties, old and new. For instance, we show the following. (i) A tight bound of O(n log d) on the cost of Greedy for d-decomposable sequences. The result builds on the recent lazy finger result of Iacono and Langerman (SODA 2016). On the other hand, we show that lazy finger alone cannot explain the efficiency of pattern avoiding sequences even in some of the simplest cases. (ii) A hierarchy of bounds using multiple lazy fingers, addressing a recent question of Iacono and Langerman. (iii) The optimality of the Move-to-root heuristic in the key-independent setting introduced by Iacono (Algorithmica 2005). (iv) A new tool that allows combining any finite number of sound structural properties. As an application, we show an upper bound on the cost of a class of sequences that all known properties fail to capture. (v) The equivalence between two families of BST properties. The observation on which this connection is based was known before - we make it explicit, and apply it to classical BST properties. (...)
Export
BibTeX
@online{Chalermsook2016PP, TITLE = {The Landscape of Bounds for Binary Search Trees}, AUTHOR = {Chalermsook, Parinya and Goswami, Mayank and Kozma, L{\'a}szl{\'o} and Mehlhorn, Kurt and Saranurak, Thatchaphol}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1603.04892}, EPRINT = {1603.04892}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Binary search trees (BSTs) with rotations can adapt to various kinds of structure in search sequences, achieving amortized access times substantially better than the Theta(log n) worst-case guarantee. Classical examples of structural properties include static optimality, sequential access, working set, key-independent optimality, and dynamic finger, all of which are now known to be achieved by the two famous online BST algorithms (Splay and Greedy). (...) In this paper, we introduce novel properties that explain the efficiency of sequences not captured by any of the previously known properties, and which provide new barriers to the dynamic optimality conjecture. We also establish connections between various properties, old and new. For instance, we show the following. (i) A tight bound of O(n log d) on the cost of Greedy for d-decomposable sequences. The result builds on the recent lazy finger result of Iacono and Langerman (SODA 2016). On the other hand, we show that lazy finger alone cannot explain the efficiency of pattern avoiding sequences even in some of the simplest cases. (ii) A hierarchy of bounds using multiple lazy fingers, addressing a recent question of Iacono and Langerman. (iii) The optimality of the Move-to-root heuristic in the key-independent setting introduced by Iacono (Algorithmica 2005). (iv) A new tool that allows combining any finite number of sound structural properties. As an application, we show an upper bound on the cost of a class of sequences that all known properties fail to capture. (v) The equivalence between two families of BST properties. The observation on which this connection is based was known before -- we make it explicit, and apply it to classical BST properties. (...)}, }
Endnote
%0 Report %A Chalermsook, Parinya %A Goswami, Mayank %A Kozma, L&#225;szl&#243; %A Mehlhorn, Kurt %A Saranurak, Thatchaphol %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T The Landscape of Bounds for Binary Search Trees : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-FCC9-9 %U http://arxiv.org/abs/1603.04892 %D 2016 %X Binary search trees (BSTs) with rotations can adapt to various kinds of structure in search sequences, achieving amortized access times substantially better than the Theta(log n) worst-case guarantee. Classical examples of structural properties include static optimality, sequential access, working set, key-independent optimality, and dynamic finger, all of which are now known to be achieved by the two famous online BST algorithms (Splay and Greedy). (...) In this paper, we introduce novel properties that explain the efficiency of sequences not captured by any of the previously known properties, and which provide new barriers to the dynamic optimality conjecture. We also establish connections between various properties, old and new. For instance, we show the following. (i) A tight bound of O(n log d) on the cost of Greedy for d-decomposable sequences. The result builds on the recent lazy finger result of Iacono and Langerman (SODA 2016). On the other hand, we show that lazy finger alone cannot explain the efficiency of pattern avoiding sequences even in some of the simplest cases. (ii) A hierarchy of bounds using multiple lazy fingers, addressing a recent question of Iacono and Langerman. (iii) The optimality of the Move-to-root heuristic in the key-independent setting introduced by Iacono (Algorithmica 2005). (iv) A new tool that allows combining any finite number of sound structural properties. As an application, we show an upper bound on the cost of a class of sequences that all known properties fail to capture. (v) The equivalence between two families of BST properties. The observation on which this connection is based was known before - we make it explicit, and apply it to classical BST properties. (...) %K Computer Science, Data Structures and Algorithms, cs.DS
[154]
P. Chalermsook and D. Vaz, “A Note on Fractional Coloring and the Integrality gap of LP for Maximum Weight Independent Set,” Electronic Notes in Discrete Mathematics (Proc. CTW 2016), vol. 55, 2016.
Export
BibTeX
@article{Chalermsook2016113, TITLE = {A Note on Fractional Coloring and the Integrality gap of {LP} for Maximum Weight Independent Set}, AUTHOR = {Chalermsook, Parinya and Vaz, Daniel}, LANGUAGE = {eng}, ISSN = {1571-0653}, DOI = {10.1016/j.endm.2016.10.029}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, JOURNAL = {Electronic Notes in Discrete Mathematics (Proc. CTW)}, VOLUME = {55}, PAGES = {113--116}, BOOKTITLE = {14th Cologne-Twente Workshop on Graphs and Combinatorial Optimization (CTW16)}, }
Endnote
%0 Journal Article %A Chalermsook, Parinya %A Vaz, Daniel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Note on Fractional Coloring and the Integrality gap of LP for Maximum Weight Independent Set : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5753-F %R 10.1016/j.endm.2016.10.029 %7 2016 %D 2016 %J Electronic Notes in Discrete Mathematics %V 55 %& 113 %P 113 - 116 %I Elsevier %C Amsterdam %@ false %B 14th Cologne-Twente Workshop on Graphs and Combinatorial Optimization (CTW16) %O CTW 2016
[155]
L. S. Chandran, D. Issac, and S. Zhou, “Hadwiger’s Conjecture and Squares of Chordal Graphs,” in Computing and Combinatorics (COCOON 2016), Ho Chi Minh City, Vietnam, 2016.
Export
BibTeX
@inproceedings{SunilChandran2016, TITLE = {Hadwiger{\textquoteright}s Conjecture and Squares of Chordal Graphs}, AUTHOR = {Chandran, L. Sunil and Issac, Davis and Zhou, Sanming}, LANGUAGE = {eng}, DOI = {10.1007/978-3-319-42634-1_34}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Computing and Combinatorics (COCOON 2016)}, EDITOR = {Dinh, Thang N. and Thai, My T.}, PAGES = {417--428}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9797}, ADDRESS = {Ho Chi Minh City, Vietnam}, }
Endnote
%0 Conference Proceedings %A Chandran, L. Sunil %A Issac, Davis %A Zhou, Sanming %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Hadwiger&#8217;s Conjecture and Squares of Chordal Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-53D4-2 %R 10.1007/978-3-319-42634-1_34 %D 2016 %B 22nd International Computing and Combinatorics Conference %Z date of event: 2016-08-02 - 2016-08-04 %C Ho Chi Minh City, Vietnam %B Computing and Combinatorics %E Dinh, Thang N.; Thai, My T. %P 417 - 428 %I Springer %B Lecture Notes in Computer Science %N 9797
[156]
B. Charron-Bost, M. Függer, and T. Nowak, “Fast, Robust, Quantizable Approximate Consensus,” in 43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016), Rome, Italy, 2016.
Export
BibTeX
@inproceedings{CFN16:icalp, TITLE = {Fast, Robust, Quantizable Approximate Consensus}, AUTHOR = {Charron-Bost, Bernadette and F{\"u}gger, Matthias and Nowak, Thomas}, LANGUAGE = {eng}, ISBN = {978-3-95977-013-2}, URL = {urn:nbn:de:0030-drops-62812}, DOI = {10.4230/LIPIcs.ICALP.2016.137}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016)}, EDITOR = {Chatzigiannakis, Ioannis and Mitzenmacher, Michael and Rabani, Yuval and Sangiorgi, Davide}, PAGES = {1--14}, EID = {137}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {55}, ADDRESS = {Rome, Italy}, }
Endnote
%0 Conference Proceedings %A Charron-Bost, Bernadette %A F&#252;gger, Matthias %A Nowak, Thomas %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Fast, Robust, Quantizable Approximate Consensus : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-608B-1 %R 10.4230/LIPIcs.ICALP.2016.137 %U urn:nbn:de:0030-drops-62812 %D 2016 %B 43rd International Colloquium on Automata, Languages, and Programming %Z date of event: 2016-07-12 - 2016-07-15 %C Rome, Italy %B 43rd International Colloquium on Automata, Languages, and Programming %E Chatzigiannakis, Ioannis; Mitzenmacher, Michael; Rabani, Yuval; Sangiorgi, Davide %P 1 - 14 %Z sequence number: 137 %I Schloss Dagstuhl %@ 978-3-95977-013-2 %B Leibniz International Proceedings in Informatics %N 55 %U http://drops.dagstuhl.de/opus/volltexte/2016/6281/http://drops.dagstuhl.de/doku/urheberrecht1.html
[157]
Y. K. Cheun and R. Cole, “A Unified Approach to Analyzing Asynchronous Coordinate Descent and Tatonnement,” 2016. [Online]. Available: http://arxiv.org/abs/1612.09171. (arXiv: 1612.09171)
Abstract
This paper concerns asynchrony in iterative processes, focusing on gradient descent and tatonnement, a fundamental price dynamic. Gradient descent is an important class of iterative algorithms for minimizing convex functions. Classically, gradient descent has been a sequential and synchronous process, although distributed and asynchronous variants have been studied since the 1980s. Coordinate descent is a commonly studied version of gradient descent. In this paper, we focus on asynchronous coordinate descent on convex functions $F:\mathbb{R}^n\rightarrow\mathbb{R}$ of the form $F(x) = f(x) + \sum_{k=1}^n \Psi_k(x_k)$, where $f:\mathbb{R}^n\rightarrow\mathbb{R}$ is a smooth convex function, and each $\Psi_k:\mathbb{R}\rightarrow\mathbb{R}$ is a univariate and possibly non-smooth convex function. Such functions occur in many data analysis and machine learning problems. We give new analyses of cyclic coordinate descent, a parallel asynchronous stochastic coordinate descent, and a rather general worst-case parallel asynchronous coordinate descent. For all of these, we either obtain sharply improved bounds, or provide the first analyses. Our analyses all use a common amortized framework. The application of this framework to the asynchronous stochastic version requires some new ideas, for it is not obvious how to ensure a uniform distribution where it is needed in the face of asynchronous actions that may undo uniformity. We believe that our approach may well be applicable to the analysis of other iterative asynchronous stochastic processes. We extend the framework to show that an asynchronous version of tatonnement, a fundamental price dynamic widely studied in general equilibrium theory, converges toward a market equilibrium for Fisher markets with CES utilities or Leontief utilities, for which tatonnement is equivalent to coordinate descent.
Export
BibTeX
@online{Cheungarxiv16, TITLE = {A Unified Approach to Analyzing Asynchronous Coordinate Descent and Tatonnement}, AUTHOR = {Cheun, Yun Kuen and Cole, Richard}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1612.09171}, EPRINT = {1612.09171}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {This paper concerns asynchrony in iterative processes, focusing on gradient descent and tatonnement, a fundamental price dynamic. Gradient descent is an important class of iterative algorithms for minimizing convex functions. Classically, gradient descent has been a sequential and synchronous process, although distributed and asynchronous variants have been studied since the 1980s. Coordinate descent is a commonly studied version of gradient descent. In this paper, we focus on asynchronous coordinate descent on convex functions $F:\mathbb{R}^n\rightarrow\mathbb{R}$ of the form $F(x) = f(x) + \sum_{k=1}^n \Psi_k(x_k)$, where $f:\mathbb{R}^n\rightarrow\mathbb{R}$ is a smooth convex function, and each $\Psi_k:\mathbb{R}\rightarrow\mathbb{R}$ is a univariate and possibly non-smooth convex function. Such functions occur in many data analysis and machine learning problems. We give new analyses of cyclic coordinate descent, a parallel asynchronous stochastic coordinate descent, and a rather general worst-case parallel asynchronous coordinate descent. For all of these, we either obtain sharply improved bounds, or provide the first analyses. Our analyses all use a common amortized framework. The application of this framework to the asynchronous stochastic version requires some new ideas, for it is not obvious how to ensure a uniform distribution where it is needed in the face of asynchronous actions that may undo uniformity. We believe that our approach may well be applicable to the analysis of other iterative asynchronous stochastic processes. We extend the framework to show that an asynchronous version of tatonnement, a fundamental price dynamic widely studied in general equilibrium theory, converges toward a market equilibrium for Fisher markets with CES utilities or Leontief utilities, for which tatonnement is equivalent to coordinate descent.}, }
Endnote
%0 Report %A Cheun, Yun Kuen %A Cole, Richard %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T A Unified Approach to Analyzing Asynchronous Coordinate Descent and Tatonnement : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5014-A %U http://arxiv.org/abs/1612.09171 %D 2016 %X This paper concerns asynchrony in iterative processes, focusing on gradient descent and tatonnement, a fundamental price dynamic. Gradient descent is an important class of iterative algorithms for minimizing convex functions. Classically, gradient descent has been a sequential and synchronous process, although distributed and asynchronous variants have been studied since the 1980s. Coordinate descent is a commonly studied version of gradient descent. In this paper, we focus on asynchronous coordinate descent on convex functions $F:\mathbb{R}^n\rightarrow\mathbb{R}$ of the form $F(x) = f(x) + \sum_{k=1}^n \Psi_k(x_k)$, where $f:\mathbb{R}^n\rightarrow\mathbb{R}$ is a smooth convex function, and each $\Psi_k:\mathbb{R}\rightarrow\mathbb{R}$ is a univariate and possibly non-smooth convex function. Such functions occur in many data analysis and machine learning problems. We give new analyses of cyclic coordinate descent, a parallel asynchronous stochastic coordinate descent, and a rather general worst-case parallel asynchronous coordinate descent. For all of these, we either obtain sharply improved bounds, or provide the first analyses. Our analyses all use a common amortized framework. The application of this framework to the asynchronous stochastic version requires some new ideas, for it is not obvious how to ensure a uniform distribution where it is needed in the face of asynchronous actions that may undo uniformity. We believe that our approach may well be applicable to the analysis of other iterative asynchronous stochastic processes. We extend the framework to show that an asynchronous version of tatonnement, a fundamental price dynamic widely studied in general equilibrium theory, converges toward a market equilibrium for Fisher markets with CES utilities or Leontief utilities, for which tatonnement is equivalent to coordinate descent. %K Mathematics, Optimization and Control, math.OC,Computer Science, Data Structures and Algorithms, cs.DS,Computer Science, Computer Science and Game Theory, cs.GT,Mathematics, Dynamical Systems, math.DS
[158]
L. Chiantini, C. Ikenmeyer, J. M. Landsberg, and G. Ottaviani, “The Geometry of Rank Decompositions of Matrix Multiplication I: 2x2 Matrices,” 2016. [Online]. Available: http://arxiv.org/abs/1610.08364. (arXiv: 1610.08364)
Abstract
This is the first in a series of papers on rank decompositions of the matrix multiplication tensor. In this paper we: establish general facts about rank decompositions of tensors, describe potential ways to search for new matrix multiplication decompositions, give a geometric proof of the theorem of Burichenko's theorem establishing the symmetry group of Strassen's algorithm, and present two particularly nice subfamilies in the Strassen family of decompositions.
Export
BibTeX
@online{CILO:16, TITLE = {The Geometry of Rank Decompositions of Matrix Multiplication I: 2x2 Matrices}, AUTHOR = {Chiantini, Luca and Ikenmeyer, Christian and Landsberg, J. M. and Ottaviani, Giorgio}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1610.08364}, EPRINT = {1610.08364}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {This is the first in a series of papers on rank decompositions of the matrix multiplication tensor. In this paper we: establish general facts about rank decompositions of tensors, describe potential ways to search for new matrix multiplication decompositions, give a geometric proof of the theorem of Burichenko's theorem establishing the symmetry group of Strassen's algorithm, and present two particularly nice subfamilies in the Strassen family of decompositions.}, }
Endnote
%0 Report %A Chiantini, Luca %A Ikenmeyer, Christian %A Landsberg, J. M. %A Ottaviani, Giorgio %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T The Geometry of Rank Decompositions of Matrix Multiplication I: 2x2 Matrices : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4F8B-3 %U http://arxiv.org/abs/1610.08364 %D 2016 %X This is the first in a series of papers on rank decompositions of the matrix multiplication tensor. In this paper we: establish general facts about rank decompositions of tensors, describe potential ways to search for new matrix multiplication decompositions, give a geometric proof of the theorem of Burichenko's theorem establishing the symmetry group of Strassen's algorithm, and present two particularly nice subfamilies in the Strassen family of decompositions. %K Computer Science, Computational Complexity, cs.CC,Mathematics, Algebraic Geometry, math.AG,
[159]
A. Choudhary, M. Kerber, and S. Raghvendra, “Polynomial-Sized Topological Approximations Using The Permutahedron,” 2016. [Online]. Available: http://arxiv.org/abs/1601.02732. (arXiv: 1601.02732)
Abstract
Classical methods to model topological properties of point clouds, such as the Vietoris-Rips complex, suffer from the combinatorial explosion of complex sizes. We propose a novel technique to approximate a multi-scale filtration of the Rips complex with improved bounds for size: precisely, for $n$ points in $\mathbb{R}^d$, we obtain a $O(d)$-approximation with at most $n2^{O(d \log k)}$ simplices of dimension $k$ or lower. In conjunction with dimension reduction techniques, our approach yields a $O(\mathrm{polylog} (n))$-approximation of size $n^{O(1)}$ for Rips filtrations on arbitrary metric spaces. This result stems from high-dimensional lattice geometry and exploits properties of the permutahedral lattice, a well-studied structure in discrete geometry. Building on the same geometric concept, we also present a lower bound result on the size of an approximate filtration: we construct a point set for which every $(1+\epsilon)$-approximation of the \v{C}ech filtration has to contain $n^{\Omega(\log\log n)}$ features, provided that $\epsilon <\frac{1}{\log^{1+c} n}$ for $c\in(0,1)$.
Export
BibTeX
@online{ChoudharyarXiv2016, TITLE = {Polynomial-Sized Topological Approximations Using The Permutahedron}, AUTHOR = {Choudhary, Aruni and Kerber, Michael and Raghvendra, Sharath}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1601.02732}, EPRINT = {1601.02732}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Classical methods to model topological properties of point clouds, such as the Vietoris-Rips complex, suffer from the combinatorial explosion of complex sizes. We propose a novel technique to approximate a multi-scale filtration of the Rips complex with improved bounds for size: precisely, for $n$ points in $\mathbb{R}^d$, we obtain a $O(d)$-approximation with at most $n2^{O(d \log k)}$ simplices of dimension $k$ or lower. In conjunction with dimension reduction techniques, our approach yields a $O(\mathrm{polylog} (n))$-approximation of size $n^{O(1)}$ for Rips filtrations on arbitrary metric spaces. This result stems from high-dimensional lattice geometry and exploits properties of the permutahedral lattice, a well-studied structure in discrete geometry. Building on the same geometric concept, we also present a lower bound result on the size of an approximate filtration: we construct a point set for which every $(1+\epsilon)$-approximation of the \v{C}ech filtration has to contain $n^{\Omega(\log\log n)}$ features, provided that $\epsilon <\frac{1}{\log^{1+c} n}$ for $c\in(0,1)$.}, }
Endnote
%0 Report %A Choudhary, Aruni %A Kerber, Michael %A Raghvendra, Sharath %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Polynomial-Sized Topological Approximations Using The Permutahedron : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0280-D %U http://arxiv.org/abs/1601.02732 %D 2016 %X Classical methods to model topological properties of point clouds, such as the Vietoris-Rips complex, suffer from the combinatorial explosion of complex sizes. We propose a novel technique to approximate a multi-scale filtration of the Rips complex with improved bounds for size: precisely, for $n$ points in $\mathbb{R}^d$, we obtain a $O(d)$-approximation with at most $n2^{O(d \log k)}$ simplices of dimension $k$ or lower. In conjunction with dimension reduction techniques, our approach yields a $O(\mathrm{polylog} (n))$-approximation of size $n^{O(1)}$ for Rips filtrations on arbitrary metric spaces. This result stems from high-dimensional lattice geometry and exploits properties of the permutahedral lattice, a well-studied structure in discrete geometry. Building on the same geometric concept, we also present a lower bound result on the size of an approximate filtration: we construct a point set for which every $(1+\epsilon)$-approximation of the \v{C}ech filtration has to contain $n^{\Omega(\log\log n)}$ features, provided that $\epsilon <\frac{1}{\log^{1+c} n}$ for $c\in(0,1)$. %K Computer Science, Computational Geometry, cs.CG,Mathematics, Algebraic Topology, math.AT,
[160]
A. Choudhary, M. Kerber, and S. Raghvendra, “Polynomial-Sized Topological Approximations Using the Permutahedron,” in 32nd International Symposium on Computational Geometry (SoCG 2016), Boston, MA, USA, 2016.
Export
BibTeX
@inproceedings{ChoudharySoCG2016, TITLE = {Polynomial-Sized Topological Approximations Using the Permutahedron}, AUTHOR = {Choudhary, Aruni and Kerber, Michael and Raghvendra, Sharat}, LANGUAGE = {eng}, ISBN = {978-3-95977-009-5}, URL = {urn:nbn:de:0030-drops-59236}, DOI = {10.4230/LIPIcs.SoCG.2016.31}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {32nd International Symposium on Computational Geometry (SoCG 2016)}, EDITOR = {Fekete, S{\'a}ndor and Lubiw, Anna}, EID = {31}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {51}, ADDRESS = {Boston, MA, USA}, }
Endnote
%0 Conference Proceedings %A Choudhary, Aruni %A Kerber, Michael %A Raghvendra, Sharat %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Polynomial-Sized Topological Approximations Using the Permutahedron : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-025D-1 %R 10.4230/LIPIcs.SoCG.2016.31 %U urn:nbn:de:0030-drops-59236 %D 2016 %B 32nd International Symposium on Computational Geometry %Z date of event: 2016-06-14 - 2016-06-17 %C Boston, MA, USA %B 32nd International Symposium on Computational Geometry %E Fekete, S&#225;ndor; Lubiw, Anna %Z sequence number: 31 %I Schloss Dagstuhl %@ 978-3-95977-009-5 %B Leibniz International Proceedings in Informatics %N 51 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2016/5923/
[161]
A. Clementi, P. Crescenzi, C. Doerr, P. Fraigniaud, F. Pasquale, and R. Silvestri, “Rumor Spreading in Random Evolving Graphs,” Random Structures and Algorithms, vol. 48, no. 2, 2016.
Export
BibTeX
@article{Doerr14RSA, TITLE = {Rumor Spreading in Random Evolving Graphs}, AUTHOR = {Clementi, Andrea and Crescenzi, Pierluigi and Doerr, Carola and Fraigniaud, Pierre and Pasquale, Francesco and Silvestri, Riccardo}, LANGUAGE = {eng}, ISSN = {1042-9832}, DOI = {10.1002/rsa.20586}, PUBLISHER = {Wiley}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Random Structures and Algorithms}, VOLUME = {48}, NUMBER = {2}, PAGES = {290--312}, }
Endnote
%0 Journal Article %A Clementi, Andrea %A Crescenzi, Pierluigi %A Doerr, Carola %A Fraigniaud, Pierre %A Pasquale, Francesco %A Silvestri, Riccardo %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Rumor Spreading in Random Evolving Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0024-546C-7 %R 10.1002/rsa.20586 %7 2015-03-30 %D 2016 %J Random Structures and Algorithms %V 48 %N 2 %& 290 %P 290 - 312 %I Wiley %C New York, NY %@ false
[162]
C. Croitoru, “Bipartite Digraphs Debates,” in 9th Multidisciplinary Workshop on Advances in Preference Handling (MPref 2015), Buenos Aires, Argentinia, 2016.
Export
BibTeX
@inproceedings{croibipartdebates15, TITLE = {Bipartite Digraphs Debates}, AUTHOR = {Croitoru, Cosmina}, LANGUAGE = {eng}, YEAR = {2015}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {9th Multidisciplinary Workshop on Advances in Preference Handling (MPref 2015)}, ADDRESS = {Buenos Aires, Argentinia}, }
Endnote
%0 Conference Proceedings %A Croitoru, Cosmina %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Bipartite Digraphs Debates : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5776-3 %D 2016 %B 9th Multidisciplinary Workshop on Advances in Preference Handling %Z date of event: 2015-07-27 - 2015-07-27 %C Buenos Aires, Argentinia %B 9th Multidisciplinary Workshop on Advances in Preference Handling
[163]
C. Croitoru and K. Mehlhorn, “Opposition Frameworks,” in Logics in Artificial Intelligence (JELIA 2016), Larnaca, Cyprus, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/jelia/CroitoruM16, TITLE = {Opposition Frameworks}, AUTHOR = {Croitoru, Cosmina and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISBN = {978-3-319-48757-1}, DOI = {10.1007/978-3-319-48758-8_13}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Logics in Artificial Intelligence (JELIA 2016)}, EDITOR = {Michael, Loizos and Kakas, Antonios}, PAGES = {190--206}, SERIES = {Lecture Notes in Artificial Intelligence}, VOLUME = {10021}, ADDRESS = {Larnaca, Cyprus}, }
Endnote
%0 Conference Proceedings %A Croitoru, Cosmina %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Opposition Frameworks : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-57A8-4 %R 10.1007/978-3-319-48758-8_13 %D 2016 %B 15th European Conference on Logics in Artificial Intelligence %Z date of event: 2016-11-09 - 2016-11-11 %C Larnaca, Cyprus %B Logics in Artificial Intelligence %E Michael, Loizos; Kakas, Antonios %P 190 - 206 %I Springer %@ 978-3-319-48757-1 %B Lecture Notes in Artificial Intelligence %N 10021
[164]
M. Cygan, M. Pilipczuk, M. Pilipczuk, E. J. van Leeuwen, and M. Wrochna, “Polynomial Kernelization for Removing Induced Claws and Diamonds,” in Graph-Theoretic Concepts in Computer Science (WG 2015), Garching, Germany, 2016.
Export
BibTeX
@inproceedings{CyganWG2015, TITLE = {Polynomial Kernelization for Removing Induced Claws and Diamonds}, AUTHOR = {Cygan, Marek and Pilipczuk, Marcin and Pilipczuk, Micha{\l} and van Leeuwen, Erik Jan and Wrochna, Marcin}, LANGUAGE = {eng}, ISBN = {978-3-662-53173-0}, DOI = {10.1007/978-3-662-53174-7_31}, PUBLISHER = {Springer}, YEAR = {2015}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Graph-Theoretic Concepts in Computer Science (WG 2015)}, EDITOR = {Mayr, Ernst W.}, PAGES = {440--455}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9224}, ADDRESS = {Garching, Germany}, }
Endnote
%0 Conference Proceedings %A Cygan, Marek %A Pilipczuk, Marcin %A Pilipczuk, Micha&#322; %A van Leeuwen, Erik Jan %A Wrochna, Marcin %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Polynomial Kernelization for Removing Induced Claws and Diamonds : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-309B-7 %R 10.1007/978-3-662-53174-7_31 %D 2016 %B 41st International Workshop on Graph-Theoretic Concepts in Computer Science %Z date of event: 2015-06-17 - 2015-06-19 %C Garching, Germany %B Graph-Theoretic Concepts in Computer Science %E Mayr, Ernst W. %P 440 - 455 %I Springer %@ 978-3-662-53173-0 %B Lecture Notes in Computer Science %N 9224
[165]
J. Dams, M. Hoefer, and T. Kesselheim, “Jamming-resistant Learning in Wireless Networks,” IEEE/ACM Transactions on Networking, vol. 24, no. 5, 2016.
Export
BibTeX
@article{Dams2016, TITLE = {Jamming-resistant Learning in Wireless Networks}, AUTHOR = {Dams, Johannes and Hoefer, Martin and Kesselheim, Thomas}, LANGUAGE = {eng}, DOI = {10.1109/TNET.2015.2486622}, PUBLISHER = {IEEE}, ADDRESS = {Piscataway, NJ}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {IEEE/ACM Transactions on Networking}, VOLUME = {24}, NUMBER = {5}, PAGES = {2809--2818}, }
Endnote
%0 Journal Article %A Dams, Johannes %A Hoefer, Martin %A Kesselheim, Thomas %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Jamming-resistant Learning in Wireless Networks : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-07E9-A %R 10.1109/TNET.2015.2486622 %7 2016 %D 2016 %J IEEE/ACM Transactions on Networking %V 24 %N 5 %& 2809 %P 2809 - 2818 %I IEEE %C Piscataway, NJ
[166]
O. Darwish, “Market Equilibrium Computation for the Linear Arrow-Debreu Model,” Universität des Saarlandes, Saarbrücken, 2016.
Abstract
The problem of market equilibrium is defined as the problem of finding prices for the goods such that the supply in the market is equal to the demand. The problem is applicable to several market models, like the linear Arrow-Debreu model, which is one of the fundamental economic market models. Over the years, various algorithms have been developed to compute the market equilibrium of the linear Arrow-Debreu model. In 2013, Duan and Mehlhorn presented the first combinatorial polynomial time algorithm for computing the market equilibrium of this model. In this thesis, we optimize, generalize, and implement the Duan-Mehlhorn algorithm. We present a novel algorithm for computing balanced ows in equality networks, which is an application of parametric ows. This algorithm outperforms the current best algorithm for computing balanced ows; hence, it improves Duan-Mehlhorn's algorithm by almost a factor of n, which is the size of the network. Moreover, we generalize Duan-Mehlhorn's algorithm by relaxing some of its assumptions. Finally, we describe our approach for implementing Duan-Mehlhorn's algorithm. The preliminary results of our implementation - based on random utility instances - show that the running time of the implementation scales significantly better than the theoretical time complexity.
Export
BibTeX
@mastersthesis{DarwishMaster2016, TITLE = {Market Equilibrium Computation for the Linear Arrow-Debreu Model}, AUTHOR = {Darwish, Omar}, LANGUAGE = {eng}, SCHOOL = {Universit{\"a}t des Saarlandes}, ADDRESS = {Saarbr{\"u}cken}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016-03-31}, ABSTRACT = {The problem of market equilibrium is defined as the problem of finding prices for the goods such that the supply in the market is equal to the demand. The problem is applicable to several market models, like the linear Arrow-Debreu model, which is one of the fundamental economic market models. Over the years, various algorithms have been developed to compute the market equilibrium of the linear Arrow-Debreu model. In 2013, Duan and Mehlhorn presented the first combinatorial polynomial time algorithm for computing the market equilibrium of this model. In this thesis, we optimize, generalize, and implement the Duan-Mehlhorn algorithm. We present a novel algorithm for computing balanced ows in equality networks, which is an application of parametric ows. This algorithm outperforms the current best algorithm for computing balanced ows; hence, it improves Duan-Mehlhorn's algorithm by almost a factor of n, which is the size of the network. Moreover, we generalize Duan-Mehlhorn's algorithm by relaxing some of its assumptions. Finally, we describe our approach for implementing Duan-Mehlhorn's algorithm. The preliminary results of our implementation -- based on random utility instances -- show that the running time of the implementation scales significantly better than the theoretical time complexity.}, }
Endnote
%0 Thesis %A Darwish, Omar %Y Mehlhorn, Kurt %A referee: Hoefer, Martin %+ International Max Planck Research School, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Market Equilibrium Computation for the Linear Arrow-Debreu Model : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-41D0-C %I Universit&#228;t des Saarlandes %C Saarbr&#252;cken %D 2016 %8 31.03.2016 %P 73 p. %V master %9 master %X The problem of market equilibrium is defined as the problem of finding prices for the goods such that the supply in the market is equal to the demand. The problem is applicable to several market models, like the linear Arrow-Debreu model, which is one of the fundamental economic market models. Over the years, various algorithms have been developed to compute the market equilibrium of the linear Arrow-Debreu model. In 2013, Duan and Mehlhorn presented the first combinatorial polynomial time algorithm for computing the market equilibrium of this model. In this thesis, we optimize, generalize, and implement the Duan-Mehlhorn algorithm. We present a novel algorithm for computing balanced ows in equality networks, which is an application of parametric ows. This algorithm outperforms the current best algorithm for computing balanced ows; hence, it improves Duan-Mehlhorn's algorithm by almost a factor of n, which is the size of the network. Moreover, we generalize Duan-Mehlhorn's algorithm by relaxing some of its assumptions. Finally, we describe our approach for implementing Duan-Mehlhorn's algorithm. The preliminary results of our implementation - based on random utility instances - show that the running time of the implementation scales significantly better than the theoretical time complexity.
[167]
O. Darwish and K. Mehlhorn, “Improved Balanced Flow Computation Using Parametric Flow,” Information Processing Letters, vol. 116, no. 9, 2016.
Export
BibTeX
@article{DarwishMehlhorn2016, TITLE = {Improved Balanced Flow Computation Using Parametric Flow}, AUTHOR = {Darwish, Omar and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISSN = {0020-0190}, DOI = {10.1016/j.ipl.2016.04.008}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Information Processing Letters}, VOLUME = {116}, NUMBER = {9}, PAGES = {560--563}, }
Endnote
%0 Journal Article %A Darwish, Omar %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Improved Balanced Flow Computation Using Parametric Flow : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-FC7B-A %R 10.1016/j.ipl.2016.04.008 %7 2016-04-22 %D 2016 %J Information Processing Letters %V 116 %N 9 %& 560 %P 560 - 563 %I Elsevier %C Amsterdam %@ false
[168]
P. J. de Rezende, C. C. de Souza, S. Friedrichs, M. Hemmer, A. Kröller, and D. C. Tozoni, “Engineering Art Galleries,” in Algorithm Engineering, Berlin: Springer, 2016.
Export
BibTeX
@incollection{deRezende2016, TITLE = {Engineering Art Galleries}, AUTHOR = {de Rezende, Pedro J. and de Souza, Cid C. and Friedrichs, Stephan and Hemmer, Michael and Kr{\"o}ller, Alexander and Tozoni, Davi C.}, LANGUAGE = {eng}, ISBN = {978-3-319-49486-9}, DOI = {10.1007/978-3-319-49487-6_12}, PUBLISHER = {Springer}, ADDRESS = {Berlin}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Algorithm Engineering}, EDITOR = {Kliemann, Lasse and Sanders, Peter}, PAGES = {379--417}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9220}, }
Endnote
%0 Book Section %A de Rezende, Pedro J. %A de Souza, Cid C. %A Friedrichs, Stephan %A Hemmer, Michael %A Kr&#246;ller, Alexander %A Tozoni, Davi C. %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Engineering Art Galleries : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-524E-B %R 10.1007/978-3-319-49487-6_12 %D 2016 %B Algorithm Engineering %E Kliemann, Lasse; Sanders, Peter %P 379 - 417 %I Springer %C Berlin %@ 978-3-319-49486-9 %S Lecture Notes in Computer Science %N 9220
[169]
B. Doerr, C. Doerr, S. Moran, and S. Moran, “Simple and Optimal Randomized Fault-tolerant Rumor Spreading,” Distributed Computing, vol. 29, no. 2, 2016.
Export
BibTeX
@article{Doerr2016, TITLE = {Simple and Optimal Randomized Fault-tolerant Rumor Spreading}, AUTHOR = {Doerr, Benjamin and Doerr, Carola and Moran, Shay and Moran, Shlomo}, LANGUAGE = {eng}, ISSN = {0178-2770}, DOI = {10.1007/s00446-014-0238-z}, PUBLISHER = {Springer International}, ADDRESS = {Berlin}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Distributed Computing}, VOLUME = {29}, NUMBER = {2}, PAGES = {89--104}, }
Endnote
%0 Journal Article %A Doerr, Benjamin %A Doerr, Carola %A Moran, Shay %A Moran, Shlomo %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Simple and Optimal Randomized Fault-tolerant Rumor Spreading : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-C4BD-B %R 10.1007/s00446-014-0238-z %7 2014 %D 2016 %J Distributed Computing %V 29 %N 2 %& 89 %P 89 - 104 %I Springer International %C Berlin %@ false
[170]
B. Doerr, C. Doerr, S. Moran, and S. Moran, “Simple and Optimal Fault-tolerant Rumor Spreading,” Distributed Computing, vol. 29, no. 2, 2016.
Export
BibTeX
@article{DoerrDMM14, TITLE = {Simple and Optimal Fault-tolerant Rumor Spreading}, AUTHOR = {Doerr, Benjamin and Doerr, Carola and Moran, Shay and Moran, Shlomo}, LANGUAGE = {eng}, ISSN = {0178-2770}, DOI = {10.1007/s00446-014-0238-z}, PUBLISHER = {Springer International}, ADDRESS = {Berlin}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Distributed Computing}, VOLUME = {29}, NUMBER = {2}, PAGES = {89--104}, }
Endnote
%0 Journal Article %A Doerr, Benjamin %A Doerr, Carola %A Moran, Shay %A Moran, Shlomo %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Simple and Optimal Fault-tolerant Rumor Spreading : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0024-5456-8 %R 10.1007/s00446-014-0238-z %7 2014-12-31 %D 2016 %J Distributed Computing %V 29 %N 2 %& 89 %P 89 - 104 %I Springer International %C Berlin %@ false
[171]
B. Doerr and M. Künnemann, “Improved Protocols and Hardness Results for the Two-Player Cryptogenography Problem,” in 43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016), Rome, Italy, 2016.
Export
BibTeX
@inproceedings{DoerrICALP2016, TITLE = {Improved Protocols and Hardness Results for the Two-Player Cryptogenography Problem}, AUTHOR = {Doerr, Benjamin and K{\"u}nnemann, Marvin}, LANGUAGE = {eng}, ISBN = {978-3-95977-013-2}, URL = {urn:nbn:de:0030-drops-62946}, DOI = {10.4230/LIPIcs.ICALP.2016.150}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016)}, EDITOR = {Chatzigiannakis, Ioannis and Mitzenmacher, Michael and Rabani, Yuval and Sangiorgi, Davide}, EID = {150}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {55}, ADDRESS = {Rome, Italy}, }
Endnote
%0 Conference Proceedings %A Doerr, Benjamin %A K&#252;nnemann, Marvin %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Improved Protocols and Hardness Results for the Two-Player Cryptogenography Problem : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-4289-F %R 10.4230/LIPIcs.ICALP.2016.150 %U urn:nbn:de:0030-drops-62946 %D 2016 %B 43rd International Colloquium on Automata, Languages, and Programming %Z date of event: 2016-07-12 - 2016-07-15 %C Rome, Italy %B 43rd International Colloquium on Automata, Languages, and Programming %E Chatzigiannakis, Ioannis; Mitzenmacher, Michael; Rabani, Yuval; Sangiorgi, Davide %Z sequence number: 150 %I Schloss Dagstuhl %@ 978-3-95977-013-2 %B Leibniz International Proceedings in Informatics %N 55 %U http://drops.dagstuhl.de/opus/volltexte/2016/6294http://drops.dagstuhl.de/doku/urheberrecht1.html
[172]
B. Doerr, C. Doerr, R. Spöhel, and H. Thomas, “Playing Mastermind with Many Colors,” Journal of the ACM, vol. 63, no. 5, 2016.
Export
BibTeX
@article{DoerrJACM2016, TITLE = {Playing {M}astermind with Many Colors}, AUTHOR = {Doerr, Benjamin and Doerr, Carola and Sp{\"o}hel, Reto and Thomas, Henning}, LANGUAGE = {eng}, ISSN = {0004-5411}, DOI = {10.1145/2987372}, PUBLISHER = {ACM}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Journal of the ACM}, VOLUME = {63}, NUMBER = {5}, EID = {42}, }
Endnote
%0 Journal Article %A Doerr, Benjamin %A Doerr, Carola %A Sp&#246;hel, Reto %A Thomas, Henning %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Playing Mastermind with Many Colors : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-BC24-0 %R 10.1145/2987372 %7 2016 %D 2016 %J Journal of the ACM %O JACM %V 63 %N 5 %Z sequence number: 42 %I ACM %C New York, NY %@ false
[173]
D. Dolev, K. Heljanko, M. Järvisalo, J. H. Korhonen, C. Lenzen, J. Rybicki, J. Suomela, and S. Wieringa, “Synchronous Counting and Computational Algorithm Design,” Journal of Computer and System Sciences, vol. 82, no. 2, 2016.
Export
BibTeX
@article{Dolev2016, TITLE = {Synchronous Counting and Computational Algorithm Design}, AUTHOR = {Dolev, Danny and Heljanko, Keijo and J{\"a}rvisalo, Matti and Korhonen, Janne H. and Lenzen, Christoph and Rybicki, Joel and Suomela, Jukka and Wieringa, Siert}, LANGUAGE = {eng}, ISSN = {0022-0000}, DOI = {10.1016/j.jcss.2015.09.002}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Journal of Computer and System Sciences}, VOLUME = {82}, NUMBER = {2}, PAGES = {310--332}, }
Endnote
%0 Journal Article %A Dolev, Danny %A Heljanko, Keijo %A J&#228;rvisalo, Matti %A Korhonen, Janne H. %A Lenzen, Christoph %A Rybicki, Joel %A Suomela, Jukka %A Wieringa, Siert %+ External Organizations External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T Synchronous Counting and Computational Algorithm Design : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0029-4258-B %R 10.1016/j.jcss.2015.09.002 %D 2016 %J Journal of Computer and System Sciences %V 82 %N 2 %& 310 %P 310 - 332 %I Elsevier %C Amsterdam %@ false
[174]
D. Dolev, M. Függer, C. Lenzen, M. Perner, and U. Schmid, “HEX: Scaling Honeycombs is Easier than Scaling Clock Trees,” Journal of Computer and System Sciences, vol. 82, no. 5, 2016.
Export
BibTeX
@article{Dolev2016a, TITLE = {{HEX}: {S}caling Honeycombs is Easier than Scaling Clock Trees}, AUTHOR = {Dolev, Danny and F{\"u}gger, Matthias and Lenzen, Christoph and Perner, Martin and Schmid, Ulrich}, LANGUAGE = {eng}, ISSN = {0022-0000}, DOI = {10.1016/j.jcss.2016.03.001}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Journal of Computer and System Sciences}, VOLUME = {82}, NUMBER = {5}, PAGES = {929--956}, }
Endnote
%0 Journal Article %A Dolev, Danny %A F&#252;gger, Matthias %A Lenzen, Christoph %A Perner, Martin %A Schmid, Ulrich %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T HEX: Scaling Honeycombs is Easier than Scaling Clock Trees : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-41BE-F %R 10.1016/j.jcss.2016.03.001 %7 2016-03-15 %D 2016 %J Journal of Computer and System Sciences %V 82 %N 5 %& 929 %P 929 - 956 %I Elsevier %C Amsterdam %@ false
[175]
R. Duan, J. Garg, and K. Mehlhorn, “An Improved Combinatorial Polynomial Algorithm for the Linear Arrow-Debreu Market,” in Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2016), Arlington, VA, USA, 2016.
Abstract
We present an improved combinatorial algorithm for the computation of equilibrium prices in the linear Arrow-Debreu model. For a market with $n$ agents and integral utilities bounded by $U$, the algorithm runs in $O(n^7 \log^3 (nU))$ time. This improves upon the previously best algorithm of Ye by a factor of $\tOmega(n)$. The algorithm refines the algorithm described by Duan and Mehlhorn and improves it by a factor of $\tOmega(n^3)$. The improvement comes from a better understanding of the iterative price adjustment process, the improved balanced flow computation for nondegenerate instances, and a novel perturbation technique for achieving nondegeneracy.
Export
BibTeX
@inproceedings{DuanSODA2016, TITLE = {An Improved Combinatorial Polynomial Algorithm for the Linear {Arrow}-{Debreu} Market}, AUTHOR = {Duan, Ran and Garg, Jugal and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISBN = {978-1-61197-433-1}, DOI = {10.1137/1.9781611974331.ch7}, PUBLISHER = {SIAM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, ABSTRACT = {We present an improved combinatorial algorithm for the computation of equilibrium prices in the linear Arrow-Debreu model. For a market with $n$ agents and integral utilities bounded by $U$, the algorithm runs in $O(n^7 \log^3 (nU))$ time. This improves upon the previously best algorithm of Ye by a factor of $\tOmega(n)$. The algorithm refines the algorithm described by Duan and Mehlhorn and improves it by a factor of $\tOmega(n^3)$. The improvement comes from a better understanding of the iterative price adjustment process, the improved balanced flow computation for nondegenerate instances, and a novel perturbation technique for achieving nondegeneracy.}, BOOKTITLE = {Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2016)}, EDITOR = {Krauthgamer, Robert}, PAGES = {90--106}, ADDRESS = {Arlington, VA, USA}, }
Endnote
%0 Conference Proceedings %A Duan, Ran %A Garg, Jugal %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T An Improved Combinatorial Polynomial Algorithm for the Linear Arrow-Debreu Market : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0028-81DD-9 %R 10.1137/1.9781611974331.ch7 %D 2016 %B Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2016-01-10 - 2016-01-12 %C Arlington, VA, USA %X We present an improved combinatorial algorithm for the computation of equilibrium prices in the linear Arrow-Debreu model. For a market with $n$ agents and integral utilities bounded by $U$, the algorithm runs in $O(n^7 \log^3 (nU))$ time. This improves upon the previously best algorithm of Ye by a factor of $\tOmega(n)$. The algorithm refines the algorithm described by Duan and Mehlhorn and improves it by a factor of $\tOmega(n^3)$. The improvement comes from a better understanding of the iterative price adjustment process, the improved balanced flow computation for nondegenerate instances, and a novel perturbation technique for achieving nondegeneracy. %B Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms %E Krauthgamer, Robert %P 90 - 106 %I SIAM %@ 978-1-61197-433-1
[176]
K. Dutta and A. Ghosh, “On Subgraphs of Bounded Degeneracy in Hypergraphs,” in Graph-Theoretic Concepts in Computer Science (WG 2016), Istanbul, Turkey, 2016.
Export
BibTeX
@inproceedings{DuttaWG2016, TITLE = {On Subgraphs of Bounded Degeneracy in Hypergraphs}, AUTHOR = {Dutta, Kunal and Ghosh, Arijit}, LANGUAGE = {eng}, ISBN = {978-3-662-53535-6}, DOI = {10.1007/978-3-662-53536-3_25}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Graph-Theoretic Concepts in Computer Science (WG 2016)}, EDITOR = {Heggernes, Pinar}, PAGES = {295--306}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9941}, ADDRESS = {Istanbul, Turkey}, }
Endnote
%0 Conference Proceedings %A Dutta, Kunal %A Ghosh, Arijit %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T On Subgraphs of Bounded Degeneracy in Hypergraphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-7987-B %R 10.1007/978-3-662-53536-3_25 %D 2016 %B 42nd International Workshop on Graph-Theoretic Concepts in Computer Science %Z date of event: 2016-06-22 - 2016-06-24 %C Istanbul, Turkey %B Graph-Theoretic Concepts in Computer Science %E Heggernes, Pinar %P 295 - 306 %I Springer %@ 978-3-662-53535-6 %B Lecture Notes in Computer Science %N 9941
[177]
K. Dutta and C. R. Subramanian, “Improved Bounds on Induced Acyclic Subgraphs in Random Digraphs,” SIAM Journal on Discrete Mathematics, vol. 30, no. 3, 2016.
Export
BibTeX
@article{DBLP:journals/siamdm/DuttaS16, TITLE = {Improved Bounds on Induced Acyclic Subgraphs in Random Digraphs}, AUTHOR = {Dutta, Kunal and Subramanian, C. R.}, LANGUAGE = {eng}, ISSN = {0895-4801}, DOI = {10.1137/140980181}, PUBLISHER = {SIAM}, ADDRESS = {Philadelphia, Pa.}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {SIAM Journal on Discrete Mathematics}, VOLUME = {30}, NUMBER = {3}, PAGES = {1848--1865}, }
Endnote
%0 Journal Article %A Dutta, Kunal %A Subramanian, C. R. %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Improved Bounds on Induced Acyclic Subgraphs in Random Digraphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-65BD-5 %R 10.1137/140980181 %7 2016 %D 2016 %J SIAM Journal on Discrete Mathematics %V 30 %N 3 %& 1848 %P 1848 - 1865 %I SIAM %C Philadelphia, Pa. %@ false
[178]
K. Dutta, E. Ezra, and A. Ghosh, “Two Proofs for Shallow Packings,” Discrete & Computational Geometry, vol. 56, no. 4, 2016.
Export
BibTeX
@article{DBLP:journals/dcg/DuttaEG16, TITLE = {Two Proofs for Shallow Packings}, AUTHOR = {Dutta, Kunal and Ezra, Esther and Ghosh, Arijit}, LANGUAGE = {eng}, ISSN = {0179-5376}, DOI = {10.1007/s00454-016-9824-0}, PUBLISHER = {Springer}, ADDRESS = {London}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Discrete \& Computational Geometry}, VOLUME = {56}, NUMBER = {4}, PAGES = {910--939}, }
Endnote
%0 Journal Article %A Dutta, Kunal %A Ezra, Esther %A Ghosh, Arijit %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Two Proofs for Shallow Packings : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-65A4-C %R 10.1007/s00454-016-9824-0 %7 2016 %D 2016 %J Discrete & Computational Geometry %V 56 %N 4 %& 910 %P 910 - 939 %I Springer %C London %@ false
[179]
P. Dütting, M. Feldman, T. Kesselheim, and B. Lucier, “Posted Prices, Smoothness, and Combinatorial Prophet Inequalities,” 2016. [Online]. Available: http://arxiv.org/abs/1612.03161. (arXiv: 1612.03161)
Abstract
We present a general framework for proving combinatorial prophet inequalities and constructing posted-price mechanisms. Our framework applies to stochastic welfare optimization problems, in which buyers arrive sequentially and make utility-maximizing purchases. Our analysis takes the form of an extension theorem: we derive sufficient conditions for achieving welfare bounds in the special case of deterministic valuations, then prove that these bounds extend directly to stochastic settings. Furthermore, our welfare bounds compose in the sense that the welfare guarantees are preserved when buyers participate in many optimization problems simultaneously. Our sufficient conditions have a natural economic interpretation, and our approach is closely connected to the smoothness framework for bounding the price of anarchy of mechanisms. We show that many smooth mechanisms can be recast as posted-price mechanisms with comparable performance guarantees. We illustrate the power of our framework in a range of applications, including combinatorial auctions, matroids, and sparse packing programs, where we unify and improve many of the previously known results.
Export
BibTeX
@online{DBLP:journals/corr/DuttingFKL16, TITLE = {Posted Prices, Smoothness, and Combinatorial Prophet Inequalities}, AUTHOR = {D{\"u}tting, Paul and Feldman, Michal and Kesselheim, Thomas and Lucier, Brendan}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1612.03161}, EPRINT = {1612.03161}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We present a general framework for proving combinatorial prophet inequalities and constructing posted-price mechanisms. Our framework applies to stochastic welfare optimization problems, in which buyers arrive sequentially and make utility-maximizing purchases. Our analysis takes the form of an extension theorem: we derive sufficient conditions for achieving welfare bounds in the special case of deterministic valuations, then prove that these bounds extend directly to stochastic settings. Furthermore, our welfare bounds compose in the sense that the welfare guarantees are preserved when buyers participate in many optimization problems simultaneously. Our sufficient conditions have a natural economic interpretation, and our approach is closely connected to the smoothness framework for bounding the price of anarchy of mechanisms. We show that many smooth mechanisms can be recast as posted-price mechanisms with comparable performance guarantees. We illustrate the power of our framework in a range of applications, including combinatorial auctions, matroids, and sparse packing programs, where we unify and improve many of the previously known results.}, }
Endnote
%0 Report %A D&#252;tting, Paul %A Feldman, Michal %A Kesselheim, Thomas %A Lucier, Brendan %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Posted Prices, Smoothness, and Combinatorial Prophet Inequalities : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E77-8 %U http://arxiv.org/abs/1612.03161 %D 2016 %X We present a general framework for proving combinatorial prophet inequalities and constructing posted-price mechanisms. Our framework applies to stochastic welfare optimization problems, in which buyers arrive sequentially and make utility-maximizing purchases. Our analysis takes the form of an extension theorem: we derive sufficient conditions for achieving welfare bounds in the special case of deterministic valuations, then prove that these bounds extend directly to stochastic settings. Furthermore, our welfare bounds compose in the sense that the welfare guarantees are preserved when buyers participate in many optimization problems simultaneously. Our sufficient conditions have a natural economic interpretation, and our approach is closely connected to the smoothness framework for bounding the price of anarchy of mechanisms. We show that many smooth mechanisms can be recast as posted-price mechanisms with comparable performance guarantees. We illustrate the power of our framework in a range of applications, including combinatorial auctions, matroids, and sparse packing programs, where we unify and improve many of the previously known results. %K Computer Science, Computer Science and Game Theory, cs.GT,Computer Science, Data Structures and Algorithms, cs.DS
[180]
P. Dütting and T. Kesselheim, “Best-Response Dynamics in Combinatorial Auctions with Item Bidding,” 2016. [Online]. Available: http://arxiv.org/abs/1607.04149. (arXiv: 1607.04149)
Abstract
In a combinatorial auction with item bidding, agents participate in multiple single-item second-price auctions at once. As some items might be substitutes, agents need to strategize in order to maximize their utilities. A number of results indicate that high welfare can be achieved this way, giving bounds on the welfare at equilibrium. Recently, however, criticism has been raised that equilibria are hard to compute and therefore unlikely to be attained. In this paper, we take a different perspective. We study simple best-response dynamics. That is, agents are activated one after the other and each activated agent updates his strategy myopically to a best response against the other agents' current strategies. Often these dynamics may take exponentially long before they converge or they may not converge at all. However, as we show, convergence is not even necessary for good welfare guarantees. Given that agents' bid updates are aggressive enough but not too aggressive, the game will remain in states of good welfare after each agent has updated his bid at least once. In more detail, we show that if agents have fractionally subadditive valuations, natural dynamics reach and remain in a state that provides a $1/3$ approximation to the optimal welfare after each agent has updated his bid at least once. For subadditive valuations, we can guarantee a $\Omega(1/\log m)$ approximation in case of $m$ items that applies after each agent has updated his bid at least once and at any point after that. The latter bound is complemented by a negative result, showing that no kind of best-response dynamics can guarantee more than a $o(\log \log m/\log m)$ fraction of the optimal social welfare.
Export
BibTeX
@online{DBLP:journals/corr/DuttingK16, TITLE = {Best-Response Dynamics in Combinatorial Auctions with Item Bidding}, AUTHOR = {D{\"u}tting, Paul and Kesselheim, Thomas}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1607.04149}, EPRINT = {1607.04149}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {In a combinatorial auction with item bidding, agents participate in multiple single-item second-price auctions at once. As some items might be substitutes, agents need to strategize in order to maximize their utilities. A number of results indicate that high welfare can be achieved this way, giving bounds on the welfare at equilibrium. Recently, however, criticism has been raised that equilibria are hard to compute and therefore unlikely to be attained. In this paper, we take a different perspective. We study simple best-response dynamics. That is, agents are activated one after the other and each activated agent updates his strategy myopically to a best response against the other agents' current strategies. Often these dynamics may take exponentially long before they converge or they may not converge at all. However, as we show, convergence is not even necessary for good welfare guarantees. Given that agents' bid updates are aggressive enough but not too aggressive, the game will remain in states of good welfare after each agent has updated his bid at least once. In more detail, we show that if agents have fractionally subadditive valuations, natural dynamics reach and remain in a state that provides a $1/3$ approximation to the optimal welfare after each agent has updated his bid at least once. For subadditive valuations, we can guarantee a $\Omega(1/\log m)$ approximation in case of $m$ items that applies after each agent has updated his bid at least once and at any point after that. The latter bound is complemented by a negative result, showing that no kind of best-response dynamics can guarantee more than a $o(\log \log m/\log m)$ fraction of the optimal social welfare.}, }
Endnote
%0 Report %A D&#252;tting, Paul %A Kesselheim, Thomas %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Best-Response Dynamics in Combinatorial Auctions with Item Bidding : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E6B-4 %U http://arxiv.org/abs/1607.04149 %D 2016 %X In a combinatorial auction with item bidding, agents participate in multiple single-item second-price auctions at once. As some items might be substitutes, agents need to strategize in order to maximize their utilities. A number of results indicate that high welfare can be achieved this way, giving bounds on the welfare at equilibrium. Recently, however, criticism has been raised that equilibria are hard to compute and therefore unlikely to be attained. In this paper, we take a different perspective. We study simple best-response dynamics. That is, agents are activated one after the other and each activated agent updates his strategy myopically to a best response against the other agents' current strategies. Often these dynamics may take exponentially long before they converge or they may not converge at all. However, as we show, convergence is not even necessary for good welfare guarantees. Given that agents' bid updates are aggressive enough but not too aggressive, the game will remain in states of good welfare after each agent has updated his bid at least once. In more detail, we show that if agents have fractionally subadditive valuations, natural dynamics reach and remain in a state that provides a $1/3$ approximation to the optimal welfare after each agent has updated his bid at least once. For subadditive valuations, we can guarantee a $\Omega(1/\log m)$ approximation in case of $m$ items that applies after each agent has updated his bid at least once and at any point after that. The latter bound is complemented by a negative result, showing that no kind of best-response dynamics can guarantee more than a $o(\log \log m/\log m)$ fraction of the optimal social welfare. %K Computer Science, Computer Science and Game Theory, cs.GT,Computer Science, Data Structures and Algorithms, cs.DS
[181]
K. Elbassioni, K. Mehlhorn, and F. Ramezani, “Towards More Practical Linear Programming-based Techniques for Algorithmic Mechanism Design,” Theory of Computing Systems, vol. 59, no. 4, 2016.
Export
BibTeX
@article{ElbassioniTCS2016, TITLE = {Towards More Practical Linear Programming-based Techniques for Algorithmic Mechanism Design}, AUTHOR = {Elbassioni, Khaled and Mehlhorn, Kurt and Ramezani, Fahimeh}, LANGUAGE = {eng}, ISSN = {1432-4350}, DOI = {10.1007/s00224-016-9704-2}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Theory of Computing Systems}, VOLUME = {59}, NUMBER = {4}, PAGES = {641--663}, }
Endnote
%0 Journal Article %A Elbassioni, Khaled %A Mehlhorn, Kurt %A Ramezani, Fahimeh %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Towards More Practical Linear Programming-based Techniques for Algorithmic Mechanism Design : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-0B1C-9 %R 10.1007/s00224-016-9704-2 %7 2016 %D 2016 %J Theory of Computing Systems %V 59 %N 4 %& 641 %P 641 - 663 %I Springer %C New York, NY %@ false
[182]
A. Elmasry, M. He, J. I. Munro, and P. K. Nicholson, “Dynamic Range Majority Data Structures,” Theoretical Computer Science, vol. 647, 2016.
Export
BibTeX
@article{ElmasryTCS2016, TITLE = {Dynamic Range Majority Data Structures}, AUTHOR = {Elmasry, Amr and He, Meng and Munro, J. Ian and Nicholson, Patrick K.}, LANGUAGE = {eng}, ISSN = {0304-3975}, DOI = {10.1016/j.tcs.2016.07.039}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Theoretical Computer Science}, VOLUME = {647}, PAGES = {59--73}, }
Endnote
%0 Journal Article %A Elmasry, Amr %A He, Meng %A Munro, J. Ian %A Nicholson, Patrick K. %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Dynamic Range Majority Data Structures : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-9AE7-D %R 10.1016/j.tcs.2016.07.039 %7 2016 %D 2016 %J Theoretical Computer Science %V 647 %& 59 %P 59 - 73 %I Elsevier %C Amsterdam %@ false
[183]
C. Engels, R. B. V. Rao, and K. Sreenivasaiah, “Lower Bounds for Projections of Power Symmetric Polynomials,” Electronic Colloquium on Computational Complexity (ECCC): Report Series, vol. 153, 2016.
Export
BibTeX
@article{EngelsRS16, TITLE = {Lower Bounds for Projections of Power Symmetric Polynomials}, AUTHOR = {Engels, Christian and Rao, Raghavendra B. V. and Sreenivasaiah, Karteek}, LANGUAGE = {eng}, ISSN = {1433-8092}, PUBLISHER = {Weizmann Institute of Science}, ADDRESS = {Rehovot}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, JOURNAL = {Electronic Colloquium on Computational Complexity (ECCC): Report Series}, VOLUME = {153}, PAGES = {1--17}, }
Endnote
%0 Journal Article %A Engels, Christian %A Rao, Raghavendra B. V. %A Sreenivasaiah, Karteek %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Lower Bounds for Projections of Power Symmetric Polynomials : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5F70-2 %7 2016 %D 2016 %J Electronic Colloquium on Computational Complexity (ECCC): Report Series %V 153 %& 1 %P 1 - 17 %I Weizmann Institute of Science %C Rehovot %@ false %U https://eccc.weizmann.ac.il/report/2016/153/
[184]
G. Even, M. Medina, and A. Rosén, “A Constant Approximation Algorithm for Scheduling Packets on Line Networks,” in 24th Annual European Symposium on Algorithms (ESA 2016), Aarhus, Denmark, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/esa/EvenMR16, TITLE = {A Constant Approximation Algorithm for Scheduling Packets on Line Networks}, AUTHOR = {Even, Guy and Medina, Moti and Ros{\'e}n, Adi}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-015-6}, URL = {urn:nbn:de:0030-drops-63524}, DOI = {10.4230/LIPIcs.ESA.2016.40}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {24th Annual European Symposium on Algorithms (ESA 2016)}, EDITOR = {Sankowski, Piotr and Zaroliagis, Christos}, PAGES = {1--16}, EID = {40}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {57}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Even, Guy %A Medina, Moti %A Ros&#233;n, Adi %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T A Constant Approximation Algorithm for Scheduling Packets on Line Networks : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5ED5-6 %R 10.4230/LIPIcs.ESA.2016.40 %U urn:nbn:de:0030-drops-63524 %D 2016 %B 24th Annual European Symposium on Algorithms %Z date of event: 2016-08-22 - 2016-08-26 %C Aarhus, Denmark %B 24th Annual European Symposium on Algorithms %E Sankowski, Piotr; Zaroliagis, Christos %P 1 - 16 %Z sequence number: 40 %I Schloss Dagstuhl %@ 978-3-95977-015-6 %B Leibniz International Proceedings in Informatics %N 57 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2016/6352/http://drops.dagstuhl.de/doku/urheberrecht1.html
[185]
G. Even, M. Medina, and B. Patt-Shamir, “On-Line Path Computation and Function Placement in SDNs,” in Stabilization, Safety, and Security of Distributed Systems (SSS 2016), Lyon, France, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/sss/EvenMP16, TITLE = {On-Line Path Computation and Function Placement in {SDN}s}, AUTHOR = {Even, Guy and Medina, Moti and Patt-Shamir, Boaz}, LANGUAGE = {eng}, ISBN = {978-3-319-49258-2}, DOI = {10.1007/978-3-319-49259-9_11}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Stabilization, Safety, and Security of Distributed Systems (SSS 2016)}, EDITOR = {Bonakdarpour, Borzoo and Petit, Franck}, EID = {147}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10083}, ADDRESS = {Lyon, France}, }
Endnote
%0 Conference Proceedings %A Even, Guy %A Medina, Moti %A Patt-Shamir, Boaz %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On-Line Path Computation and Function Placement in SDNs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5ED3-A %R 10.1007/978-3-319-49259-9_11 %D 2016 %B 18th International Symposium on Stabilization, Safety, and Security of Distributed Systems %Z date of event: 2016-11-08 - 2016-11-10 %C Lyon, France %B Stabilization, Safety, and Security of Distributed Systems %E Bonakdarpour, Borzoo; Petit, Franck %P - 131 %Z sequence number: 147 %I Springer %@ 978-3-319-49258-2 %B Lecture Notes in Computer Science %N 10083
[186]
G. Even, R. Levi, M. Medina, and A. Rosen, “Sublinear Random Access Generators for Preferential Attachment Graphs,” 2016. [Online]. Available: http://arxiv.org/abs/1602.06159. (arXiv: 1602.06159)
Abstract
We consider the problem of generating random graphs in evolving random graph models. In the standard approach, the whole graph is chosen randomly according to the distribution of the model before answering queries to the adjacency lists of the graph. Instead, we propose to answer queries by generating the graphs on-the-fly while respecting the probability space of the random graph model. We focus on two random graph models: the Barab{\'{a}}si-Albert Preferential Attachment model (BA-graphs) and the random recursive tree model. We present sublinear randomized generating algorithms for both models. Per query, the running time, the increase in space, and the number of random bits consumed are $\poly\log(n)$ with probability $1-1/\poly(n)$, where $n$ denotes the number of vertices. This result shows that, although the BA random graph model is defined sequentially, random access is possible without chronological evolution. In addition to a conceptual contribution, on-the-fly generation of random graphs can serve as a tool for simulating sublinear algorithms over large BA-graphs.
Export
BibTeX
@online{DBLP:journals/corr/EvenLMR16, TITLE = {Sublinear Random Access Generators for Preferential Attachment Graphs}, AUTHOR = {Even, Guy and Levi, Reut and Medina, Moti and Rosen, Adi}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1602.06159}, EPRINT = {1602.06159}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We consider the problem of generating random graphs in evolving random graph models. In the standard approach, the whole graph is chosen randomly according to the distribution of the model before answering queries to the adjacency lists of the graph. Instead, we propose to answer queries by generating the graphs on-the-fly while respecting the probability space of the random graph model. We focus on two random graph models: the Barab{\'{a}}si-Albert Preferential Attachment model (BA-graphs) and the random recursive tree model. We present sublinear randomized generating algorithms for both models. Per query, the running time, the increase in space, and the number of random bits consumed are $\poly\log(n)$ with probability $1-1/\poly(n)$, where $n$ denotes the number of vertices. This result shows that, although the BA random graph model is defined sequentially, random access is possible without chronological evolution. In addition to a conceptual contribution, on-the-fly generation of random graphs can serve as a tool for simulating sublinear algorithms over large BA-graphs.}, }
Endnote
%0 Report %A Even, Guy %A Levi, Reut %A Medina, Moti %A Rosen, Adi %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Sublinear Random Access Generators for Preferential Attachment Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5EC7-6 %U http://arxiv.org/abs/1602.06159 %D 2016 %X We consider the problem of generating random graphs in evolving random graph models. In the standard approach, the whole graph is chosen randomly according to the distribution of the model before answering queries to the adjacency lists of the graph. Instead, we propose to answer queries by generating the graphs on-the-fly while respecting the probability space of the random graph model. We focus on two random graph models: the Barab{\'{a}}si-Albert Preferential Attachment model (BA-graphs) and the random recursive tree model. We present sublinear randomized generating algorithms for both models. Per query, the running time, the increase in space, and the number of random bits consumed are $\poly\log(n)$ with probability $1-1/\poly(n)$, where $n$ denotes the number of vertices. This result shows that, although the BA random graph model is defined sequentially, random access is possible without chronological evolution. In addition to a conceptual contribution, on-the-fly generation of random graphs can serve as a tool for simulating sublinear algorithms over large BA-graphs. %K Computer Science, Data Structures and Algorithms, cs.DS
[187]
F. V. Fomin, P. Heggernes, and E. J. van Leeuwen, “The Firefighter Problem on Graph Classes,” Theoretical Computer Science, vol. 613, 2016.
Export
BibTeX
@article{FominTCS2016, TITLE = {The {Firefighter} Problem on Graph Classes}, AUTHOR = {Fomin, Fedor V. and Heggernes, Pinar and van Leeuwen, Erik Jan}, LANGUAGE = {eng}, ISSN = {0304-3975}, DOI = {10.1016/j.tcs.2015.11.024}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Theoretical Computer Science}, VOLUME = {613}, PAGES = {38--50}, }
Endnote
%0 Journal Article %A Fomin, Fedor V. %A Heggernes, Pinar %A van Leeuwen, Erik Jan %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T The Firefighter Problem on Graph Classes : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0029-B8FB-F %R 10.1016/j.tcs.2015.11.024 %7 2015-11-23 %D 2016 %J Theoretical Computer Science %V 613 %& 38 %P 38 - 50 %I Elsevier %C Amsterdam %@ false
[188]
N. Fountoulakis, M. Khosla, and K. Panagiotou, “The Multiple-orientability Thresholds for Random Hypergraphs,” Combinatorics, Probability and Computing, vol. 25, no. 6, 2016.
Export
BibTeX
@article{Fountoulakis2016, TITLE = {The Multiple-orientability Thresholds for Random Hypergraphs}, AUTHOR = {Fountoulakis, Nikolaos and Khosla, Megha and Panagiotou, Konstantinos}, LANGUAGE = {eng}, ISSN = {0963-5483}, DOI = {10.1017/S0963548315000334}, PUBLISHER = {Cambridge University Press}, ADDRESS = {Cambridge, England}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Combinatorics, Probability and Computing}, VOLUME = {25}, NUMBER = {6}, PAGES = {870--908}, }
Endnote
%0 Journal Article %A Fountoulakis, Nikolaos %A Khosla, Megha %A Panagiotou, Konstantinos %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T The Multiple-orientability Thresholds for Random Hypergraphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-B7B8-8 %R 10.1017/S0963548315000334 %7 2016 %D 2016 %J Combinatorics, Probability and Computing %V 25 %N 6 %& 870 %P 870 - 908 %I Cambridge University Press %C Cambridge, England %@ false
[189]
S. Friedrichs, M. Hemmer, J. King, and C. Schmidt, “The Continuous 1.5D Terrain Guarding Problem: Discretization, Optimal Solutions, and PTAS,” Journal of Computational Geometry, vol. 7, no. 1, 2016.
Export
BibTeX
@article{FriedrichsHemmerSchmidt2016, TITLE = {The Continuous 1.5{D} Terrain Guarding Problem: {D}iscretization, Optimal Solutions, and {PTAS}}, AUTHOR = {Friedrichs, Stephan and Hemmer, Michael and King, James and Schmidt, Christiane}, LANGUAGE = {eng}, ISSN = {1920-180X}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, JOURNAL = {Journal of Computational Geometry}, VOLUME = {7}, NUMBER = {1}, PAGES = {256--284}, }
Endnote
%0 Journal Article %A Friedrichs, Stephan %A Hemmer, Michael %A King, James %A Schmidt, Christiane %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations External Organizations %T The Continuous 1.5D Terrain Guarding Problem: Discretization, Optimal Solutions, and PTAS : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-C405-7 %7 2016 %D 2016 %J Journal of Computational Geometry %O JoCG %V 7 %N 1 %& 256 %P 256 - 284 %@ false %U http://jocg.org/index.php/jocg/article/view/242
[190]
S. Friedrichs and C. Lenzen, “Parallel Metric Tree Embedding based on an Algebraic View on Moore-Bellman-Ford,” in SPAA’16, 28th ACM Symposium on Parallelism in Algorithms and Architectures, Pacific Grove, CA, USA, 2016.
Export
BibTeX
@inproceedings{FriedrichsSPAA2016, TITLE = {Parallel Metric Tree Embedding based on an Algebraic View on {Moore}-{Bellman}-{Ford}}, AUTHOR = {Friedrichs, Stephan and Lenzen, Christoph}, LANGUAGE = {eng}, ISBN = {978-1-4503-4210-0}, DOI = {10.1145/2935764.2935777}, PUBLISHER = {ACM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {SPAA'16, 28th ACM Symposium on Parallelism in Algorithms and Architectures}, DEBUG = {author: Gilbert, Seth}, EDITOR = {Scheideler, Christian}, PAGES = {455--466}, ADDRESS = {Pacific Grove, CA, USA}, }
Endnote
%0 Conference Proceedings %A Friedrichs, Stephan %A Lenzen, Christoph %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Parallel Metric Tree Embedding based on an Algebraic View on Moore-Bellman-Ford : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0302-3 %R 10.1145/2935764.2935777 %D 2016 %B 28th ACM Symposium on Parallelism in Algorithms and Architectures %Z date of event: 2016-07-11 - 2016-07-13 %C Pacific Grove, CA, USA %B SPAA'16 %E Scheideler, Christian; Gilbert, Seth %P 455 - 466 %I ACM %@ 978-1-4503-4210-0
[191]
S. Friedrichs, M. Függer, and C. Lenzen, “Metastability-Containing Circuits,” 2016. [Online]. Available: http://arxiv.org/abs/1606.06570. (arXiv: 1606.06570)
Abstract
Communication across unsynchronized clock domains is inherently vulnerable to metastable upsets; no digital circuit can deterministically avoid, resolve, or detect metastability (Marino, 1981). Traditionally, a possibly metastable input is stored in synchronizers, decreasing the odds of maintained metastability over time. This approach costs time, and does not guarantee success. We propose a fundamentally different approach: It is possible to \emph{contain} metastability by logical masking, so that it cannot infect the entire circuit. This technique guarantees a limited degree of metastability in---and uncertainty about---the output. We present a synchronizer-free, fault-tolerant clock synchronization algorithm as application, synchronizing clock domains and thus enabling metastability-free communication. At the heart of our approach lies a model for metastability in synchronous clocked digital circuits. Metastability is propagated in a worst-case fashion, allowing to derive deterministic guarantees, without and unlike synchronizers. The proposed model permits positive results while at the same time reproducing established impossibility results regarding avoidance, resolution, and detection of metastability. Furthermore, we fully classify which functions can be computed by synchronous circuits with standard registers, and show that masking registers are computationally strictly more powerful.
Export
BibTeX
@online{Friedrichs_Fuegger_Lenzen2016, TITLE = {Metastability-Containing Circuits}, AUTHOR = {Friedrichs, Stephan and F{\"u}gger, Matthias and Lenzen, Christoph}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1606.06570}, EPRINT = {1606.06570}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Communication across unsynchronized clock domains is inherently vulnerable to metastable upsets; no digital circuit can deterministically avoid, resolve, or detect metastability (Marino, 1981). Traditionally, a possibly metastable input is stored in synchronizers, decreasing the odds of maintained metastability over time. This approach costs time, and does not guarantee success. We propose a fundamentally different approach: It is possible to \emph{contain} metastability by logical masking, so that it cannot infect the entire circuit. This technique guarantees a limited degree of metastability in---and uncertainty about---the output. We present a synchronizer-free, fault-tolerant clock synchronization algorithm as application, synchronizing clock domains and thus enabling metastability-free communication. At the heart of our approach lies a model for metastability in synchronous clocked digital circuits. Metastability is propagated in a worst-case fashion, allowing to derive deterministic guarantees, without and unlike synchronizers. The proposed model permits positive results while at the same time reproducing established impossibility results regarding avoidance, resolution, and detection of metastability. Furthermore, we fully classify which functions can be computed by synchronous circuits with standard registers, and show that masking registers are computationally strictly more powerful.}, }
Endnote
%0 Report %A Friedrichs, Stephan %A F&#252;gger, Matthias %A Lenzen, Christoph %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Metastability-Containing Circuits : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0304-0 %U http://arxiv.org/abs/1606.06570 %D 2016 %X Communication across unsynchronized clock domains is inherently vulnerable to metastable upsets; no digital circuit can deterministically avoid, resolve, or detect metastability (Marino, 1981). Traditionally, a possibly metastable input is stored in synchronizers, decreasing the odds of maintained metastability over time. This approach costs time, and does not guarantee success. We propose a fundamentally different approach: It is possible to \emph{contain} metastability by logical masking, so that it cannot infect the entire circuit. This technique guarantees a limited degree of metastability in---and uncertainty about---the output. We present a synchronizer-free, fault-tolerant clock synchronization algorithm as application, synchronizing clock domains and thus enabling metastability-free communication. At the heart of our approach lies a model for metastability in synchronous clocked digital circuits. Metastability is propagated in a worst-case fashion, allowing to derive deterministic guarantees, without and unlike synchronizers. The proposed model permits positive results while at the same time reproducing established impossibility results regarding avoidance, resolution, and detection of metastability. Furthermore, we fully classify which functions can be computed by synchronous circuits with standard registers, and show that masking registers are computationally strictly more powerful. %K Computer Science, Distributed, Parallel, and Cluster Computing, cs.DC
[192]
M. Függer, T. Nowak, and U. Schmid, “Unfaithful Glitch Propagation in Existing Binary Circuit Models,” IEEE Transactions on Computers, vol. 65, no. 3, 2016.
Export
BibTeX
@article{Fuegger2016, TITLE = {Unfaithful Glitch Propagation in Existing Binary Circuit Models}, AUTHOR = {F{\"u}gger, Matthias and Nowak, Thomas and Schmid, Ulrich}, LANGUAGE = {eng}, ISSN = {0018-9340}, DOI = {10.1109/TC.2015.2435791}, PUBLISHER = {IEEE}, ADDRESS = {Piscataway, NJ}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {IEEE Transactions on Computers}, VOLUME = {65}, NUMBER = {3}, PAGES = {964--978}, }
Endnote
%0 Journal Article %A F&#252;gger, Matthias %A Nowak, Thomas %A Schmid, Ulrich %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Unfaithful Glitch Propagation in Existing Binary Circuit Models : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-1161-D %R 10.1109/TC.2015.2435791 %7 2016 %D 2016 %J IEEE Transactions on Computers %V 65 %N 3 %& 964 %P 964 - 978 %I IEEE %C Piscataway, NJ %@ false
[193]
S. Garg and G. Philip, “Raising The Bar For Vertex Cover: Fixed-parameter Tractability Above A Higher Guarantee,” in Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2016), Arlington, VA, USA, 2016.
Export
BibTeX
@inproceedings{GargSODA2015, TITLE = {Raising The Bar For Vertex Cover: {F}ixed-parameter Tractability Above A Higher Guarantee}, AUTHOR = {Garg, Shivam and Philip, Geevarghese}, LANGUAGE = {eng}, ISBN = {978-1-61197-433-1}, DOI = {10.1137/1.9781611974331.ch80}, PUBLISHER = {SIAM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2016)}, EDITOR = {Krauthgamer, Robert}, PAGES = {1152--1166}, ADDRESS = {Arlington, VA, USA}, }
Endnote
%0 Conference Proceedings %A Garg, Shivam %A Philip, Geevarghese %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Raising The Bar For Vertex Cover: Fixed-parameter Tractability Above A Higher Guarantee : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0028-8205-0 %R 10.1137/1.9781611974331.ch80 %D 2016 %B Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2016-01-10 - 2016-01-12 %C Arlington, VA, USA %B Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms %E Krauthgamer, Robert %P 1152 - 1166 %I SIAM %@ 978-1-61197-433-1
[194]
F. Gesmundo, C. Ikenmeyer, and G. Panova, “Geometric complexity theory and matrix powering,” 2016. [Online]. Available: http://arxiv.org/abs/1611.00827. (arXiv: 1611.00827)
Abstract
Valiant's famous determinant versus permanent problem is the flagship problem in algebraic complexity theory. Mulmuley and Sohoni (Siam J Comput 2001, 2008) introduced geometric complexity theory, an approach to study this and related problems via algebraic geometry and representation theory. Their approach works by multiplying the permanent polynomial with a high power of a linear form (a process called padding) and then comparing the orbit closures of the determinant and the padded permanent. This padding was recently used heavily to show no-go results for the method of shifted partial derivatives (Efremenko, Landsberg, Schenck, Weyman, 2016) and for geometric complexity theory (Ikenmeyer Panova, FOCS 2016 and B\"urgisser, Ikenmeyer Panova, FOCS 2016). Following a classical homogenization result of Nisan (STOC 1991) we replace the determinant in geometric complexity theory with the trace of a variable matrix power. This gives an equivalent but much cleaner homogeneous formulation of geometric complexity theory in which the padding is removed. This radically changes the representation theoretic questions involved to prove complexity lower bounds. We prove that in this homogeneous formulation there are no orbit occurrence obstructions that prove even superlinear lower bounds on the complexity of the permanent. This is the first no-go result in geometric complexity theory that rules out superlinear lower bounds in some model. Interestingly---in contrast to the determinant---the trace of a variable matrix power is not uniquely determined by its stabilizer.
Export
BibTeX
@online{GIP:16, TITLE = {Geometric complexity theory and matrix powering}, AUTHOR = {Gesmundo, Fulvio and Ikenmeyer, Christian and Panova, Greta}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1611.00827}, EPRINT = {1611.00827}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Valiant's famous determinant versus permanent problem is the flagship problem in algebraic complexity theory. Mulmuley and Sohoni (Siam J Comput 2001, 2008) introduced geometric complexity theory, an approach to study this and related problems via algebraic geometry and representation theory. Their approach works by multiplying the permanent polynomial with a high power of a linear form (a process called padding) and then comparing the orbit closures of the determinant and the padded permanent. This padding was recently used heavily to show no-go results for the method of shifted partial derivatives (Efremenko, Landsberg, Schenck, Weyman, 2016) and for geometric complexity theory (Ikenmeyer Panova, FOCS 2016 and B\"urgisser, Ikenmeyer Panova, FOCS 2016). Following a classical homogenization result of Nisan (STOC 1991) we replace the determinant in geometric complexity theory with the trace of a variable matrix power. This gives an equivalent but much cleaner homogeneous formulation of geometric complexity theory in which the padding is removed. This radically changes the representation theoretic questions involved to prove complexity lower bounds. We prove that in this homogeneous formulation there are no orbit occurrence obstructions that prove even superlinear lower bounds on the complexity of the permanent. This is the first no-go result in geometric complexity theory that rules out superlinear lower bounds in some model. Interestingly---in contrast to the determinant---the trace of a variable matrix power is not uniquely determined by its stabilizer.}, }
Endnote
%0 Report %A Gesmundo, Fulvio %A Ikenmeyer, Christian %A Panova, Greta %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Geometric complexity theory and matrix powering : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4F88-9 %U http://arxiv.org/abs/1611.00827 %D 2016 %X Valiant's famous determinant versus permanent problem is the flagship problem in algebraic complexity theory. Mulmuley and Sohoni (Siam J Comput 2001, 2008) introduced geometric complexity theory, an approach to study this and related problems via algebraic geometry and representation theory. Their approach works by multiplying the permanent polynomial with a high power of a linear form (a process called padding) and then comparing the orbit closures of the determinant and the padded permanent. This padding was recently used heavily to show no-go results for the method of shifted partial derivatives (Efremenko, Landsberg, Schenck, Weyman, 2016) and for geometric complexity theory (Ikenmeyer Panova, FOCS 2016 and B\"urgisser, Ikenmeyer Panova, FOCS 2016). Following a classical homogenization result of Nisan (STOC 1991) we replace the determinant in geometric complexity theory with the trace of a variable matrix power. This gives an equivalent but much cleaner homogeneous formulation of geometric complexity theory in which the padding is removed. This radically changes the representation theoretic questions involved to prove complexity lower bounds. We prove that in this homogeneous formulation there are no orbit occurrence obstructions that prove even superlinear lower bounds on the complexity of the permanent. This is the first no-go result in geometric complexity theory that rules out superlinear lower bounds in some model. Interestingly---in contrast to the determinant---the trace of a variable matrix power is not uniquely determined by its stabilizer. %K Computer Science, Computational Complexity, cs.CC,Mathematics, Representation Theory, math.RT,
[195]
P. A. Golovach, D. Paulusma, and E. J. van Leeuwen, “Induced Disjoint Paths in Circular-Arc Graphs in Linear Time,” Theoretical Computer Science, vol. 640, 2016.
Export
BibTeX
@article{GolovachTCS2016, TITLE = {Induced Disjoint Paths in Circular-Arc Graphs in Linear Time}, AUTHOR = {Golovach, Petr A. and Paulusma, Dani{\"e}l and van Leeuwen, Erik Jan}, LANGUAGE = {eng}, ISSN = {0304-3975}, DOI = {10.1016/j.tcs.2016.06.003}, PUBLISHER = {Elsevier}, ADDRESS = {Amsterdam}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Theoretical Computer Science}, VOLUME = {640}, PAGES = {70--83}, }
Endnote
%0 Journal Article %A Golovach, Petr A. %A Paulusma, Dani&#235;l %A van Leeuwen, Erik Jan %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Induced Disjoint Paths in Circular-Arc Graphs in Linear Time : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-459D-2 %R 10.1016/j.tcs.2016.06.003 %7 2016 %D 2016 %J Theoretical Computer Science %V 640 %& 70 %P 70 - 83 %I Elsevier %C Amsterdam %@ false
[196]
M. Göös, J. Hirvonen, R. Levi, M. Medina, and J. Suomela, “Non-local Probes Do Not Help with Many Graph Problems,” in Distributed Computing (DISC 2016), Paris, France, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/wdag/GoosHLMS16, TITLE = {Non-local Probes Do Not Help with Many Graph Problems}, AUTHOR = {G{\"o}{\"o}s, Mika and Hirvonen, Juho and Levi, Reut and Medina, Moti and Suomela, Jukka}, LANGUAGE = {eng}, ISBN = {978-3-662-53425-0}, DOI = {10.1007/978-3-662-53426-7_15}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Distributed Computing (DISC 2016)}, EDITOR = {Gavoille, Cyril and Ilcinkas, David}, PAGES = {201--214}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9888}, ADDRESS = {Paris, France}, }
Endnote
%0 Conference Proceedings %A G&#246;&#246;s, Mika %A Hirvonen, Juho %A Levi, Reut %A Medina, Moti %A Suomela, Jukka %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Non-local Probes Do Not Help with Many Graph Problems : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5ECF-5 %R 10.1007/978-3-662-53426-7_15 %D 2016 %B 30th International Symposium on Distributed Computing %Z date of event: 2016-09-27 - 2016-09-29 %C Paris, France %B Distributed Computing %E Gavoille, Cyril; Ilcinkas, David %P 201 - 214 %I Springer %@ 978-3-662-53425-0 %B Lecture Notes in Computer Science %N 9888
[197]
T. Harks, M. Hoefer, K. Schewior, and A. Skopalik, “Routing Games with Progressive Filling,” IEEE/ACM Transactions on Networking, vol. 24, no. 4, 2016.
Export
BibTeX
@article{Harks2016, TITLE = {Routing Games with Progressive Filling}, AUTHOR = {Harks, Tobias and Hoefer, Martin and Schewior, Kevin and Skopalik, Alexander}, LANGUAGE = {eng}, ISBN = {978-14799-3360-0}, DOI = {10.1109/TNET.2015.2468571}, PUBLISHER = {IEEE}, ADDRESS = {Piscataway, NJ}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {IEEE/ACM Transactions on Networking}, VOLUME = {24}, NUMBER = {4}, PAGES = {2553--2562}, }
Endnote
%0 Journal Article %A Harks, Tobias %A Hoefer, Martin %A Schewior, Kevin %A Skopalik, Alexander %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Routing Games with Progressive Filling : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-7EC8-4 %R 10.1109/TNET.2015.2468571 %7 2016 %D 2016 %J IEEE/ACM Transactions on Networking %V 24 %N 4 %& 2553 %P 2553 - 2562 %I IEEE %C Piscataway, NJ %@ 978-14799-3360-0
[198]
M. Henzinger, S. Krinninger, and D. Nanongkai, “A Deterministic Almost-Tight Distributed Algorithm for Approximating Single-Source Shortest Paths,” in Proceedings of the 48th Annual ACM SIGACT Symposium on Theory of Computing (STOC 2016), Cambridge, MA, USA, 2016.
Export
BibTeX
@inproceedings{HenzingerKrin16, TITLE = {A Deterministic Almost-Tight Distributed Algorithm for Approximating Single-Source Shortest Paths}, AUTHOR = {Henzinger, Monika and Krinninger, Sebastian and Nanongkai, Danupon}, LANGUAGE = {eng}, DOI = {10.1145/2897518.2897638}, PUBLISHER = {ACM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {Proceedings of the 48th Annual ACM SIGACT Symposium on Theory of Computing (STOC 2016)}, PAGES = {489--498}, ADDRESS = {Cambridge, MA, USA}, }
Endnote
%0 Conference Proceedings %A Henzinger, Monika %A Krinninger, Sebastian %A Nanongkai, Danupon %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T A Deterministic Almost-Tight Distributed Algorithm for Approximating Single-Source Shortest Paths : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-52C0-5 %R 10.1145/2897518.2897638 %D 2016 %B 48th Annual ACM SIGACT Symposium on Theory of Computing %Z date of event: 2016-06-19 - 2016-06-21 %C Cambridge, MA, USA %B Proceedings of the 48th Annual ACM SIGACT Symposium on Theory of Computing %P 489 - 498 %I ACM
[199]
S. Heydrich and R. van Stee, “Improved Lower Bounds for Online Hypercube Packing,” 2016. [Online]. Available: http://arxiv.org/abs/1607.01229. (arXiv: 1607.01229)
Abstract
Packing a given sequence of items into as few bins as possible in an online fashion is a widely studied problem. We improve lower bounds for packing hypercubes into bins in two or more dimensions, once for general algorithms (in two dimensions) and once for an important subclass, so-called Harmonic-type algorithms (in two or more dimensions). Lastly, we show that two adaptions of the ideas from the best known one-dimensional packing algorithm to square packing also do not help to break the barrier of 2.
Export
BibTeX
@online{HeydrichS16, TITLE = {Improved Lower Bounds for Online Hypercube Packing}, AUTHOR = {Heydrich, Sandy and van Stee, Rob}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1607.01229}, EPRINT = {1607.01229}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Packing a given sequence of items into as few bins as possible in an online fashion is a widely studied problem. We improve lower bounds for packing hypercubes into bins in two or more dimensions, once for general algorithms (in two dimensions) and once for an important subclass, so-called Harmonic-type algorithms (in two or more dimensions). Lastly, we show that two adaptions of the ideas from the best known one-dimensional packing algorithm to square packing also do not help to break the barrier of 2.}, }
Endnote
%0 Report %A Heydrich, Sandy %A van Stee, Rob %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Improved Lower Bounds for Online Hypercube Packing : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-54AF-0 %U http://arxiv.org/abs/1607.01229 %D 2016 %X Packing a given sequence of items into as few bins as possible in an online fashion is a widely studied problem. We improve lower bounds for packing hypercubes into bins in two or more dimensions, once for general algorithms (in two dimensions) and once for an important subclass, so-called Harmonic-type algorithms (in two or more dimensions). Lastly, we show that two adaptions of the ideas from the best known one-dimensional packing algorithm to square packing also do not help to break the barrier of 2. %K Computer Science, Data Structures and Algorithms, cs.DS
[200]
S. Heydrich and R. van Stee, “Beating the Harmonic Lower Bound for Online Bin Packing,” in 43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016), Rome, Italy, 2016.
Export
BibTeX
@inproceedings{HeydrichICALP2016, TITLE = {Beating the Harmonic Lower Bound for Online Bin Packing}, AUTHOR = {Heydrich, Sandy and van Stee, Rob}, LANGUAGE = {eng}, ISBN = {978-3-95977-013-2}, URL = {urn:nbn:de:0030-drops-63214}, DOI = {10.4230/LIPIcs.ICALP.2016.41}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016)}, EDITOR = {Chatzigiannakis, Ioannis and Mitzenmacher, Michael and Rabani, Yuval and Sangiorgi, Davide}, EID = {41}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {55}, ADDRESS = {Rome, Italy}, }
Endnote
%0 Conference Proceedings %A Heydrich, Sandy %A van Stee, Rob %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Beating the Harmonic Lower Bound for Online Bin Packing : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-4294-6 %R 10.4230/LIPIcs.ICALP.2016.41 %U urn:nbn:de:0030-drops-63214 %D 2016 %B 43rd International Colloquium on Automata, Languages, and Programming %Z date of event: 2016-07-12 - 2016-07-15 %C Rome, Italy %B 43rd International Colloquium on Automata, Languages, and Programming %E Chatzigiannakis, Ioannis; Mitzenmacher, Michael; Rabani, Yuval; Sangiorgi, Davide %Z sequence number: 41 %I Schloss Dagstuhl %@ 978-3-95977-013-2 %B Leibniz International Proceedings in Informatics %N 55 %U http://drops.dagstuhl.de/opus/volltexte/2016/6321/http://drops.dagstuhl.de/doku/urheberrecht1.html
[201]
M. Hoefer and R. Lavi, “Preface to Special Issue on Algorithmic Game Theory,” Theory of Computing Systems, vol. 59, no. 4, 2016.
Export
BibTeX
@article{Hoefer_Lavi2016, TITLE = {Preface to Special Issue on Algorithmic Game Theory}, AUTHOR = {Hoefer, Martin and Lavi, Ron}, LANGUAGE = {eng}, ISSN = {1432-4350}, DOI = {10.1007/s00224-016-9721-1}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Theory of Computing Systems}, VOLUME = {59}, NUMBER = {4}, PAGES = {561--562}, }
Endnote
%0 Journal Article %A Hoefer, Martin %A Lavi, Ron %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Preface to Special Issue on Algorithmic Game Theory : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-2DD4-F %R 10.1007/s00224-016-9721-1 %7 2016 %D 2016 %J Theory of Computing Systems %V 59 %N 4 %& 561 %P 561 - 562 %I Springer %C New York, NY %@ false
[202]
M. Hoefer, T. Kesselheim, and B. Vöcking, “Truthfulness and Stochastic Dominance with Monetary Transfers,” ACM Transactions on Economics and Computation, vol. 4, no. 2, 2016.
Export
BibTeX
@article{HoeferTEC2016, TITLE = {Truthfulness and Stochastic Dominance with Monetary Transfers}, AUTHOR = {Hoefer, Martin and Kesselheim, Thomas and V{\"o}cking, Berthold}, LANGUAGE = {eng}, DOI = {10.1145/2847522}, PUBLISHER = {ACM}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {ACM Transactions on Economics and Computation}, VOLUME = {4}, NUMBER = {2}, EID = {11}, }
Endnote
%0 Journal Article %A Hoefer, Martin %A Kesselheim, Thomas %A V&#246;cking, Berthold %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Truthfulness and Stochastic Dominance with Monetary Transfers : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0216-0 %R 10.1145/2847522 %7 2016 %D 2016 %J ACM Transactions on Economics and Computation %V 4 %N 2 %Z sequence number: 11 %I ACM %C New York, NY
[203]
M. Hoefer, T. Kesselheim, and B. Kodric, “Smoothness for Simultaneous Composition of Mechanisms with Admission,” in Web and Internet Economics (WINE 2016), Montréal, Canada, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/wine/HoeferKK16, TITLE = {Smoothness for Simultaneous Composition of Mechanisms with Admission}, AUTHOR = {Hoefer, Martin and Kesselheim, Thomas and Kodric, Bojana}, LANGUAGE = {eng}, ISBN = {978-3-662-54109-8}, DOI = {10.1007/978-3-662-54110-4_21}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Web and Internet Economics (WINE 2016)}, EDITOR = {Cai, Yang and Vetta, Adrian}, PAGES = {294--308}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10123}, ADDRESS = {Montr{\'e}al, Canada}, }
Endnote
%0 Conference Proceedings %A Hoefer, Martin %A Kesselheim, Thomas %A Kodric, Bojana %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Smoothness for Simultaneous Composition of Mechanisms with Admission : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E66-E %R 10.1007/978-3-662-54110-4_21 %D 2016 %B 12th International Conference on Web and Internet Economics %Z date of event: 2016-12-11 - 2016-12-14 %C Montr&#233;al, Canada %B Web and Internet Economics %E Cai, Yang; Vetta, Adrian %P 294 - 308 %I Springer %@ 978-3-662-54109-8 %B Lecture Notes in Computer Science %N 10123
[204]
C.-C. Huang, T. Kavitha, K. Mehlhorn, and D. Michail, “Fair Matchings and Related Problems,” Algorithmica, vol. 74, no. 3, 2016.
Export
BibTeX
@article{Huang2016, TITLE = {Fair Matchings and Related Problems}, AUTHOR = {Huang, Chien-Chung and Kavitha, Telikepalli and Mehlhorn, Kurt and Michail, Dimitrios}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-015-9994-9}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Algorithmica}, VOLUME = {74}, NUMBER = {3}, PAGES = {1184--1203}, }
Endnote
%0 Journal Article %A Huang, Chien-Chung %A Kavitha, Telikepalli %A Mehlhorn, Kurt %A Michail, Dimitrios %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Fair Matchings and Related Problems : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-115F-4 %R 10.1007/s00453-015-9994-9 %7 2016 %D 2016 %J Algorithmica %V 74 %N 3 %& 1184 %P 1184 - 1203 %I Springer %C New York, NY %@ false
[205]
C.-C. Huang and S. Ott, “A Combinatorial Approximation Algorithm for Graph Balancing with Light Hyper Edges,” in 24th Annual European Symposium on Algorithms (ESA 2016), Aarhus, Denmark, 2016.
Export
BibTeX
@inproceedings{HuangOtt16, TITLE = {A Combinatorial Approximation Algorithm for Graph Balancing with Light Hyper Edges}, AUTHOR = {Huang, Chien-Chung and Ott, Sebastian}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-015-6}, URL = {urn:nbn:de:0030-drops-63919}, DOI = {10.4230/LIPIcs.ESA.2016.49}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {24th Annual European Symposium on Algorithms (ESA 2016)}, EDITOR = {Sankowski, Piotr and Zaroliagis, Christos}, PAGES = {1--15}, EID = {49}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {57}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Huang, Chien-Chung %A Ott, Sebastian %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Combinatorial Approximation Algorithm for Graph Balancing with Light Hyper Edges : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5420-0 %R 10.4230/LIPIcs.ESA.2016.49 %U urn:nbn:de:0030-drops-63919 %D 2016 %B 24th Annual European Symposium on Algorithms %Z date of event: 2016-08-22 - 2016-08-26 %C Aarhus, Denmark %B 24th Annual European Symposium on Algorithms %E Sankowski, Piotr; Zaroliagis, Christos %P 1 - 15 %Z sequence number: 49 %I Schloss Dagstuhl %@ 978-3-95977-015-6 %B Leibniz International Proceedings in Informatics %N 57 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2016/6391/http://drops.dagstuhl.de/doku/urheberrecht1.html
[206]
C. Ikenmeyer and S. Mengel, “On the Relative Power of Reduction Notions in Arithmetic Circuit Complexity,” 2016. [Online]. Available: http://arxiv.org/abs/1609.05942. (arXiv: 1609.05942)
Abstract
We show that the two main reduction notions in arithmetic circuit complexity, p-projections and c-reductions, differ in power. We do so by showing unconditionally that there are polynomials that are VNP-complete under c-reductions but not under p-projections. We also show that the question of which polynomials are VNP-complete under which type of reductions depends on the underlying field.
Export
BibTeX
@online{IM:16, TITLE = {On the Relative Power of Reduction Notions in Arithmetic Circuit Complexity}, AUTHOR = {Ikenmeyer, Christian and Mengel, Stefan}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1609.05942}, EPRINT = {1609.05942}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We show that the two main reduction notions in arithmetic circuit complexity, p-projections and c-reductions, differ in power. We do so by showing unconditionally that there are polynomials that are VNP-complete under c-reductions but not under p-projections. We also show that the question of which polynomials are VNP-complete under which type of reductions depends on the underlying field.}, }
Endnote
%0 Report %A Ikenmeyer, Christian %A Mengel, Stefan %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On the Relative Power of Reduction Notions in Arithmetic Circuit Complexity : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4F8E-E %U http://arxiv.org/abs/1609.05942 %D 2016 %X We show that the two main reduction notions in arithmetic circuit complexity, p-projections and c-reductions, differ in power. We do so by showing unconditionally that there are polynomials that are VNP-complete under c-reductions but not under p-projections. We also show that the question of which polynomials are VNP-complete under which type of reductions depends on the underlying field. %K Computer Science, Computational Complexity, cs.CC,
[207]
C. Ikenmeyer and J. M. Landsberg, “On the Complexity of the Permanent in Various Computational Models,” 2016. [Online]. Available: http://arxiv.org/abs/1610.00159. (arXiv: 1610.00159)
Abstract
We answer a question in [Landsberg, Ressayre, 2015], showing the regular determinantal complexity of the determinant det_m is O(m^3). We answer questions in, and generalize results of [Aravind, Joglekar, 2015], showing there is no rank one determinantal expression for perm_m or det_m when m >= 3. Finally we state and prove several "folklore" results relating different models of computation.
Export
BibTeX
@online{IL:16, TITLE = {On the Complexity of the Permanent in Various Computational Models}, AUTHOR = {Ikenmeyer, Christian and Landsberg, J. M.}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1610.00159}, EPRINT = {1610.00159}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We answer a question in [Landsberg, Ressayre, 2015], showing the regular determinantal complexity of the determinant det_m is O(m^3). We answer questions in, and generalize results of [Aravind, Joglekar, 2015], showing there is no rank one determinantal expression for perm_m or det_m when m >= 3. Finally we state and prove several "folklore" results relating different models of computation.}, }
Endnote
%0 Report %A Ikenmeyer, Christian %A Landsberg, J. M. %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T On the Complexity of the Permanent in Various Computational Models : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4F1E-C %U http://arxiv.org/abs/1610.00159 %D 2016 %X We answer a question in [Landsberg, Ressayre, 2015], showing the regular determinantal complexity of the determinant det_m is O(m^3). We answer questions in, and generalize results of [Aravind, Joglekar, 2015], showing there is no rank one determinantal expression for perm_m or det_m when m >= 3. Finally we state and prove several "folklore" results relating different models of computation. %K Computer Science, Computational Complexity, cs.CC,
[208]
A. Jeż, “One-variable Word Equations in Linear Time,” Algorithmica, vol. 74, no. 1, 2016.
Export
BibTeX
@article{JezAlgorithmica2016, TITLE = {One-variable Word Equations in Linear Time}, AUTHOR = {Je{\.z}, Artur}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-014-9931-3}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Algorithmica}, VOLUME = {74}, NUMBER = {1}, PAGES = {1--48}, }
Endnote
%0 Journal Article %A Je&#380;, Artur %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society %T One-variable Word Equations in Linear Time : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0029-75F8-F %R 10.1007/s00453-014-9931-3 %7 2014 %D 2016 %J Algorithmica %V 74 %N 1 %& 1 %P 1 - 48 %I Springer %C New York, NY %@ false
[209]
M. John and A. Karrenbauer, “A Novel SDP Relaxation for the Quadratic Assignment Problem Using Cut Pseudo Bases,” in Combinatorial Optimization (ISCO 2016), Vietri sul Mare, Italy, 2016.
Export
BibTeX
@inproceedings{JohnISCO2016, TITLE = {A Novel {SDP} Relaxation for the Quadratic Assignment Problem Using Cut Pseudo Bases}, AUTHOR = {John, Maximilian and Karrenbauer, Andreas}, LANGUAGE = {eng}, ISBN = {978-3-319-45586-0}, DOI = {10.1007/978-3-319-45587-7_36}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Combinatorial Optimization (ISCO 2016)}, EDITOR = {Cerulli, Raffaele and Fujishige, Satoru and Mahjoub, Ridha A.}, PAGES = {414--425}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9849}, ADDRESS = {Vietri sul Mare, Italy}, }
Endnote
%0 Conference Proceedings %A John, Maximilian %A Karrenbauer, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Novel SDP Relaxation for the Quadratic Assignment Problem Using Cut Pseudo Bases : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-8417-5 %R 10.1007/978-3-319-45587-7_36 %D 2016 %B 4th International Symposium on Combinatorial Optimization %Z date of event: 2016-05-16 - 2016-05-18 %C Vietri sul Mare, Italy %B Combinatorial Optimization %E Cerulli, Raffaele; Fujishige, Satoru; Mahjoub, Ridha A. %P 414 - 425 %I Springer %@ 978-3-319-45586-0 %B Lecture Notes in Computer Science %N 9849
[210]
I. Kanj, C. Komusiewicz, M. Sorge, and E. J. van Leeuwen, “Parameterized Algorithms for Recognizing Monopolar and 2-Subcolorable Graphs,” in 15th Scandinavian Symposium and Workshops on Algorithm Theory (SWAT 2016), Reykjavik, Iceland, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/swat/KanjKSL16, TITLE = {Parameterized Algorithms for Recognizing Monopolar and 2-Subcolorable Graphs}, AUTHOR = {Kanj, Iyad and Komusiewicz, Christian and Sorge, Manuel and van Leeuwen, Erik Jan}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-011-8}, URL = {urn:nbn:de:0030-drops-60360}, DOI = {10.4230/LIPIcs.SWAT.2016.14}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {15th Scandinavian Symposium and Workshops on Algorithm Theory (SWAT 2016)}, EDITOR = {Pagh, Rasmus}, PAGES = {1--14}, EID = {14}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {53}, ADDRESS = {Reykjavik, Iceland}, }
Endnote
%0 Conference Proceedings %A Kanj, Iyad %A Komusiewicz, Christian %A Sorge, Manuel %A van Leeuwen, Erik Jan %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Parameterized Algorithms for Recognizing Monopolar and 2-Subcolorable Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5367-9 %U urn:nbn:de:0030-drops-60360 %R 10.4230/LIPIcs.SWAT.2016.14 %D 2016 %B 15th Scandinavian Symposium and Workshops on Algorithm Theory %Z date of event: 2016-06-22 - 2016-06-24 %C Reykjavik, Iceland %B 15th Scandinavian Symposium and Workshops on Algorithm Theory %E Pagh, Rasmus %P 1 - 14 %Z sequence number: 14 %I Schloss Dagstuhl %@ 978-3-95977-011-8 %B Leibniz International Proceedings in Informatics %N 53 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2016/6036/http://drops.dagstuhl.de/doku/urheberrecht1.html
[211]
M. Kerber, D. Morozov, and A. Nigmetov, “Geometry Helps to Compare Persistence Diagrams,” in Proceedings of the Eighteenth Workshop on Algorithm Engineering and Experiments (ALENEX 2016), Arlington, VA, USA, 2016.
Export
BibTeX
@inproceedings{KerberALENEX2016, TITLE = {Geometry Helps to Compare Persistence Diagrams}, AUTHOR = {Kerber, Michael and Morozov, Dimitriy and Nigmetov, Amur}, LANGUAGE = {eng}, ISBN = {978-1-61197-431-7}, DOI = {10.1137/1.9781611974317.9}, PUBLISHER = {SIAM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {Proceedings of the Eighteenth Workshop on Algorithm Engineering and Experiments (ALENEX 2016)}, EDITOR = {Goodrich, Michael and Mitzenmacher, Michael}, PAGES = {103--112}, ADDRESS = {Arlington, VA, USA}, }
Endnote
%0 Conference Proceedings %A Kerber, Michael %A Morozov, Dimitriy %A Nigmetov, Amur %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Geometry Helps to Compare Persistence Diagrams : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0259-9 %R 10.1137/1.9781611974317.9 %D 2016 %B Eighteenth Workshop on Algorithm Engineering and Experiments %Z date of event: 2016-01-10 - 2016-01-10 %C Arlington, VA, USA %B Proceedings of the Eighteenth Workshop on Algorithm Engineering and Experiments %E Goodrich, Michael; Mitzenmacher, Michael %P 103 - 112 %I SIAM %@ 978-1-61197-431-7
[212]
M. Kerber, D. R. Sheehy, and P. Skraba, “Persistent Homology and Nested Dissection,” in Proceedings of the Twenty-Seventh ACM-SIAM Annual Symposium on Discrete Algorithms (SODA 2016), Arlington, VA, USA, 2016.
Export
BibTeX
@inproceedings{KerberSODA2015, TITLE = {Persistent Homology and Nested Dissection}, AUTHOR = {Kerber, Michael and Sheehy, Donald R. and Skraba, Primoz}, LANGUAGE = {eng}, ISBN = {978-1-61197-433-1}, DOI = {10.1137/1.9781611974331.ch86}, PUBLISHER = {SIAM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Proceedings of the Twenty-Seventh ACM-SIAM Annual Symposium on Discrete Algorithms (SODA 2016)}, EDITOR = {Krauthgamer, Robert}, PAGES = {1234--1245}, ADDRESS = {Arlington, VA, USA}, }
Endnote
%0 Conference Proceedings %A Kerber, Michael %A Sheehy, Donald R. %A Skraba, Primoz %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Persistent Homology and Nested Dissection : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0267-9 %R 10.1137/1.9781611974331.ch86 %D 2016 %B Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2016-01-10 - 2016-01-12 %C Arlington, VA, USA %B Proceedings of the Twenty-Seventh ACM-SIAM Annual Symposium on Discrete Algorithms %E Krauthgamer, Robert %P 1234 - 1245 %I SIAM %@ 978-1-61197-433-1
[213]
M. Kerber, D. Morozov, and A. Nigmetov, “Geometry Helps to Compare Persistence Diagrams,” 2016. [Online]. Available: http://arxiv.org/abs/1606.03357. (arXiv: 1606.03357)
Abstract
Exploiting geometric structure to improve the asymptotic complexity of discrete assignment problems is a well-studied subject. In contrast, the practical advantages of using geometry for such problems have not been explored. We implement geometric variants of the Hopcroft--Karp algorithm for bottleneck matching (based on previous work by Efrat el al.) and of the auction algorithm by Bertsekas for Wasserstein distance computation. Both implementations use k-d trees to replace a linear scan with a geometric proximity query. Our interest in this problem stems from the desire to compute distances between persistence diagrams, a problem that comes up frequently in topological data analysis. We show that our geometric matching algorithms lead to a substantial performance gain, both in running time and in memory consumption, over their purely combinatorial counterparts. Moreover, our implementation significantly outperforms the only other implementation available for comparing persistence diagrams.
Export
BibTeX
@online{KerberarXiv2016, TITLE = {Geometry Helps to Compare Persistence Diagrams}, AUTHOR = {Kerber, Michael and Morozov, Dmitriy and Nigmetov, Arnur}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1606.03357}, EPRINT = {1606.03357}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Exploiting geometric structure to improve the asymptotic complexity of discrete assignment problems is a well-studied subject. In contrast, the practical advantages of using geometry for such problems have not been explored. We implement geometric variants of the Hopcroft--Karp algorithm for bottleneck matching (based on previous work by Efrat el al.) and of the auction algorithm by Bertsekas for Wasserstein distance computation. Both implementations use k-d trees to replace a linear scan with a geometric proximity query. Our interest in this problem stems from the desire to compute distances between persistence diagrams, a problem that comes up frequently in topological data analysis. We show that our geometric matching algorithms lead to a substantial performance gain, both in running time and in memory consumption, over their purely combinatorial counterparts. Moreover, our implementation significantly outperforms the only other implementation available for comparing persistence diagrams.}, }
Endnote
%0 Report %A Kerber, Michael %A Morozov, Dmitriy %A Nigmetov, Arnur %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Geometry Helps to Compare Persistence Diagrams : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-029A-6 %U http://arxiv.org/abs/1606.03357 %D 2016 %X Exploiting geometric structure to improve the asymptotic complexity of discrete assignment problems is a well-studied subject. In contrast, the practical advantages of using geometry for such problems have not been explored. We implement geometric variants of the Hopcroft--Karp algorithm for bottleneck matching (based on previous work by Efrat el al.) and of the auction algorithm by Bertsekas for Wasserstein distance computation. Both implementations use k-d trees to replace a linear scan with a geometric proximity query. Our interest in this problem stems from the desire to compute distances between persistence diagrams, a problem that comes up frequently in topological data analysis. We show that our geometric matching algorithms lead to a substantial performance gain, both in running time and in memory consumption, over their purely combinatorial counterparts. Moreover, our implementation significantly outperforms the only other implementation available for comparing persistence diagrams. %K Computer Science, Computational Geometry, cs.CG
[214]
T. Kesselheim and A. Tönnis, “Submodular Secretary Problems: Cardinality, Matching, and Linear Constraints,” 2016. [Online]. Available: http://arxiv.org/abs/1607.08805. (arXiv: 1607.08805)
Abstract
We study various generalizations of the secretary problem with submodular objective functions. Generally, a set of requests is revealed step-by-step to an algorithm in random order. For each request, one option has to be selected so as to maximize a monotone submodular function while ensuring feasibility. For our results, we assume that we are given an offline algorithm computing an $\alpha$-approximation for the respective problem. This way, we separate computational limitations from the ones due to the online nature. When only focusing on the online aspect, we can assume $\alpha = 1$. In the submodular secretary problem, feasibility constraints are cardinality constraints. That is, out of a randomly ordered stream of entities, one has to select a subset size $k$. For this problem, we present a $0.31\alpha$-competitive algorithm for all $k$, which asymptotically reaches competitive ratio $\frac{\alpha}{e}$ for large $k$. In submodular secretary matching, one side of a bipartite graph is revealed online. Upon arrival, each node has to be matched permanently to an offline node or discarded irrevocably. We give an $\frac{\alpha}{4}$-competitive algorithm. In both cases, we improve over previously best known competitive ratios, using a generalization of the algorithm for the classic secretary problem. Furthermore, we give an $O(\alpha d^{-\frac{2}{B-1}})$-competitive algorithm for submodular function maximization subject to linear packing constraints. Here, $d$ is the column sparsity, that is the maximal number of none-zero entries in a column of the constraint matrix, and $B$ is the minimal capacity of the constraints. Notably, this bound is independent of the total number of constraints. We improve the algorithm to be $O(\alpha d^{-\frac{1}{B-1}})$-competitive if both $d$ and $B$ are known to the algorithm beforehand.
Export
BibTeX
@online{DBLP:journals/corr/KesselheimT16a, TITLE = {Submodular Secretary Problems: {C}ardinality, Matching, and Linear Constraints}, AUTHOR = {Kesselheim, Thomas and T{\"o}nnis, Andreas}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1607.08805}, EPRINT = {1607.08805}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We study various generalizations of the secretary problem with submodular objective functions. Generally, a set of requests is revealed step-by-step to an algorithm in random order. For each request, one option has to be selected so as to maximize a monotone submodular function while ensuring feasibility. For our results, we assume that we are given an offline algorithm computing an $\alpha$-approximation for the respective problem. This way, we separate computational limitations from the ones due to the online nature. When only focusing on the online aspect, we can assume $\alpha = 1$. In the submodular secretary problem, feasibility constraints are cardinality constraints. That is, out of a randomly ordered stream of entities, one has to select a subset size $k$. For this problem, we present a $0.31\alpha$-competitive algorithm for all $k$, which asymptotically reaches competitive ratio $\frac{\alpha}{e}$ for large $k$. In submodular secretary matching, one side of a bipartite graph is revealed online. Upon arrival, each node has to be matched permanently to an offline node or discarded irrevocably. We give an $\frac{\alpha}{4}$-competitive algorithm. In both cases, we improve over previously best known competitive ratios, using a generalization of the algorithm for the classic secretary problem. Furthermore, we give an $O(\alpha d^{-\frac{2}{B-1}})$-competitive algorithm for submodular function maximization subject to linear packing constraints. Here, $d$ is the column sparsity, that is the maximal number of none-zero entries in a column of the constraint matrix, and $B$ is the minimal capacity of the constraints. Notably, this bound is independent of the total number of constraints. We improve the algorithm to be $O(\alpha d^{-\frac{1}{B-1}})$-competitive if both $d$ and $B$ are known to the algorithm beforehand.}, }
Endnote
%0 Report %A Kesselheim, Thomas %A T&#246;nnis, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Submodular Secretary Problems: Cardinality, Matching, and Linear Constraints : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E71-3 %U http://arxiv.org/abs/1607.08805 %D 2016 %X We study various generalizations of the secretary problem with submodular objective functions. Generally, a set of requests is revealed step-by-step to an algorithm in random order. For each request, one option has to be selected so as to maximize a monotone submodular function while ensuring feasibility. For our results, we assume that we are given an offline algorithm computing an $\alpha$-approximation for the respective problem. This way, we separate computational limitations from the ones due to the online nature. When only focusing on the online aspect, we can assume $\alpha = 1$. In the submodular secretary problem, feasibility constraints are cardinality constraints. That is, out of a randomly ordered stream of entities, one has to select a subset size $k$. For this problem, we present a $0.31\alpha$-competitive algorithm for all $k$, which asymptotically reaches competitive ratio $\frac{\alpha}{e}$ for large $k$. In submodular secretary matching, one side of a bipartite graph is revealed online. Upon arrival, each node has to be matched permanently to an offline node or discarded irrevocably. We give an $\frac{\alpha}{4}$-competitive algorithm. In both cases, we improve over previously best known competitive ratios, using a generalization of the algorithm for the classic secretary problem. Furthermore, we give an $O(\alpha d^{-\frac{2}{B-1}})$-competitive algorithm for submodular function maximization subject to linear packing constraints. Here, $d$ is the column sparsity, that is the maximal number of none-zero entries in a column of the constraint matrix, and $B$ is the minimal capacity of the constraints. Notably, this bound is independent of the total number of constraints. We improve the algorithm to be $O(\alpha d^{-\frac{1}{B-1}})$-competitive if both $d$ and $B$ are known to the algorithm beforehand. %K Computer Science, Data Structures and Algorithms, cs.DS
[215]
T. Kesselheim and A. Tönnis, “Think Eternally: Improved Algorithms for the Temp Secretary Problem and Extensions,” 2016. [Online]. Available: http://arxiv.org/abs/1606.06926. (arXiv: 1606.06926)
Abstract
The \emph{Temp Secretary Problem} was recently introduced by Fiat et al. It is a generalization of the Secretary Problem, in which commitments are temporary for a fixed duration. We present a simple online algorithm with improved performance guarantees for cases already considered by Fiat et al.\ and give competitive ratios for new generalizations of the problem. In the classical setting, where candidates have identical contract durations $\gamma \ll 1$ and we are allowed to hire up to $B$ candidates simultaneously, our algorithm is $(\frac{1}{2} - O(\sqrt{\gamma}))$-competitive. For large $B$, the bound improves to $1 - O\left(\frac{1}{\sqrt{B}}\right) - O(\sqrt{\gamma})$. Furthermore we generalize the problem from cardinality constraints towards general packing constraints. We achieve a competitive ratio of $1 - O\left(\sqrt{\frac{(1+\log d + \log B)}{B}}\right) -O(\sqrt{\gamma})$, where $d$ is the sparsity of the constraint matrix and $B$ is generalized to the capacity ratio of linear constraints. Additionally we extend the problem towards arbitrary hiring durations. Our algorithmic approach is a relaxation that aggregates all temporal constraints into a non-temporal constraint. Then we apply a linear scaling algorithm that, on every arrival, computes a tentative solution on the input that is known up to this point. This tentative solution uses the non-temporal, relaxed constraints scaled down linearly by the amount of time that has already passed.
Export
BibTeX
@online{DBLP:journals/corr/KesselheimT16, TITLE = {Think Eternally: {I}mproved Algorithms for the Temp Secretary Problem and Extensions}, AUTHOR = {Kesselheim, Thomas and T{\"o}nnis, Andreas}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1606.06926}, EPRINT = {1606.06926}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {The \emph{Temp Secretary Problem} was recently introduced by Fiat et al. It is a generalization of the Secretary Problem, in which commitments are temporary for a fixed duration. We present a simple online algorithm with improved performance guarantees for cases already considered by Fiat et al.\ and give competitive ratios for new generalizations of the problem. In the classical setting, where candidates have identical contract durations $\gamma \ll 1$ and we are allowed to hire up to $B$ candidates simultaneously, our algorithm is $(\frac{1}{2} -- O(\sqrt{\gamma}))$-competitive. For large $B$, the bound improves to $1 -- O\left(\frac{1}{\sqrt{B}}\right) -- O(\sqrt{\gamma})$. Furthermore we generalize the problem from cardinality constraints towards general packing constraints. We achieve a competitive ratio of $1 - O\left(\sqrt{\frac{(1+\log d + \log B)}{B}}\right) -O(\sqrt{\gamma})$, where $d$ is the sparsity of the constraint matrix and $B$ is generalized to the capacity ratio of linear constraints. Additionally we extend the problem towards arbitrary hiring durations. Our algorithmic approach is a relaxation that aggregates all temporal constraints into a non-temporal constraint. Then we apply a linear scaling algorithm that, on every arrival, computes a tentative solution on the input that is known up to this point. This tentative solution uses the non-temporal, relaxed constraints scaled down linearly by the amount of time that has already passed.}, }
Endnote
%0 Report %A Kesselheim, Thomas %A T&#246;nnis, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Think Eternally: Improved Algorithms for the Temp Secretary Problem and Extensions : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E68-A %U http://arxiv.org/abs/1606.06926 %D 2016 %X The \emph{Temp Secretary Problem} was recently introduced by Fiat et al. It is a generalization of the Secretary Problem, in which commitments are temporary for a fixed duration. We present a simple online algorithm with improved performance guarantees for cases already considered by Fiat et al.\ and give competitive ratios for new generalizations of the problem. In the classical setting, where candidates have identical contract durations $\gamma \ll 1$ and we are allowed to hire up to $B$ candidates simultaneously, our algorithm is $(\frac{1}{2} - O(\sqrt{\gamma}))$-competitive. For large $B$, the bound improves to $1 - O\left(\frac{1}{\sqrt{B}}\right) - O(\sqrt{\gamma})$. Furthermore we generalize the problem from cardinality constraints towards general packing constraints. We achieve a competitive ratio of $1 - O\left(\sqrt{\frac{(1+\log d + \log B)}{B}}\right) -O(\sqrt{\gamma})$, where $d$ is the sparsity of the constraint matrix and $B$ is generalized to the capacity ratio of linear constraints. Additionally we extend the problem towards arbitrary hiring durations. Our algorithmic approach is a relaxation that aggregates all temporal constraints into a non-temporal constraint. Then we apply a linear scaling algorithm that, on every arrival, computes a tentative solution on the input that is known up to this point. This tentative solution uses the non-temporal, relaxed constraints scaled down linearly by the amount of time that has already passed. %K Computer Science, Data Structures and Algorithms, cs.DS
[216]
T. Kesselheim and A. Tönnis, “Think Eternally: Improved Algorithms for the Temp Secretary Problem and Extensions,” in 24th Annual European Symposium on Algorithms (ESA 2016), Aarhus, Denmark, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/esa/KesselheimT16, TITLE = {Think Eternally: {I}mproved Algorithms for the Temp Secretary Problem and Extensions}, AUTHOR = {Kesselheim, Thomas and T{\"o}nnis, Andreas}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-015-6}, URL = {urn:nbn:de:0030-drops-63966}, DOI = {10.4230/LIPIcs.ESA.2016.54}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {24th Annual European Symposium on Algorithms (ESA 2016)}, EDITOR = {Sankowski, Piotr and Zaroliagis, Christos}, PAGES = {1--17}, EID = {54}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {57}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Kesselheim, Thomas %A T&#246;nnis, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Think Eternally: Improved Algorithms for the Temp Secretary Problem and Extensions : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-4E60-9 %U urn:nbn:de:0030-drops-63966 %R 10.4230/LIPIcs.ESA.2016.54 %D 2016 %B 24th Annual European Symposium on Algorithms %Z date of event: 2016-08-22 - 2016-08-26 %C Aarhus, Denmark %B 24th Annual European Symposium on Algorithms %E Sankowski, Piotr; Zaroliagis, Christos %P 1 - 17 %Z sequence number: 54 %I Schloss Dagstuhl %@ 978-3-95977-015-6 %B Leibniz International Proceedings in Informatics %N 57 %@ false %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2016/6396/
[217]
P. Khanchandani and C. Lenzen, “Self-stabilizing Byzantine Clock Synchronization with Optimal Precision,” in Stabilization, Safety, and Security of Distributed Systems (SSS 2016), Lyon, France, 2016.
Export
BibTeX
@inproceedings{KhanchandaniSSS2016, TITLE = {Self-stabilizing {B}yzantine Clock Synchronization with Optimal Precision}, AUTHOR = {Khanchandani, Pankaj and Lenzen, Christoph}, LANGUAGE = {eng}, ISBN = {978-3-319-49258-2}, DOI = {10.1007/978-3-319-49259-9_18}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Stabilization, Safety, and Security of Distributed Systems (SSS 2016)}, EDITOR = {Bonakdarpour, Borzoo and Petit, Franck}, PAGES = {213--230}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10083}, ADDRESS = {Lyon, France}, }
Endnote
%0 Conference Proceedings %A Khanchandani, Pankaj %A Lenzen, Christoph %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Self-stabilizing Byzantine Clock Synchronization with Optimal Precision : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-571E-9 %R 10.1007/978-3-319-49259-9_18 %D 2016 %B 18th International Symposium on Stabilization, Safety, and Security of Distributed Systems %Z date of event: 2016-11-08 - 2016-11-10 %C Lyon, France %B Stabilization, Safety, and Security of Distributed Systems %E Bonakdarpour, Borzoo; Petit, Franck %P 213 - 230 %I Springer %@ 978-3-319-49258-2 %B Lecture Notes in Computer Science %N 10083
[218]
P. Khanchandani and C. Lenzen, “Self-stabilizing Byzantine Clock Synchronization with Optimal Precision,” 2016. [Online]. Available: http://arxiv.org/abs/1609.09281. (arXiv: 1609.09281)
Abstract
We revisit the approach to Byzantine fault-tolerant clock synchronization based on approximate agreement introduced by Lynch and Welch. Our contribution is threefold: (1) We provide a slightly refined variant of the algorithm yielding improved bounds on the skew that can be achieved and the sustainable frequency offsets. (2) We show how to extend the technique to also synchronize clock rates. This permits less frequent communication without significant loss of precision, provided that clock rates change sufficiently slowly. (3) We present a coupling scheme that allows to make these algorithms self-stabilizing while preserving their high precision. The scheme utilizes a low-precision, but self-stabilizing algorithm for the purpose of recovery.
Export
BibTeX
@online{Khanchandani2016, TITLE = {Self-stabilizing Byzantine Clock Synchronization with Optimal Precision}, AUTHOR = {Khanchandani, Pankaj and Lenzen, Christoph}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1609.09281}, EPRINT = {1609.09281}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We revisit the approach to Byzantine fault-tolerant clock synchronization based on approximate agreement introduced by Lynch and Welch. Our contribution is threefold: (1) We provide a slightly refined variant of the algorithm yielding improved bounds on the skew that can be achieved and the sustainable frequency offsets. (2) We show how to extend the technique to also synchronize clock rates. This permits less frequent communication without significant loss of precision, provided that clock rates change sufficiently slowly. (3) We present a coupling scheme that allows to make these algorithms self-stabilizing while preserving their high precision. The scheme utilizes a low-precision, but self-stabilizing algorithm for the purpose of recovery.}, }
Endnote
%0 Report %A Khanchandani, Pankaj %A Lenzen, Christoph %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Self-stabilizing Byzantine Clock Synchronization with Optimal Precision : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-8437-C %U http://arxiv.org/abs/1609.09281 %D 2016 %X We revisit the approach to Byzantine fault-tolerant clock synchronization based on approximate agreement introduced by Lynch and Welch. Our contribution is threefold: (1) We provide a slightly refined variant of the algorithm yielding improved bounds on the skew that can be achieved and the sustainable frequency offsets. (2) We show how to extend the technique to also synchronize clock rates. This permits less frequent communication without significant loss of precision, provided that clock rates change sufficiently slowly. (3) We present a coupling scheme that allows to make these algorithms self-stabilizing while preserving their high precision. The scheme utilizes a low-precision, but self-stabilizing algorithm for the purpose of recovery. %K Computer Science, Distributed, Parallel, and Cluster Computing, cs.DC
[219]
A. Kinali, F. Huemer, and C. Lenzen, “Fault-Tolerant Clock Synchronization with High Precision,” in ISVLSI 2016, Pittsburgh, PA, USA, 2016.
Export
BibTeX
@inproceedings{KinaliVLSI2016, TITLE = {Fault-Tolerant Clock Synchronization with High Precision}, AUTHOR = {Kinali, Attila and Huemer, Florian and Lenzen, Christoph}, LANGUAGE = {eng}, ISBN = {78-1-4673-9038-5}, DOI = {10.1109/ISVLSI.2016.88}, PUBLISHER = {IEEE Computer Society}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {ISVLSI 2016}, PAGES = {490--495}, ADDRESS = {Pittsburgh, PA, USA}, }
Endnote
%0 Conference Proceedings %A Kinali, Attila %A Huemer, Florian %A Lenzen, Christoph %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Fault-Tolerant Clock Synchronization with High Precision : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-842C-6 %R 10.1109/ISVLSI.2016.88 %D 2016 %B IEEE Computer Society Annual Symposium on VLSI %Z date of event: 2016-07-11 - 2016-07-13 %C Pittsburgh, PA, USA %B ISVLSI 2016 %P 490 - 495 %I IEEE Computer Society %@ 78-1-4673-9038-5
[220]
A. Kobel, F. Rouillier, and M. Sagraloff, “Computing Real Roots of Real Polynomials ... and now For Real!,” 2016. [Online]. Available: http://arxiv.org/abs/1605.00410. (arXiv: 1605.00410)
Abstract
Very recent work introduces an asymptotically fast subdivision algorithm, denoted ANewDsc, for isolating the real roots of a univariate real polynomial. The method combines Descartes' Rule of Signs to test intervals for the existence of roots, Newton iteration to speed up convergence against clusters of roots, and approximate computation to decrease the required precision. It achieves record bounds on the worst-case complexity for the considered problem, matching the complexity of Pan's method for computing all complex roots and improving upon the complexity of other subdivision methods by several magnitudes. In the article at hand, we report on an implementation of ANewDsc on top of the RS root isolator. RS is a highly efficient realization of the classical Descartes method and currently serves as the default real root solver in Maple. We describe crucial design changes within ANewDsc and RS that led to a high-performance implementation without harming the theoretical complexity of the underlying algorithm. With an excerpt of our extensive collection of benchmarks, available online at http://anewdsc.mpi-inf.mpg.de/, we illustrate that the theoretical gain in performance of ANewDsc over other subdivision methods also transfers into practice. These experiments also show that our new implementation outperforms both RS and mature competitors by magnitudes for notoriously hard instances with clustered roots. For all other instances, we avoid almost any overhead by integrating additional optimizations and heuristics.
Export
BibTeX
@online{KobelarXiv2016, TITLE = {Computing Real Roots of Real Polynomials \ldots{} and now For Real!}, AUTHOR = {Kobel, Alexander and Rouillier, Fabrice and Sagraloff, Michael}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1605.00410}, EPRINT = {1605.00410}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Very recent work introduces an asymptotically fast subdivision algorithm, denoted ANewDsc, for isolating the real roots of a univariate real polynomial. The method combines Descartes' Rule of Signs to test intervals for the existence of roots, Newton iteration to speed up convergence against clusters of roots, and approximate computation to decrease the required precision. It achieves record bounds on the worst-case complexity for the considered problem, matching the complexity of Pan's method for computing all complex roots and improving upon the complexity of other subdivision methods by several magnitudes. In the article at hand, we report on an implementation of ANewDsc on top of the RS root isolator. RS is a highly efficient realization of the classical Descartes method and currently serves as the default real root solver in Maple. We describe crucial design changes within ANewDsc and RS that led to a high-performance implementation without harming the theoretical complexity of the underlying algorithm. With an excerpt of our extensive collection of benchmarks, available online at http://anewdsc.mpi-inf.mpg.de/, we illustrate that the theoretical gain in performance of ANewDsc over other subdivision methods also transfers into practice. These experiments also show that our new implementation outperforms both RS and mature competitors by magnitudes for notoriously hard instances with clustered roots. For all other instances, we avoid almost any overhead by integrating additional optimizations and heuristics.}, }
Endnote
%0 Report %A Kobel, Alexander %A Rouillier, Fabrice %A Sagraloff, Michael %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Computing Real Roots of Real Polynomials ... and now For Real! : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-029D-F %U http://arxiv.org/abs/1605.00410 %D 2016 %X Very recent work introduces an asymptotically fast subdivision algorithm, denoted ANewDsc, for isolating the real roots of a univariate real polynomial. The method combines Descartes' Rule of Signs to test intervals for the existence of roots, Newton iteration to speed up convergence against clusters of roots, and approximate computation to decrease the required precision. It achieves record bounds on the worst-case complexity for the considered problem, matching the complexity of Pan's method for computing all complex roots and improving upon the complexity of other subdivision methods by several magnitudes. In the article at hand, we report on an implementation of ANewDsc on top of the RS root isolator. RS is a highly efficient realization of the classical Descartes method and currently serves as the default real root solver in Maple. We describe crucial design changes within ANewDsc and RS that led to a high-performance implementation without harming the theoretical complexity of the underlying algorithm. With an excerpt of our extensive collection of benchmarks, available online at http://anewdsc.mpi-inf.mpg.de/, we illustrate that the theoretical gain in performance of ANewDsc over other subdivision methods also transfers into practice. These experiments also show that our new implementation outperforms both RS and mature competitors by magnitudes for notoriously hard instances with clustered roots. For all other instances, we avoid almost any overhead by integrating additional optimizations and heuristics. %K Computer Science, Mathematical Software, cs.MS,Computer Science, Numerical Analysis, cs.NA,Computer Science, Symbolic Computation, cs.SC,Mathematics, Numerical Analysis, math.NA,
[221]
A. Kobel, F. Rouillier, and M. Sagraloff, “Computing Real Roots of Real Polynomials ... and now For Real!,” in ISSAC 2016, 41st International Symposium on Symbolic and Algebraic Computation, Waterloo, Canada, 2016.
Export
BibTeX
@inproceedings{KobelISSAC2016, TITLE = {Computing Real Roots of Real Polynomials \ldots{} and now For Real!}, AUTHOR = {Kobel, Alexander and Rouillier, Fabrice and Sagraloff, Michael}, LANGUAGE = {eng}, ISBN = {978-1-4503-4380-0}, DOI = {10.1145/2930889.2930937}, PUBLISHER = {ACM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {ISSAC 2016, 41st International Symposium on Symbolic and Algebraic Computation}, EDITOR = {Rosenkranz, Markus}, PAGES = {303--310}, ADDRESS = {Waterloo, Canada}, }
Endnote
%0 Conference Proceedings %A Kobel, Alexander %A Rouillier, Fabrice %A Sagraloff, Michael %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Computing Real Roots of Real Polynomials ... and now For Real! : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-02AB-F %R 10.1145/2930889.2930937 %D 2016 %B 41st International Symposium on Symbolic and Algebraic Computation %Z date of event: 2016-06-19 - 2016-06-22 %C Waterloo, Canada %B ISSAC 2016 %E Rosenkranz, Markus %P 303 - 310 %I ACM %@ 978-1-4503-4380-0
[222]
P. Kolev and K. Mehlhorn, “A Note On Spectral Clustering,” in 24th Annual European Symposium on Algorithms (ESA 2016), Aarhus, Denmark, 2016.
Export
BibTeX
@inproceedings{KolevESA2016, TITLE = {A Note On Spectral Clustering}, AUTHOR = {Kolev, Pavel and Mehlhorn, Kurt}, LANGUAGE = {eng}, ISSN = {1868-8969}, ISBN = {978-3-95977-015-6}, URL = {urn:nbn:de:0030-drops-63994}, DOI = {10.4230/LIPIcs.ESA.2016.57}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {24th Annual European Symposium on Algorithms (ESA 2016)}, EDITOR = {Sankowski, Piotr and Zaroliagis, Christos}, PAGES = {1--14}, EID = {57}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {57}, ADDRESS = {Aarhus, Denmark}, }
Endnote
%0 Conference Proceedings %A Kolev, Pavel %A Mehlhorn, Kurt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T A Note On Spectral Clustering : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-4797-D %U urn:nbn:de:0030-drops-63994 %R 10.4230/LIPIcs.ESA.2016.57 %D 2016 %B 24th Annual European Symposium on Algorithms %Z date of event: 2016-08-22 - 2016-08-26 %C Aarhus, Denmark %B 24th Annual European Symposium on Algorithms %E Sankowski, Piotr; Zaroliagis, Christos %P 1 - 14 %Z sequence number: 57 %I Schloss Dagstuhl %@ 978-3-95977-015-6 %B Leibniz International Proceedings in Informatics %N 57 %@ false %U http://drops.dagstuhl.de/opus/volltexte/2016/6399/http://drops.dagstuhl.de/doku/urheberrecht1.html
[223]
G. Kol, S. Moran, A. Shpilka, and A. Yehudayoff, “Direct Sum Fails for Zero-Error Average Communication,” Algorithmica, vol. 76, no. 3, 2016.
Export
BibTeX
@article{Kol2016, TITLE = {Direct Sum Fails for Zero-Error Average Communication}, AUTHOR = {Kol, Gillat and Moran, Shay and Shpilka, Amir and Yehudayoff, Amir}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-016-0144-9}, PUBLISHER = {Springer}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Algorithmica}, VOLUME = {76}, NUMBER = {3}, PAGES = {782--795}, }
Endnote
%0 Journal Article %A Kol, Gillat %A Moran, Shay %A Shpilka, Amir %A Yehudayoff, Amir %+ External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T Direct Sum Fails for Zero-Error Average Communication : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-41D1-1 %R 10.1007/s00453-016-0144-9 %7 2016-03-28 %D 2016 %J Algorithmica %V 76 %N 3 %& 782 %P 782 - 795 %I Springer %C New York, NY %@ false
[224]
S. Kratsch, D. Marx, and M. Wahlström, “Parameterized Complexity and Kernelizability of Max Ones and Exact Ones Problems,” ACM Transactions on Computation Theory, vol. 8, no. 1, 2016.
Export
BibTeX
@article{Kratsch2016, TITLE = {Parameterized Complexity and Kernelizability of {Max Ones} and {Exact Ones} Problems}, AUTHOR = {Kratsch, Stefan and Marx, D{\'a}niel and Wahlstr{\"o}m, Magnus}, LANGUAGE = {eng}, DOI = {10.1145/2858787}, PUBLISHER = {ACM}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {ACM Transactions on Computation Theory}, VOLUME = {8}, NUMBER = {1}, EID = {1}, }
Endnote
%0 Journal Article %A Kratsch, Stefan %A Marx, D&#225;niel %A Wahlstr&#246;m, Magnus %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Parameterized Complexity and Kernelizability of Max Ones and Exact Ones Problems : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002A-4326-D %R 10.1145/2858787 %7 2016 %D 2016 %J ACM Transactions on Computation Theory %O TOCT %V 8 %N 1 %Z sequence number: 1 %I ACM %C New York, NY
[225]
S. Kratsch, G. Philip, and S. Ray, “Point Line Cover: The Easy Kernel is Essentially Tight,” ACM Transactions on Algorithms, vol. 12, no. 3, 2016.
Export
BibTeX
@article{Kratsch2016TALG, TITLE = {Point Line Cover: The Easy Kernel is Essentially Tight}, AUTHOR = {Kratsch, Stefan and Philip, Geevarghese and Ray, Saurabh}, LANGUAGE = {eng}, ISSN = {1549-6325}, DOI = {10.1145/2832912}, PUBLISHER = {Association for Computing Machinery}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {ACM Transactions on Algorithms}, VOLUME = {12}, NUMBER = {3}, EID = {40}, }
Endnote
%0 Journal Article %A Kratsch, Stefan %A Philip, Geevarghese %A Ray, Saurabh %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Point Line Cover: The Easy Kernel is Essentially Tight : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-195B-8 %R 10.1145/2832912 %7 2016 %D 2016 %J ACM Transactions on Algorithms %V 12 %N 3 %Z sequence number: 40 %I Association for Computing Machinery %C New York, NY %@ false
[226]
A. Krebs, N. Limaye, M. Mahajan, and K. Sreenivasaiah, “Small Depth Proof Systems,” ACM Transactions on Computation Theory, vol. 9, no. 1, 2016.
Export
BibTeX
@article{KrebsLMS16, TITLE = {Small Depth Proof Systems}, AUTHOR = {Krebs, Andreas and Limaye, Nutan and Mahajan, Meena and Sreenivasaiah, Karteek}, LANGUAGE = {eng}, ISSN = {1942-3454}, DOI = {10.1145/2956229}, PUBLISHER = {ACM}, ADDRESS = {New York, NY}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {ACM Transactions on Computation Theory}, VOLUME = {9}, NUMBER = {1}, PAGES = {1--26}, EID = {2}, }
Endnote
%0 Journal Article %A Krebs, Andreas %A Limaye, Nutan %A Mahajan, Meena %A Sreenivasaiah, Karteek %+ External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Small Depth Proof Systems : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5F76-5 %R 10.1145/2956229 %7 2016 %D 2016 %J ACM Transactions on Computation Theory %O TOCT %V 9 %N 1 %& 1 %P 1 - 26 %Z sequence number: 2 %I ACM %C New York, NY %@ false
[227]
S. Krinninger, “Schnellere Approximationsalgorithmen zur Partiell-Dynamischen Berechnung Kürzester Wege,” in Ausgezeichnete Informatikdissertationen 2015, Bonn: GI, 2016.
Export
BibTeX
@incollection{KrinLNI_Diss16, TITLE = {{Schnellere Approximationsalgorithmen zur Partiell-Dynamischen Berechnung K{\"u}rzester Wege}}, AUTHOR = {Krinninger, Sebastian}, LANGUAGE = {deu}, ISBN = {978-3-88579-975-7}, PUBLISHER = {GI}, ADDRESS = {Bonn}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Ausgezeichnete Informatikdissertationen 2015}, EDITOR = {H{\"o}lldobler, Steffen}, PAGES = {279--288}, SERIES = {Lecture Notes in Informatics -- Dissertations}, VOLUME = {16}, }
Endnote
%0 Book Section %A Krinninger, Sebastian %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Schnellere Approximationsalgorithmen zur Partiell-Dynamischen Berechnung K&#252;rzester Wege : %G deu %U http://hdl.handle.net/11858/00-001M-0000-002C-5117-B %D 2016 %B Ausgezeichnete Informatikdissertationen 2015 %E H&#246;lldobler, Steffen %P 279 - 288 %I GI %C Bonn %@ 978-3-88579-975-7 %S Lecture Notes in Informatics - Dissertations %N 16
[228]
M. Künnemann, “Tight(er) Bounds for Similarity Measures, Smoothed Approximation and Broadcasting,” Universität des Saarlandes, Saarbrücken, 2016.
Export
BibTeX
@phdthesis{Kuennemannphd2016, TITLE = {Tight(er) Bounds for Similarity Measures, Smoothed Approximation and Broadcasting}, AUTHOR = {K{\"u}nnemann, Marvin}, LANGUAGE = {eng}, URL = {urn:nbn:de:bsz:291-scidok-65991}, SCHOOL = {Universit{\"a}t des Saarlandes}, ADDRESS = {Saarbr{\"u}cken}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, }
Endnote
%0 Thesis %A K&#252;nnemann, Marvin %Y Doerr, Benjamin %A referee: Mehlhorn, Kurt %A referee: Welzl, Emo %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society International Max Planck Research School, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Tight(er) Bounds for Similarity Measures, Smoothed Approximation and Broadcasting : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-423A-3 %U urn:nbn:de:bsz:291-scidok-65991 %I Universit&#228;t des Saarlandes %C Saarbr&#252;cken %D 2016 %P XI, 223 p. %V phd %9 phd %U http://scidok.sulb.uni-saarland.de/volltexte/2016/6599/http://scidok.sulb.uni-saarland.de/doku/lic_ohne_pod.php?la=de
[229]
A. Kurpisz, M. Mastrolilli, C. Mathieu, T. Mömke, V. Verdugo, and A. Wiese, “Semidefinite and Linear Programming Integrality Gaps for Scheduling Identical Machines,” in Integer Programming and Combinatorial Optimization (IPCO 2016), Liège, Belgium, 2016.
Export
BibTeX
@inproceedings{KurpiszIPCO2016, TITLE = {Semidefinite and Linear Programming Integrality Gaps for Scheduling Identical Machines}, AUTHOR = {Kurpisz, Adam and Mastrolilli, Monaldo and Mathieu, Claire and M{\"o}mke, Tobias and Verdugo, Victor and Wiese, Andreas}, LANGUAGE = {eng}, ISBN = {978-3-319-33460-8}, DOI = {10.1007/978-3-319-33461-5_13}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Integer Programming and Combinatorial Optimization (IPCO 2016)}, EDITOR = {Louveaux, Quentin and Skutella, Martin}, PAGES = {152--163}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {9682}, ADDRESS = {Li{\`e}ge, Belgium}, }
Endnote
%0 Conference Proceedings %A Kurpisz, Adam %A Mastrolilli, Monaldo %A Mathieu, Claire %A M&#246;mke, Tobias %A Verdugo, Victor %A Wiese, Andreas %+ External Organizations External Organizations External Organizations External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Semidefinite and Linear Programming Integrality Gaps for Scheduling Identical Machines : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-0240-0 %R 10.1007/978-3-319-33461-5_13 %D 2016 %B 18th Conference on Integer Programming and Combinatorial Optimization %Z date of event: 2016-06-01 - 2016-06-03 %C Li&#232;ge, Belgium %B Integer Programming and Combinatorial Optimization %E Louveaux, Quentin; Skutella, Martin %P 152 - 163 %I Springer %@ 978-3-319-33460-8 %B Lecture Notes in Computer Science %N 9682
[230]
C. Lenzen and J. Rybicki, “Near-Optimal Self-Stabilising Counting and Firing Squads,” 2016. [Online]. Available: http://arxiv.org/abs/1608.00214. (arXiv: 1608.00214)
Abstract
Consider a fully-connected synchronous distributed system consisting of $n$ nodes, where up to $f$ nodes may be faulty and every node starts in an arbitrary initial state. In the synchronous counting problem, all nodes need to eventually agree on a counter that is increased by one modulo some $C$ in each round. In the self-stabilising firing squad problem, the task is to eventually guarantee that all non-faulty nodes have simultaneous responses to external inputs: if a subset of the correct nodes receive an external "go" signal as input, then all correct nodes should agree on a round (in the not-too-distant future) in which to jointly output a "fire" signal. Moreover, no node should generate a "fire" signal without some correct node having previously received a "go" signal as input. We present a framework reducing both tasks to binary consensus at very small cost: we maintain the resilience of the underlying consensus routine, while the stabilisation time and message size are, up to constant factors, bounded by the sum of the cost of the consensus routine for $f$ faults and recursively applying our scheme to $f'<f/2$ faults. For example, we obtain a deterministic algorithm for self-stabilising Byzantine firing squads with optimal resilience $f<n/3$, asymptotically optimal stabilisation and response time $O(f)$, and message size $O(\log f)$. As our framework does not restrict the type of consensus routines used, we also obtain efficient randomised solutions, and it is straightforward to adapt our framework to allow for $f<n/2$ omission or $f<n$ crash faults, respectively. Our results resolve various open questions on the two problems, most prominently whether (communication-efficient) self-stabilising Byzantine firing squads or (randomised) sublinear-time solutions for either problem exist.
Export
BibTeX
@online{Lenzen_arXiv1608.00214, TITLE = {Near-Optimal Self-Stabilising Counting and Firing Squads}, AUTHOR = {Lenzen, Christoph and Rybicki, Joel}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1608.00214}, EPRINT = {1608.00214}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Consider a fully-connected synchronous distributed system consisting of $n$ nodes, where up to $f$ nodes may be faulty and every node starts in an arbitrary initial state. In the synchronous counting problem, all nodes need to eventually agree on a counter that is increased by one modulo some $C$ in each round. In the self-stabilising firing squad problem, the task is to eventually guarantee that all non-faulty nodes have simultaneous responses to external inputs: if a subset of the correct nodes receive an external "go" signal as input, then all correct nodes should agree on a round (in the not-too-distant future) in which to jointly output a "fire" signal. Moreover, no node should generate a "fire" signal without some correct node having previously received a "go" signal as input. We present a framework reducing both tasks to binary consensus at very small cost: we maintain the resilience of the underlying consensus routine, while the stabilisation time and message size are, up to constant factors, bounded by the sum of the cost of the consensus routine for $f$ faults and recursively applying our scheme to $f'<f/2$ faults. For example, we obtain a deterministic algorithm for self-stabilising Byzantine firing squads with optimal resilience $f<n/3$, asymptotically optimal stabilisation and response time $O(f)$, and message size $O(\log f)$. As our framework does not restrict the type of consensus routines used, we also obtain efficient randomised solutions, and it is straightforward to adapt our framework to allow for $f<n/2$ omission or $f<n$ crash faults, respectively. Our results resolve various open questions on the two problems, most prominently whether (communication-efficient) self-stabilising Byzantine firing squads or (randomised) sublinear-time solutions for either problem exist.}, }
Endnote
%0 Report %A Lenzen, Christoph %A Rybicki, Joel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Near-Optimal Self-Stabilising Counting and Firing Squads : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-8434-1 %U http://arxiv.org/abs/1608.00214 %D 2016 %X Consider a fully-connected synchronous distributed system consisting of $n$ nodes, where up to $f$ nodes may be faulty and every node starts in an arbitrary initial state. In the synchronous counting problem, all nodes need to eventually agree on a counter that is increased by one modulo some $C$ in each round. In the self-stabilising firing squad problem, the task is to eventually guarantee that all non-faulty nodes have simultaneous responses to external inputs: if a subset of the correct nodes receive an external "go" signal as input, then all correct nodes should agree on a round (in the not-too-distant future) in which to jointly output a "fire" signal. Moreover, no node should generate a "fire" signal without some correct node having previously received a "go" signal as input. We present a framework reducing both tasks to binary consensus at very small cost: we maintain the resilience of the underlying consensus routine, while the stabilisation time and message size are, up to constant factors, bounded by the sum of the cost of the consensus routine for $f$ faults and recursively applying our scheme to $f'<f/2$ faults. For example, we obtain a deterministic algorithm for self-stabilising Byzantine firing squads with optimal resilience $f<n/3$, asymptotically optimal stabilisation and response time $O(f)$, and message size $O(\log f)$. As our framework does not restrict the type of consensus routines used, we also obtain efficient randomised solutions, and it is straightforward to adapt our framework to allow for $f<n/2$ omission or $f<n$ crash faults, respectively. Our results resolve various open questions on the two problems, most prominently whether (communication-efficient) self-stabilising Byzantine firing squads or (randomised) sublinear-time solutions for either problem exist. %K Computer Science, Distributed, Parallel, and Cluster Computing, cs.DC
[231]
C. Lenzen and J. Rybicki, “Near-Optimal Self-stabilising Counting and Firing Squads,” in Stabilization, Safety, and Security of Distributed Systems (SSS 2016), Lyon, France, 2016.
Export
BibTeX
@inproceedings{LenzenSSS2016, TITLE = {Near-Optimal Self-stabilising Counting and Firing Squads}, AUTHOR = {Lenzen, Christoph and Rybicki, Joel}, LANGUAGE = {eng}, ISBN = {978-3-319-49258-2}, DOI = {10.1007/978-3-319-49259-9_21}, PUBLISHER = {Springer}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Stabilization, Safety, and Security of Distributed Systems (SSS 2016)}, EDITOR = {Bonakdarpour, Borzoo and Petit, Franck}, PAGES = {263--280}, SERIES = {Lecture Notes in Computer Science}, VOLUME = {10083}, ADDRESS = {Lyon, France}, }
Endnote
%0 Conference Proceedings %A Lenzen, Christoph %A Rybicki, Joel %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T Near-Optimal Self-stabilising Counting and Firing Squads : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5728-2 %R 10.1007/978-3-319-49259-9_21 %D 2016 %B 18th International Symposium on Stabilization, Safety, and Security of Distributed Systems %Z date of event: 2016-11-08 - 2016-11-10 %C Lyon, France %B Stabilization, Safety, and Security of Distributed Systems %E Bonakdarpour, Borzoo; Petit, Franck %P 263 - 280 %I Springer %@ 978-3-319-49258-2 %B Lecture Notes in Computer Science %N 10083
[232]
C. Lenzen and R. Wattenhofer, “CLEX: Yet Another Supercomputer Architecture?,” 2016. [Online]. Available: http://arxiv.org/abs/1607.00298. (arXiv: 1607.00298)
Abstract
We propose the CLEX supercomputer topology and routing scheme. We prove that CLEX can utilize a constant fraction of the total bandwidth for point-to-point communication, at delays proportional to the sum of the number of intermediate hops and the maximum physical distance between any two nodes. Moreover, % applying an asymmetric bandwidth assignment to the links, all-to-all communication can be realized $(1+o(1))$-optimally both with regard to bandwidth and delays. This is achieved at node degrees of $n^{\varepsilon}$, for an arbitrary small constant $\varepsilon\in (0,1]$. In contrast, these results are impossible in any network featuring constant or polylogarithmic node degrees. Through simulation, we assess the benefits of an implementation of the proposed communication strategy. Our results indicate that, for a million processors, CLEX can increase bandwidth utilization and reduce average routing path length by at least factors $10$ respectively $5$ in comparison to a torus network. Furthermore, the CLEX communication scheme features several other properties, such as deadlock-freedom, inherent fault-tolerance, and canonical partition into smaller subsystems.
Export
BibTeX
@online{Lenzen_arXiv1607.00298, TITLE = {CLEX: {Y}et Another Supercomputer Architecture?}, AUTHOR = {Lenzen, Christoph and Wattenhofer, Roger}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1607.00298}, EPRINT = {1607.00298}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {We propose the CLEX supercomputer topology and routing scheme. We prove that CLEX can utilize a constant fraction of the total bandwidth for point-to-point communication, at delays proportional to the sum of the number of intermediate hops and the maximum physical distance between any two nodes. Moreover, % applying an asymmetric bandwidth assignment to the links, all-to-all communication can be realized $(1+o(1))$-optimally both with regard to bandwidth and delays. This is achieved at node degrees of $n^{\varepsilon}$, for an arbitrary small constant $\varepsilon\in (0,1]$. In contrast, these results are impossible in any network featuring constant or polylogarithmic node degrees. Through simulation, we assess the benefits of an implementation of the proposed communication strategy. Our results indicate that, for a million processors, CLEX can increase bandwidth utilization and reduce average routing path length by at least factors $10$ respectively $5$ in comparison to a torus network. Furthermore, the CLEX communication scheme features several other properties, such as deadlock-freedom, inherent fault-tolerance, and canonical partition into smaller subsystems.}, }
Endnote
%0 Report %A Lenzen, Christoph %A Wattenhofer, Roger %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations %T CLEX: Yet Another Supercomputer Architecture? : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-8431-7 %U http://arxiv.org/abs/1607.00298 %D 2016 %X We propose the CLEX supercomputer topology and routing scheme. We prove that CLEX can utilize a constant fraction of the total bandwidth for point-to-point communication, at delays proportional to the sum of the number of intermediate hops and the maximum physical distance between any two nodes. Moreover, % applying an asymmetric bandwidth assignment to the links, all-to-all communication can be realized $(1+o(1))$-optimally both with regard to bandwidth and delays. This is achieved at node degrees of $n^{\varepsilon}$, for an arbitrary small constant $\varepsilon\in (0,1]$. In contrast, these results are impossible in any network featuring constant or polylogarithmic node degrees. Through simulation, we assess the benefits of an implementation of the proposed communication strategy. Our results indicate that, for a million processors, CLEX can increase bandwidth utilization and reduce average routing path length by at least factors $10$ respectively $5$ in comparison to a torus network. Furthermore, the CLEX communication scheme features several other properties, such as deadlock-freedom, inherent fault-tolerance, and canonical partition into smaller subsystems. %K Computer Science, Distributed, Parallel, and Cluster Computing, cs.DC,Computer Science, Networking and Internet Architecture, cs.NI
[233]
C. Lenzen and M. Medina, “Efficient Metastability-Containing Gray Code 2-Sort,” in 22nd IEEE International Symposium on Asynchronous Circuits and Systems, Porto Alegre, Brazil, 2016.
Export
BibTeX
@inproceedings{LenzenASYNC2016, TITLE = {Efficient Metastability-Containing Gray Code 2-Sort}, AUTHOR = {Lenzen, Christoph and Medina, Moti}, LANGUAGE = {eng}, ISBN = {978-1-4673-9007-1}, DOI = {10.1109/ASYNC.2016.18}, PUBLISHER = {IEEE}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {22nd IEEE International Symposium on Asynchronous Circuits and Systems}, PAGES = {49--56}, ADDRESS = {Porto Alegre, Brazil}, }
Endnote
%0 Conference Proceedings %A Lenzen, Christoph %A Medina, Moti %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Efficient Metastability-Containing Gray Code 2-Sort : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-07EB-6 %R 10.1109/ASYNC.2016.18 %D 2016 %B 22nd IEEE International Symposium on Asynchronous Circuits and Systems %Z date of event: 2016-05-08 - 2016-05-11 %C Porto Alegre, Brazil %B 22nd IEEE International Symposium on Asynchronous Circuits and Systems %P 49 - 56 %I IEEE %@ 978-1-4673-9007-1
[234]
R. Levi, D. Ron, and R. Rubinfeld, “A Local Algorithm for Constructing Spanners in Minor-Free Graphs,” 2016. [Online]. Available: http://arxiv.org/abs/1604.07038. (arXiv: 1604.07038)
Abstract
Constructing a spanning tree of a graph is one of the most basic tasks in graph theory. We consider this problem in the setting of local algorithms: one wants to quickly determine whether a given edge $e$ is in a specific spanning tree, without computing the whole spanning tree, but rather by inspecting the local neighborhood of $e$. The challenge is to maintain consistency. That is, to answer queries about different edges according to the same spanning tree. Since it is known that this problem cannot be solved without essentially viewing all the graph, we consider the relaxed version of finding a spanning subgraph with $(1+\epsilon)n$ edges (where $n$ is the number of vertices and $\epsilon$ is a given sparsity parameter). It is known that this relaxed problem requires inspecting $\Omega(\sqrt{n})$ edges in general graphs, which motivates the study of natural restricted families of graphs. One such family is the family of graphs with an excluded minor. For this family there is an algorithm that achieves constant success probability, and inspects $(d/\epsilon)^{poly(h)\log(1/\epsilon)}$ edges (for each edge it is queried on), where $d$ is the maximum degree in the graph and $h$ is the size of the excluded minor. The distances between pairs of vertices in the spanning subgraph $G'$ are at most a factor of $poly(d, 1/\epsilon, h)$ larger than in $G$. In this work, we show that for an input graph that is $H$-minor free for any $H$ of size $h$, this task can be performed by inspecting only $poly(d, 1/\epsilon, h)$ edges. The distances between pairs of vertices in the spanning subgraph $G'$ are at most a factor of $\tilde{O}(h\log(d)/\epsilon)$ larger than in $G$. Furthermore, the error probability of the new algorithm is significantly improved to $\Theta(1/n)$. This algorithm can also be easily adapted to yield an efficient algorithm for the distributed setting.
Export
BibTeX
@online{DBLP:journals/corr/LeviRR16, TITLE = {A Local Algorithm for Constructing Spanners in Minor-Free Graphs}, AUTHOR = {Levi, Reut and Ron, Dana and Rubinfeld, Ronitt}, LANGUAGE = {eng}, URL = {http://arxiv.org/abs/1604.07038}, EPRINT = {1604.07038}, EPRINTTYPE = {arXiv}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, ABSTRACT = {Constructing a spanning tree of a graph is one of the most basic tasks in graph theory. We consider this problem in the setting of local algorithms: one wants to quickly determine whether a given edge $e$ is in a specific spanning tree, without computing the whole spanning tree, but rather by inspecting the local neighborhood of $e$. The challenge is to maintain consistency. That is, to answer queries about different edges according to the same spanning tree. Since it is known that this problem cannot be solved without essentially viewing all the graph, we consider the relaxed version of finding a spanning subgraph with $(1+\epsilon)n$ edges (where $n$ is the number of vertices and $\epsilon$ is a given sparsity parameter). It is known that this relaxed problem requires inspecting $\Omega(\sqrt{n})$ edges in general graphs, which motivates the study of natural restricted families of graphs. One such family is the family of graphs with an excluded minor. For this family there is an algorithm that achieves constant success probability, and inspects $(d/\epsilon)^{poly(h)\log(1/\epsilon)}$ edges (for each edge it is queried on), where $d$ is the maximum degree in the graph and $h$ is the size of the excluded minor. The distances between pairs of vertices in the spanning subgraph $G'$ are at most a factor of $poly(d, 1/\epsilon, h)$ larger than in $G$. In this work, we show that for an input graph that is $H$-minor free for any $H$ of size $h$, this task can be performed by inspecting only $poly(d, 1/\epsilon, h)$ edges. The distances between pairs of vertices in the spanning subgraph $G'$ are at most a factor of $\tilde{O}(h\log(d)/\epsilon)$ larger than in $G$. Furthermore, the error probability of the new algorithm is significantly improved to $\Theta(1/n)$. This algorithm can also be easily adapted to yield an efficient algorithm for the distributed setting.}, }
Endnote
%0 Report %A Levi, Reut %A Ron, Dana %A Rubinfeld, Ronitt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T A Local Algorithm for Constructing Spanners in Minor-Free Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5FFC-C %U http://arxiv.org/abs/1604.07038 %D 2016 %X Constructing a spanning tree of a graph is one of the most basic tasks in graph theory. We consider this problem in the setting of local algorithms: one wants to quickly determine whether a given edge $e$ is in a specific spanning tree, without computing the whole spanning tree, but rather by inspecting the local neighborhood of $e$. The challenge is to maintain consistency. That is, to answer queries about different edges according to the same spanning tree. Since it is known that this problem cannot be solved without essentially viewing all the graph, we consider the relaxed version of finding a spanning subgraph with $(1+\epsilon)n$ edges (where $n$ is the number of vertices and $\epsilon$ is a given sparsity parameter). It is known that this relaxed problem requires inspecting $\Omega(\sqrt{n})$ edges in general graphs, which motivates the study of natural restricted families of graphs. One such family is the family of graphs with an excluded minor. For this family there is an algorithm that achieves constant success probability, and inspects $(d/\epsilon)^{poly(h)\log(1/\epsilon)}$ edges (for each edge it is queried on), where $d$ is the maximum degree in the graph and $h$ is the size of the excluded minor. The distances between pairs of vertices in the spanning subgraph $G'$ are at most a factor of $poly(d, 1/\epsilon, h)$ larger than in $G$. In this work, we show that for an input graph that is $H$-minor free for any $H$ of size $h$, this task can be performed by inspecting only $poly(d, 1/\epsilon, h)$ edges. The distances between pairs of vertices in the spanning subgraph $G'$ are at most a factor of $\tilde{O}(h\log(d)/\epsilon)$ larger than in $G$. Furthermore, the error probability of the new algorithm is significantly improved to $\Theta(1/n)$. This algorithm can also be easily adapted to yield an efficient algorithm for the distributed setting. %K Computer Science, Data Structures and Algorithms, cs.DS
[235]
R. Levi, D. Ron, and R. Rubinfeld, “A Local Algorithm for Constructing Spanners in Minor-Free Graphs,” in Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2016), Paris, France, 2016.
Export
BibTeX
@inproceedings{DBLP:conf/approx/LeviRR16, TITLE = {A Local Algorithm for Constructing Spanners in Minor-Free Graphs}, AUTHOR = {Levi, Reut and Ron, Dana and Rubinfeld, Ronitt}, LANGUAGE = {eng}, ISBN = {978-3-95977-018-7}, URL = {urn:nbn:de:0030-drops-66613}, DOI = {10.4230/LIPIcs.APPROX-RANDOM.2016.38}, PUBLISHER = {Schloss Dagstuhl}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, BOOKTITLE = {Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2016)}, EDITOR = {Jansen, Klaus and Mathieu, Claire and Rolim, Jos{\'e} D. P. and Umans, Chris}, PAGES = {1--15}, EID = {38}, SERIES = {Leibniz International Proceedings in Informatics}, VOLUME = {60}, ADDRESS = {Paris, France}, }
Endnote
%0 Conference Proceedings %A Levi, Reut %A Ron, Dana %A Rubinfeld, Ronitt %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations %T A Local Algorithm for Constructing Spanners in Minor-Free Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-5FE0-8 %R 10.4230/LIPIcs.APPROX-RANDOM.2016.38 %U urn:nbn:de:0030-drops-66613 %D 2016 %B 19th International Workshop on Approximation Algorithms for Combinatorial Optimization Problems / 20th International Workshop on Randomization and Computation %Z date of event: 2016-09-07 - 2016-09-09 %C Paris, France %B Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques %E Jansen, Klaus; Mathieu, Claire; Rolim, Jos&#233; D. P.; Umans, Chris %P 1 - 15 %Z sequence number: 38 %I Schloss Dagstuhl %@ 978-3-95977-018-7 %B Leibniz International Proceedings in Informatics %N 60 %U http://drops.dagstuhl.de/doku/urheberrecht1.htmlhttp://drops.dagstuhl.de/opus/volltexte/2016/6661/
[236]
D. Lokshtanov, M. Pilipczuk, and E. J. van Leeuwen, “Independence and Efficient Domination on P6-free Graphs,” in Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2016), Arlington, VA, USA, 2016.
Export
BibTeX
@inproceedings{LokshtanovSODA2015, TITLE = {Independence and Efficient Domination on {$P_6$}-free Graphs}, AUTHOR = {Lokshtanov, Daniel and Pilipczuk, Marcin and van Leeuwen, Erik Jan}, LANGUAGE = {eng}, ISBN = {978-1-61197-433-1}, DOI = {10.1137/1.9781611974331.ch124}, PUBLISHER = {SIAM}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, BOOKTITLE = {Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms (SODA 2016)}, EDITOR = {Krauthgamer, Robert}, PAGES = {1784--1803}, ADDRESS = {Arlington, VA, USA}, }
Endnote
%0 Conference Proceedings %A Lokshtanov, Daniel %A Pilipczuk, Marcin %A van Leeuwen, Erik Jan %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Independence and Efficient Domination on P6-free Graphs : %G eng %U http://hdl.handle.net/11858/00-001M-0000-0029-2497-D %R 10.1137/1.9781611974331.ch124 %D 2016 %B Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms %Z date of event: 2016-01-10 - 2016-01-12 %C Arlington, VA, USA %B Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms %E Krauthgamer, Robert %P 1784 - 1803 %I SIAM %@ 978-1-61197-433-1
[237]
M. Mahajan, R. B. V. Rao, and K. Sreenivasaiah, “Building Above Read-Once Polynomials: Identity Testing and Hardness of Representation,” Algorithmica, vol. 76, no. 4, 2016.
Export
BibTeX
@article{Mahajan2016, TITLE = {Building Above Read-Once Polynomials: Identity Testing and Hardness of Representation}, AUTHOR = {Mahajan, Meena and Rao, Raghavendra B. V. and Sreenivasaiah, Karteek}, LANGUAGE = {eng}, ISSN = {0178-4617}, DOI = {10.1007/s00453-015-0101-z}, PUBLISHER = {Springer-Verlag}, ADDRESS = {New York}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {Algorithmica}, VOLUME = {76}, NUMBER = {4}, PAGES = {890--909}, }
Endnote
%0 Journal Article %A Mahajan, Meena %A Rao, Raghavendra B. V. %A Sreenivasaiah, Karteek %+ External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T Building Above Read-Once Polynomials: Identity Testing and Hardness of Representation : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002C-1017-F %R 10.1007/s00453-015-0101-z %7 2016 %D 2016 %J Algorithmica %V 76 %N 4 %& 890 %P 890 - 909 %I Springer-Verlag %C New York %@ false
[238]
N. Megow, M. Skutella, J. Verschae, and A. Wiese, “The Power of Recourse for Online MST and TSP,” SIAM Journal on Computing, vol. 45, no. 3, 2016.
Export
BibTeX
@article{Megow2016, TITLE = {The Power of Recourse for Online {MST} and {TSP}}, AUTHOR = {Megow, Nicole and Skutella, Martin and Verschae, Jos{\'e} and Wiese, Andreas}, LANGUAGE = {eng}, ISSN = {0097-5397}, DOI = {10.1137/130917703}, PUBLISHER = {Society for Industrial and Applied Mathematics.}, ADDRESS = {Philadelphia, PA}, YEAR = {2016}, MARGINALMARK = {$\bullet$}, DATE = {2016}, JOURNAL = {SIAM Journal on Computing}, VOLUME = {45}, NUMBER = {3}, PAGES = {859--880}, }
Endnote
%0 Journal Article %A Megow, Nicole %A Skutella, Martin %A Verschae, Jos&#233; %A Wiese, Andreas %+ Algorithms and Complexity, MPI for Informatics, Max Planck Society External Organizations External Organizations Algorithms and Complexity, MPI for Informatics, Max Planck Society %T The Power of Recourse for Online MST and TSP : %G eng %U http://hdl.handle.net/11858/00-001M-0000-002B-B051-4 %R 10.1137/130917703 %7 2016 %D 2016 %J SIAM Journal on Computing %V 45 %N 3 %& 859 %P 859 - 880 %I Society for Industrial and Applied Mathematics. %C Philadelphia, PA %@ false
[239]
K. Mehlhorn and S. Saxena, “A Still Simpler Way of Introducing the Interior-Point Method for Linear Programming,” Computer Science Review, vol. 22, 2016.