mplrs: A scalable parallel vertex/facet enumeration code
November 20, 2015 ยท Declared Dead ยท ๐ Mathematical Programming Computation
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
David Avis, Charles Jordan
arXiv ID
1511.06487
Category
cs.MS: Mathematical Software
Cross-listed
cs.CG,
cs.DC
Citations
32
Venue
Mathematical Programming Computation
Last Checked
2 months ago
Abstract
We describe a new parallel implementation, mplrs, of the vertex enumeration code lrs that uses the MPI parallel environment and can be run on a network of computers. The implementation makes use of a C wrapper that essentially uses the existing lrs code with only minor modifications. mplrs was derived from the earlier parallel implementation plrs, written by G. Roumanis in C++. plrs uses the Boost library and runs on a shared memory machine. In developing mplrs we discovered a method of balancing the parallel tree search, called budgeting, that greatly improves parallelization beyond the bottleneck encountered previously at around 32 cores. This method can be readily adapted for use in other reverse search enumeration codes. We also report some preliminary computational results comparing parallel and sequential codes for vertex/facet enumeration problems for convex polyhedra. The problems chosen span the range from simple to highly degenerate polytopes. For most problems tested, the results clearly show the advantage of using the parallel implementation mplrs of the reverse search based code lrs, even when as few as 8 cores are available. For some problems almost linear speedup was observed up to 1200 cores, the largest number of cores tested.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Mathematical Software
๐
๐
Old Age
๐
๐
Old Age
CSR5: An Efficient Storage Format for Cross-Platform Sparse Matrix-Vector Multiplication
R.I.P.
๐ป
Ghosted
Mathematical Foundations of the GraphBLAS
R.I.P.
๐ป
Ghosted
The DUNE Framework: Basic Concepts and Recent Developments
R.I.P.
๐ป
Ghosted
Format Abstraction for Sparse Tensor Algebra Compilers
R.I.P.
๐ป
Ghosted
AMReX: Block-Structured Adaptive Mesh Refinement for Multiphysics Applications
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted