[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
article
Free access

The influence of relaxed supernode partitions on the multifrontal method

Published: 01 December 1989 Publication History

Abstract

In this paper we present an algorithm for partitioning the nodes of a graph into supernodes, which improves the performance of the multifrontal method for the factorization of large, sparse matrices on vector computers. This new algorithm first partitions the graph into fundamental supernodes. Next, using a specified relaxation parameter, the supernodes are coalesced in a careful manner to create a coarser supernode partition. Using this coarser partition in the factorization generally introduces logically zero entries into the factor. This is accompanied by a decrease in the amount of sparse vector computations and data movement and an increase in the number of dense vector operations. The amount of storage required for the factor is generally increased by a small amount. On a collection of moderately sized 3-D structures, matrices speedups of 3 to 20 percent on the Cray X-MP are observed over the fundamental supernode partition which allows no logically zero entries in the factor. Using this relaxed supernode partition, the multifrontal method now factorizes the extremely sparse electric power matrices faster than the general sparse algorithm. In addition, there is potential for considerably reducing the communication requirements for an implementation of the multifrontal method on a local memory multiprocessor.

References

[1]
ASHCRAFT, C.C. A vector implementation of the multifrontal method for large sparse symmetric positive definite linear systems. Boeing Computer Services, Tech. Rep. ETA-TR-51, Seattle, Wash., 1987.
[2]
ASHCRAFT, C. C., LEWIS, J. G., AND PEYTON, B.W. A supernodal implementation of general sparse factorizations for vector computers. Boeing Computer Services, Tech. Rep. ETA-TR-52, Seattle, Wash., 1987.
[3]
ASHCRAFT, C. C., GRIMES, R. G., LEWIS, J. G., PEYTON, B. W., AND SIMON, H.D. Progress in sparse matrix methods for large linear systems on vector supercomputers. Int. J. Supercomput. Appl. 1 (1987), 10-20.
[4]
DODSON, D. S., AND LEWIS, J. G. Proposed sparse extensions to the basic linear algebra subprograms. SIGNUM Newsl. (ACM) 20 (1985), 22-25.
[5]
DODSON, D. S., GRIMES, R. G., AND LEWIS, J.G. Sparse extensions to the FORTRAN basic linear algebra subprograms. Boeing Computer Services, Tech. Rep. ETA-TR-63, Seattle, Wash., accepted for publication in ACM TOMS (1988).
[6]
DONGARRA, J. J., AND EISENSTAT, S. C. Squeezing the most out of an algorithm in CRAY Fortran. ACM Trans. Math. Softw. 10, 3 (Sept. 1984), 221-230.
[7]
DONGARRA, J. J., DU CROZ, J., HAMMARLING, S. Z., AND HANSON, R.J. An extended set of Fortran basic linear algebra subprograms. Argonne National Laboratory, Tech. Memo. 41, Argonne, Ill., 1986.
[8]
DUFF, I.S. Full matrix techniques in sparse Gaussian elimination. In Lecture Notes in Math., 912, G. A. Watson, Ed. Springer-Verlag, New York, 1982, pp. 71-84.
[9]
DUFF, I. S. Parallel implementation of multifrontal schemes. Parallel Comput. 3 (1986), 193-204.
[10]
DUFF, I. S., AND REID, J.K. The multifrontal solution of indefinite sparse symmetric linear equations. ACM Trans. Math. Softw. 9, 3 (Sept. 1983), 302-325.
[11]
DUFF, I. S., AND REID, J.K. The multifrontal solution of unsymmetric sets of linear equations. SIAM J. Sci. Star. Comput. 5 (1984), 633-641.
[12]
DUFF, I. S., GRIMES, R. G., AND LEWIS, J.G. Sparse matrix test problems. Harwell Laboratory, Tech. Rep. CSS-191, Oxfordshire, England, to appear in ACM TOMS (1988).
[13]
DUFF, I. S., GRIMES, R. G., LEWIS, J. G., AND POOLE, W.G. Sparse matrix test problems. SIGNUM Newsl. (1982), 22.
[14]
EISENSTAT, S. C., SCHULTZ, M. H., AND SHERMAN, A. A. Software for sparse gaussian elimination with limited core storage. In Sparse Matrix Proceedings. I. S. Duff and G. W. Stewart, Eds., SIAM, Philadelphia, Pa., 1978, pp. 135-153.
[15]
EISENSTAT, S. C., SCHULTZ, M. H., AND SHERMAN, A.A. Algorithms and data structures for sparse symmetric Gaussian elimination. SIAM J. Sci. Star. Comput. 2 (1981), 225-237.
[16]
GEORGE, J. A., AND LIU, J.W. Computer Solution of Large Sparse Positive Definite Systems. Prentice-Hall, Englewood Cliffs, N.J., 1981.
[17]
LEwis, J. G., AND SIMON, H.D. The impact of hardware gather/scatter on sparse Gaussian elimination. SIAM J. Sci. Star. Comput. 9 (1988), 304-311.
[18]
LIU, J.W. Modification of the minimum-degree algorithm by multiple elimination. ACM Trans. Math. Softw. 11, 2 (June 1985), 141-153.
[19]
LIU, J.W. A compact row storage scheme for Cholesky factors using elimination trees. ACM Trans. Math. Softw. 12, 2 (June 1986), 127-148.
[20]
LIU, J. W. On the storage requirement in the out-of-core multifrontal method for sparse factorization. ACM Trans. Math. Softw. 12 (1986), 249-264.
[21]
SCHREIBER, R. A new implementation of sparse Gaussian elimination. ACM Trans. Math. Softw. 8, 2 (Sept. 1982), 256-276.

Cited By

View all
  • (2024)HPS Cholesky: Hierarchical Parallelized Supernodal Cholesky with Adaptive ParametersACM Transactions on Parallel Computing10.1145/363005111:1(1-22)Online publication date: 11-Mar-2024
  • (2024)GPU Accelerated Sparse Cholesky FactorizationSC24-W: Workshops of the International Conference for High Performance Computing, Networking, Storage and Analysis10.1109/SCW63240.2024.00098(703-707)Online publication date: 17-Nov-2024
  • (2022)A Left-Looking Sparse Cholesky Parallel Algorithm for Shared Memory MultiprocessorsAdvances in Natural Computation, Fuzzy Systems and Knowledge Discovery10.1007/978-3-030-89698-0_21(198-208)Online publication date: 4-Jan-2022
  • Show More Cited By

Recommendations

Reviews

Michael Donoho Fry

The authors show how to solve large, sparse, positive definite systems of linear equations on a Cray X-MP a little faster without using a lot more memory. With its intimidating title, this paper is unlikely to draw a large audience outside of specialists in this research, but others can benefit from reading it. The paper begins with some of the history of two rival algorithms, the “multifrontal method” and the “general sparse method.” The multifrontal method had disadvantages that have since been overcome, so that it now runs two to three times as fast as the general sparse method. The improvements that overcame these disadvantages were presented in an earlier paper by Ashcraft [1]. Matrix factorization takes place in steps that trace a postorder traversal of an “elimination tree.” Expensive overhead can be reduced by collecting the tree's nodes into “supernodes.” Why is this effort to gain a few more percentage points of performance worth investigation__?__ Supercomputers spend a lot of their productive time solving this type of system of equations, but that only explains specialists' interest. The authors employ a simple idea. Back off from shrinking the problem. Make the problem a little bigger, requiring a little more memory and a few million more arithmetic operations, but arrange it so that more of the arithmetic can use the vector architecture of the machine and run at 100–125 megaflops rather than at the 8–<__?__Pub Caret>10 megaflops of scalar arithmetic. The authors do not pretend to have invented this idea themselves; they simply record the thinking that went into one use of it. The partitioning into supernodes permits more vector arithmetic. The authors give statistics on test runs so the reader can judge the success of the effort. They also invite the reader to explore other avenues, including other criteria for partitioning nodes and alternative multiprocessor implementations. The paper and its bibliography offer a well-motivated entry point for readers interested in learning about vector architecture and how matrix algorithms are developed for it. It is not easy reading, but it is accessible to a larger audience than its title would suggest.

Access critical reviews of Computing literature here

Become a reviewer for Computing Reviews.

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Mathematical Software
ACM Transactions on Mathematical Software  Volume 15, Issue 4
Dec. 1989
107 pages
ISSN:0098-3500
EISSN:1557-7295
DOI:10.1145/76909
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 December 1989
Published in TOMS Volume 15, Issue 4

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)82
  • Downloads (Last 6 weeks)13
Reflects downloads up to 12 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)HPS Cholesky: Hierarchical Parallelized Supernodal Cholesky with Adaptive ParametersACM Transactions on Parallel Computing10.1145/363005111:1(1-22)Online publication date: 11-Mar-2024
  • (2024)GPU Accelerated Sparse Cholesky FactorizationSC24-W: Workshops of the International Conference for High Performance Computing, Networking, Storage and Analysis10.1109/SCW63240.2024.00098(703-707)Online publication date: 17-Nov-2024
  • (2022)A Left-Looking Sparse Cholesky Parallel Algorithm for Shared Memory MultiprocessorsAdvances in Natural Computation, Fuzzy Systems and Knowledge Discovery10.1007/978-3-030-89698-0_21(198-208)Online publication date: 4-Jan-2022
  • (2021)STM-multifrontal QRProceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis10.1145/3458817.3476199(1-14)Online publication date: 14-Nov-2021
  • (2021)Fast Implementation of the Traveling-Salesman-Problem Method for Reordering Columns within SupernodesSIAM Journal on Matrix Analysis and Applications10.1137/20M136807042:3(1337-1364)Online publication date: 7-Sep-2021
  • (2020)k-center Clustering under Perturbation ResilienceACM Transactions on Algorithms10.1145/338142416:2(1-39)Online publication date: 9-Mar-2020
  • (2020)Testing Bounded ArboricityACM Transactions on Algorithms10.1145/338141816:2(1-22)Online publication date: 9-Mar-2020
  • (2020)High-performance sampling of generic determinantal point processesPhilosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences10.1098/rsta.2019.0059378:2166(20190059)Online publication date: 20-Jan-2020
  • (2020)Cached Gaussian elimination for simulating Stokes flow on domains with repetitive geometryJournal of Computational Physics10.1016/j.jcp.2020.109812423(109812)Online publication date: Dec-2020
  • (2020)Affecting the Relaxation Parameter in the Multifrontal MethodSustained Simulation Performance 2018 and 201910.1007/978-3-030-39181-2_17(215-224)Online publication date: 31-Mar-2020
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Full Access

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media