Skip to main content Accessibility help
×
Hostname: page-component-8448b6f56d-sxzjt Total loading time: 0 Render date: 2024-04-24T08:02:13.467Z Has data issue: false hasContentIssue false

4 - More Methods for Solving Linear Equations

Published online by Cambridge University Press:  01 June 2011

John F. Monahan
Affiliation:
North Carolina State University
Get access

Summary

Introduction

The previous chapter dwelled on the fundamental methods of matrix computations. In this chapter, more specialized methods are considered. The first topic is an alternative approach to solving general systems of equations – full elimination (often the method taught in beginning linear algebra courses), which has some advantages whenever the inverse is required. Next, our goal is reducing the effort in solving equations by exploiting the structure of a matrix. One such structure is bandedness, and the Cholesky factorization of a banded positive definite matrix is then applied to time-series computations, cutting the work from O(n3) to O(n). Next is the Toeplitz structure, also arising in time-series analysis, where the work can be reduced to O(n2) in a more general setting. Sparse matrix methods are designed to exploit unstructured patterns of zeros and so avoid unneeded work. Finally, iterative methods are discussed, beginning with iterative improvement.

Full Elimination with Complete Pivoting

Gaussian elimination creates an upper triangular matrix, column by column, by adding multiples of a row to the rows below it and placing zeros below the diagonal of each column. An alternative is to place zeros throughout that column – with the exception of the pivot position, which could be made equal to one.

Type
Chapter
Information
Publisher: Cambridge University Press
Print publication year: 2011

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

Ansley, Craig F. (1979), “An Algorithm for the Exact Likelihood of a Mixed Autoregressive-Moving Average Process,” Biometrika 66: 59–65.CrossRefGoogle Scholar
Bartels, Richard H. and Golub, Gene H. (1969), “The Simplex Method of Linear Programming Using LU Decomposition,” Communications of the ACM, 12: 266–68.CrossRefGoogle Scholar
Box, G. E. P. and Jenkins, G. M. (1972), Time Series Analysis: Forecasting and Control. San Francisco: Holden-Day.Google Scholar
Cybenko, George (1980), “The Numerical Stability of the Levinson–Durbin Algorithm for Toeplitz Systems of Equations,” SIAM Journal of Scientific and Statistical Computing 1: 303–19.CrossRefGoogle Scholar
Durbin, J. (1960), “The Fitting of Time Series Models,” Revue Internationale Institut de Statistique 28: 233–43.CrossRefGoogle Scholar
Golub, Gene H. and Loan, Charles van (1996), Matrix Computations, 3rd ed. Baltimore: Johns Hopkins University Press.Google Scholar
Harrod, W. J. and Plemmons, R. J. (1984), “Comparison of Some Direct Methods for Computing Stationary Distributions of Markov Chains,” SIAM Journal of Scientific and Statistical Computing 5: 453–69.CrossRefGoogle Scholar
Haslett, John and Raftery, Adrian (1989), “Space-Time Modelling with Long-Memory Dependence: Assessing Ireland's Wind Power Resource,” Applied Statistics 38: 1–50.CrossRefGoogle Scholar
Hosking, J. R. M. (1981), “Fractional Differencing,” Biometrika 68: 165–76.CrossRefGoogle Scholar
Jennings, Alan (1977), Matrix Computation for Engineers and Scientists. New York: Wiley.Google Scholar
Kelley, C. T. (1995), Iterative Methods for Linear and Nonlinear Equations. Philadelphia: SIAM.CrossRefGoogle Scholar
Levinson, Norman (1947), “The Weiner RMS Error Criterion in Filter Design and Prediction,” Journal of Mathematics and Physics 26: 261–78.Google Scholar
Martin, R. S. and Wilkinson, J. H. (1965), “Symmetric Decomposition of Positive Definite Band Matrices,” Numerische Mathematik 7: 355–61.CrossRefGoogle Scholar
Monahan, John F. (1984), “A Note on Enforcing Stationarity in Autoregressive-Moving Average Models,” Biometrika 71: 403–4.CrossRefGoogle Scholar
Murtagh, Bruce A. (1981), Advanced Linear ProgrammingNew York: McGraw-Hill.Google Scholar
Pagano, Marcello (1973), “When is an Autoregressive Scheme Stationary?” Communications in Statistics 1: 533–44.CrossRefGoogle Scholar
Trench, William F. (1964), “An Algorithm for the Inversion of Finite Toeplitz Matrices,” Journal of SIAM 12: 515–22.Google Scholar

Save book to Kindle

To save this book to your Kindle, first ensure coreplatform@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Available formats
×

Save book to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

Available formats
×

Save book to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

Available formats
×