Abstract
In general, one tries to approximate a tensor v by another tensor u requiring less data. The reason is twofold: the memory size should decrease and, hopefully, operations involving u should require less computational work. In fact, \({\rm u} \in \mathcal{R}_{{\rm r}}\) leads to decreasing cost for storage and operations as r decreases. However, the other side of the coin is an increasing approximation error. Correspondingly, in Sect. 9.1 two approximation strategies are presented, where either the representation rank r of u or the accuracy is prescribed. Before we study the approximation problem in general, two particular situations are discussed. Section 9.2 is devoted to r = 1, when \({\rm u} \in \mathcal{R}_{1}\) is an elementary tensor. The matrix case d = 2 is recalled in Sect. 9.3. The properties observed in the latter two sections contrast with the true tensor case studied in Sect. 9.4. Numerical algorithms solving the approximation problem will be discussed in Sect. 9.5. Modified approximation problems are addressed in Sect. 9.6.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag GmbH Berlin Heidelberg
About this chapter
Cite this chapter
Hackbusch, W. (2012). r-Term Approximation. In: Tensor Spaces and Numerical Tensor Calculus. Springer Series in Computational Mathematics, vol 42. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-28027-6_9
Download citation
DOI: https://doi.org/10.1007/978-3-642-28027-6_9
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-28026-9
Online ISBN: 978-3-642-28027-6
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)