Abstract
Data is compressed by reducing its redundancy, but this also makes the data less reliable, more prone to errors. Increasing the integrity of data, on the other hand, is done by adding check bits and parity bits, a process that increases the size of the data, thereby increasing redundancy. Data compression and data reliability are therefore opposites, and it is interesting to note that the latter is a relatively recent field, whereas the former existed even before the advent of computers. The sympathetic telegraph, discussed in the Preface, the Braille code of 1820 (Section 1.1.1), the shutter telegraph of the British admiralty [Bell et al. 90], and the Morse code of 1838 (Table 2.1) use simple, intuitive forms of compression. It therefore seems that reducing redundancy comes naturally to anyone who works on codes, but increasing it is something that “goes against the grain” in humans. This section discusses simple, intuitive compression methods that have been used in the past. Today these methods are mostly of historical interest, since they are generally inefficient and cannot compete with the modern compression methods developed during the last several decades.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2010 Springer-Verlag London
About this chapter
Cite this chapter
Salomon, D., Motta, G. (2010). Basic Techniques. In: Handbook of Data Compression. Springer, London. https://doi.org/10.1007/978-1-84882-903-9_1
Download citation
DOI: https://doi.org/10.1007/978-1-84882-903-9_1
Published:
Publisher Name: Springer, London
Print ISBN: 978-1-84882-902-2
Online ISBN: 978-1-84882-903-9
eBook Packages: Computer ScienceComputer Science (R0)