Abstract
Data is compressed by reducing its redundancy, but this also makes the data less reliable, more prone to errors. Increasing the integrity of data, on the other hand, is done by adding check bits and parity bits, a process that increases the size of the data, thereby increasing redundancy. Data compression and data reliability are therefore opposites, and it is interesting to note that the latter is a relatively recent field, whereas the former existed even before the advent of computers. The sympathetic telegraph, discussed in the Preface, the Braille code of 1820 (Section 1.1.1), and the Morse code of 1838 (Table 2.1) use simple, intuitive forms of compression. It therefore seems that reducing redundancy comes naturally to anyone who works on codes, but increasing it is something that “goes against the grain” in humans. This section discusses simple, intuitive compression methods that have been used in the past. Today these methods are mostly of historical interest, since they are generally inefficient and cannot compete with the modern compression methods developed during the last several decades.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Rights and permissions
Copyright information
© 2007 Springer-Verlag London Limited
About this chapter
Cite this chapter
(2007). Basic Techniques. In: Data Compression. Springer, London. https://doi.org/10.1007/978-1-84628-603-2_2
Download citation
DOI: https://doi.org/10.1007/978-1-84628-603-2_2
Publisher Name: Springer, London
Print ISBN: 978-1-84628-602-5
Online ISBN: 978-1-84628-603-2
eBook Packages: Computer ScienceComputer Science (R0)