Skip to main content

Image and Video Acquisition, Representation and Storage

  • Chapter
  • First Online:
Machine Learning for Audio, Image and Video Analysis

Abstract

What the reader should know to understand this chapter \(\bullet \) Elementary notions of optics and physics. \(\bullet \) Basic notions of mathematics.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 64.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 99.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The aperture controls the amount of light that reaches the camera sensor.

  2. 2.

    In a camera, the shutter is a device that allows light to pass for a determined period of time, with the aim of exposing the CCD (or CMOS) sensor to the required amount of light to create a permanent image of view. Shutter speed is the time that the shutter is open.

  3. 3.

    Brightness measures the color intensity (see Sect. 3.4.2).

  4. 4.

    This is also called the International Lighting Committee.

  5. 5.

    TIFF also provides lossy compression schemes, although they are less popular.

  6. 6.

    In this mode, JPEG produces a nolossy compression.

  7. 7.

    JPEG offers the possibility of reducing by a factor of 2 only in the horizontal direction.

  8. 8.

    CRT stands for cathode-ray tube.

  9. 9.

    Sequential Color with Memory.

References

  1. T. Acharaya and A. K. Ray. Image Processing: Principles and Applications. John Wiley and Sons, 2005.

    Google Scholar 

  2. F.L. Alt. Digital pattern recognition by moments. Journal of ACM, 11:240–258, 1962.

    Google Scholar 

  3. D. Ballard and C. Brown. Computer Vision. Academic Press, 1982.

    Google Scholar 

  4. B. E. Bayer. Color imaging array. Color us patent 3,971,065. Technical report, Eastman Kodak Company, 1976.

    Google Scholar 

  5. K. M. Bhurchandi, A. K. Ray, and P. M. Nawghare. An analytical approach for sampling the rgb color space considering physiological limitations of human vision and its application for color image analysis. In Proceedings of Indian Conference on Computer Vision, Graphics and Image Processing, pages 44–49, 2000.

    Google Scholar 

  6. J. Bormans, J. Gelissen, and A. Perkis. Mpeg-21: The 21\(^{st}\) century multimedia framework. IEEE Signal Processing Magazine, pages 53–62, 2003.

    Google Scholar 

  7. G. Buchsbaum. An analytical derivation of visual nonlinearity. IEEE Transactions on biomedical engineering, BME-27(5):237–242, 1980.

    Google Scholar 

  8. C. K. Chui. An Introduction to Wavelets. Academic Press, 1982.

    Google Scholar 

  9. T. H. Cormen, C. E. Leiserson, and R. L. Rivest. Introduction to Algorithms. MIT Press, 1990.

    Google Scholar 

  10. I. Daubechies. Ten Lectures on Wavelets. SIAM, 1992.

    Google Scholar 

  11. A. Del Bimbo. Visual Information Retrieval. Morgan Kaufman Publishers, 1999.

    Google Scholar 

  12. T. Ebrahimi. Mpeg-4 video verification model: A video encoding/decoding algorithm based on content representation. Image Communication Journal, 9(4):367–384, 1996.

    Google Scholar 

  13. K. S. Gibson and D. Nickerson. Analysis of the Munsell colour system based on measurements made in 1919 and 1926. Journal of Optical Society of America, 3(12):591–608, 1940.

    Google Scholar 

  14. R. C. Gonzalez and R. E. Woods. Digital Image Processing. Addison Wesley, 1992.

    Google Scholar 

  15. G. Healey and Q. Luong. Color in computer vision: Recent progress. In Handbook of Pattern Recognition and Computer Vision, pages 283–312. World Scientific Publishing, 1998.

    Google Scholar 

  16. M. K. Hu. Visual pattern recognition by moment invariants. IRE Transactions on Information Theory, 8:351–364, 1962.

    Google Scholar 

  17. D. A. Huffman. A method for the construction of minimum-redundancy codes. Proceedings of the IRE, 40(9):1098–1101, 1952.

    Google Scholar 

  18. L. M. Hurvich and D. Jameson. An opponent process theory of colour vision. Psychological Review, 64(6):384–404, 1957.

    Google Scholar 

  19. L. M. Hurvich and D. Jameson. Some quantitative aspects of an opponent-colors theory: IV A psychological color specification system. Journal of the Optical Society of America, 45(6):416–421, 1957.

    Google Scholar 

  20. A. K. Jain. Fundamentals of Digital Image Processing. Prentice-Hall, 1989.

    Google Scholar 

  21. D. B. Judd and G. Wyszecki. Color in Business, Science and Industry. John Wiley and Sons, 1975.

    Google Scholar 

  22. H. R. Kang. Color Technology for Electronic Imaging Devices. SPIE Optical Engineering Press, 1997.

    Google Scholar 

  23. R. Koenen, F. Pereira, and L. Chiariglione. Mpeg-4: Context and objectives. Image Communication Journal, 9(4):295–304, 1997.

    Google Scholar 

  24. E. H. Land. Color vision and the natural images. Proceedings of the National Academy of Sciences, 45(1):116–129, 1959.

    Google Scholar 

  25. D. Le Gall. Mpeg: a video compression standard for multimedia applications. Communications of the ACM, 34(4):46–58, 1991.

    Google Scholar 

  26. D. G. Lowe. Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60(2):91–110, 2004.

    Google Scholar 

  27. S. Mallat. A theory for multiresolution signal decomposition: the wavelet representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 11(7):674–693, 1998.

    Google Scholar 

  28. G. W. Meyer. Tutorial on colour science. The Visual Computer, 2(5):278–290, 1986.

    Google Scholar 

  29. M. Miyahara and Y. Yoshida. Mathematical transform of rgb colour data to munsell colour system. In SPIE Visual Communication and Image Processing ’88, pages 650–657, 1988.

    Google Scholar 

  30. A. H. Munsell. An Atlas of the Munsell System. Wassworth-Howland, 1915.

    Google Scholar 

  31. C. L. Novak and S. A. Shafer. Color Vision. Encyclopedia of Artificial Intelligence. John Wiley and Sons, 1992.

    Google Scholar 

  32. W. B. Pennebaker and J. L. Mitchell. JPEG Still Image Data Compression Standard. Chapman & Hall, 1993.

    Google Scholar 

  33. W. K. Pratt. Digital Image Processing. John Wiley and Sons, 1991.

    Google Scholar 

  34. K. R. Rao and P. Yip. Digital Cosine Transform: Algorithms, Advantages, Applications. Academic Press, 1990.

    Google Scholar 

  35. T. Sakamoto, C. Nakanishi, and T. Hase. Software pixel interpolation for digital still cameras suitable for a 32-bit mcu. IEEE Transactions on Consumer Electronics, 44(4):1342–1352, 1998.

    Google Scholar 

  36. P. Salembier and J. R. Smith. Mpeg-7 multimedia description schemes. IEEE Transactions on Circuits and Systems for Video Technology, 11(6):748–759, 2001.

    Google Scholar 

  37. G. Sharma. Digital color imaging. IEEE Transactions on Image Processing, 6(7):901–932, 1997.

    Google Scholar 

  38. A. S. Tanenbaum. Modern Operating Systems. Prentice-Hall, 2001.

    Google Scholar 

  39. E. Trucco and A. Verri. Introductory Techniques for 3-D Computer Vision. Prentice-Hall, 1998.

    Google Scholar 

  40. P. Tsai, T. Acharaya, and A. K. Ray. Adaptive fuzzy color interpolation. Journal of Electronic Imaging, 11(3):293–305, 2002.

    Google Scholar 

  41. B. A. Wandell. Foundations of Vision. Sinauer Associates, 1995.

    Google Scholar 

  42. G. Wyszecki and W. S. Stiles. Color Science. Mc Graw-Hill, 1982.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Francesco Camastra .

Problems

Problems

3.1

Show that in the XYZ model the white is represented by the triple (1,1,1).

3.2

Consider the YIQ model. Show that in a grayscale image, where R\(=\)G\(=\)B, the chrominance components I and Q are null.

3.3

Consider the HSV model. Show that in the simplest form of HSV transformation, the hue (H) become undefined when the saturation S is null.

3.4

Compute in HSV model, the coordinates of cyan, magenta and yellow.

3.5

Repeat Problem 3.4 for the HSB model.

3.6

Take a videocassette registered under the NTSC system. How will it be displayed by a PAL videocassette recorder (VCR)? Explain your answer.

3.7

Implement the Huffman coding algorithm. Test the software on the following example: consider a file formed by 10,000 A, 2,000 B, 25,000 C, 5,000 D, 40,000 E, 18,000 F. Compute how many bits are required to code the file.

3.8

Consider the file formed by 20,000 B, 2,500 C, 50,000 D, 4,000 E, 1,800 F. Compare, in terms of memory required, fix-length and Huffman coding. Does there exist a case where fix-length and Huffman coding require the same memory resources? Explain your answer.

3.9

How much memory is required to store the movie Casablanca in its uncompressed version? Assume that the movie is black/white, has 25 frame/sec (each frame is 640 \(\times \) 480 pixels), its runtime is 102 min. For sake of simplicity, do not consider the memory required to store the audio of the movie.

3.10

Repeat the Problem 3.9 for the movie Titanic. Titanic is a color movie, has 30 frame/sec, and its runtime is 194 min.

3.11

Repeat the Problem 3.10 for the high definition version of the movie Titanic. Assume that each frame i is 1,920 \(\times \) 1,240 pixels and that movie is visualized using PAL or Secam system.

3.12

Repeat Problem 3.4 for the HIS model.

3.13

Repeat Problem 3.4 for the YUV model.

3.14

Implement Hu’s moments. Test your implementation on an Image verifying that the moments are invariant w.r.t. rotation.

3.15

Write the mathematical expression of the two dimensional Wavelet transform.

3.16

Implement the Wavelet Transform using Haar scaling function as Mother Function.

3.17

Consider the second-order Hessian matrix, H, defined as follows:

$$\begin{aligned} H = \left[ \begin{array}{ll} D_{xx} \quad D_{xy} \\ D_{yx} \quad D_{yy} \end{array} \right] \end{aligned}$$
(3.72)

Let \({\textit{Tr}}(H)\) and \({\textit{Det}}(H)\) be the trace and the determinant of the matrix H, respectively. Prove that holds the following formula

$$\begin{aligned} \varGamma = \frac{{\textit{Tr}}(H)^2}{{\textit{Det}}(H)} = \frac{(r+1)^2}{r}, \end{aligned}$$
(3.73)

where r is the ratio between the larger and the smaller eigenvalue. Moreover, show that \(\varGamma \) takes the minimum when r is equal to 1.

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer-Verlag London

About this chapter

Cite this chapter

Camastra, F., Vinciarelli, A. (2015). Image and Video Acquisition, Representation and Storage. In: Machine Learning for Audio, Image and Video Analysis. Advanced Information and Knowledge Processing. Springer, London. https://doi.org/10.1007/978-1-4471-6735-8_3

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-6735-8_3

  • Published:

  • Publisher Name: Springer, London

  • Print ISBN: 978-1-4471-6734-1

  • Online ISBN: 978-1-4471-6735-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics