Skip to main content
Log in

An Annotation Engine for Supporting Video Database Population

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Segmentation, video data modeling, and annotation are indispensable operations necessary for creating and populating a video database. To support such video databases, annotation data can be collected as metadata for the database and subsequently used for indexing and query evaluation. In this paper we describe the design and development of a video annotation engine, called Vane, intended to solve this problem as a domain-independent video annotation application.

Using the Vane tool, the annotation of raw video data is achieved through metadata collection. This process, which is performed semi-automatically, produces tailored SGML documents whose purpose is to describe information about the video content. These documents constitute the metadatabase component of the video database. The video data model which has been developed for the metadata, is as open as possible for multiple domain-specific applications. The tool is currently in use to annotate a video archive comprised of educational and news video content.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. T.G. Aguierre Smith and N.C. Pincever, “Parsing movies in context,” Proc. USENIX, pp. 157–168, Summer 1991.

  2. T.G. Aguierre Smith and G. Davenport, “The stratification system: A design environment for random access video,” in 3rd Intl. Workshop on Network and Operating System Support for Digital Audio and Video, Nov. 1992.

  3. G. Ahanger and T.D.C. Little, “A survey of technologies for parsing and indexing digital video,” Journal of Visual Communication and Image Representation, Vol. 7, No. 1, pp. 28–43, March 1996.

    Google Scholar 

  4. F. Arman, A. Hsu, and M-.Y. Chiu, “Image processing on compressed data for large video databases,” in 1st ACM Intl. Conf. on Multimedia, pp. 267–272, Aug. 1993.

  5. V. Bhaskaran and K. Konstantinides, Image and Video Compression Standards: Algorithms and Architectures, Kluwer Academic Publishers, 1995.

  6. K. Böhm and T.C. Rakow, “Metadata for multimedia documents,” SIGMOD Record, Vol. 23, No. 4, pp. 21–26, Dec. 1994.

    Google Scholar 

  7. M. Carreira, J. Casebolt, G. Desrosiers, and T.D.C. Little, “Capture-time indexing paradigm, authoring, tool, and browsing environment for digital broadcast video,” Vol. SPIE 2417, pp. 380–388, May 1995.

    Google Scholar 

  8. F. Chen, M. Hearst, J. Kupiec, J. Pedersen, and L. Wilcox, “Metadata for mixed-media access,” SIGMOD Record, Vol. 23, No. 4, pp. 64–71, Dec. 1994.

    Google Scholar 

  9. G. Davenport, T.A. Smith, and N. Pincever, “Cinematic primitives for multimedia,” IEEE Computer Graphics and Applications, pp. 67–74, July 1991.

  10. M. Davis, “Media streams: An iconic visual language for video annotation,” Proc. IEEE Symposium on Visual Languages, pp. 196–202, 1993.

  11. W.I. Grosky, F. Fotouhi, and I.K. Sethi, “Using metadata for the intelligent browsing of structured media objects,” SIGMOD Record, Vol. 23, No. 4, pp. 49–56, Dec. 1994.

    Google Scholar 

  12. A. Hampapur, R. Jain, and T. Weymouth, “Digital video segmentation,” in 2nd ACM Intl. Conf. on Multimedia, pp. 357–364, 1994.

  13. R. Hjelsvold, S. Langorgen, R. Midtstraum, and O. Sandsta, “Integrated video archive tools,” in 3rd ACM Intl. Multimedia Conf., pp. 5–9, Nov. 1995.

  14. W. Klas and A. Sheth, “Metadata for digital media: Introduction to the special issue,” SIGMOD Record, Vol. 23, No. 4, pp. 19–20, Dec. 1994.

    Google Scholar 

  15. W. Kou, Digital Image Compression: Algorithms and Standards, Kluwer Academic Publishers, 1995.

  16. J. Lee and B.W. Dickinson, “Multiresolution video indexing for subband coded video databases,” in IS&T/SPIE, Conference on Storage and Retrieval for Image and Video Databases, Feb. 1994.

  17. A. Nagasaka and Y. Tanaka, “Automatic video indexing and full-video search for object appearances,” in Visual Database Systems, II, E. Knuth and L.M. Wegner (Eds.), IFIP, Elsevier Science Publishers B.V., 1992, pp. 113–127.

  18. K. Otsuji and Y. Tonomura, “Projection detecting filter for video cut detection,” in 1st ACM Intl. Conf. on Multimedia, pp. 251–257, Aug. 1993.

  19. J.K. Ousterhout, Tcl and the TK Toolkit, Addison-Wesley Publishing Company, 1994.

  20. R.W. Picard and T.P. Minka, “Vision texture for annotation,” M.I.T. Media Laboratory Perceptual Computing Section Technical Report, Vol. 1, No. 302, pp. 3–14, 1995.

    Google Scholar 

  21. L.A. Rowe, J.S. Boreczky, and C.A. Eads, “Indexes for user access to large video databases,” Vol. SPIE 2185, pp. 150–161, Feb. 1994.

    Google Scholar 

  22. Y. Tonomura, A. Akutsu, Y. Taniguchi, and G. Suzuki, “Structured video computing,” IEEE Multimedia, pp. 34–43, Fall 1994.

  23. B.B. Welch, Practical Programming in Tcl and Tk, Addison-Wesley Publishing Company: Upper Saddle River, New Jersey, 1995.

    Google Scholar 

  24. H.J. Zhang, A. Kankanhalli, and S.W. Smoliar, “Automatic partitioning of full-motion video,” ACM/Springer Multimedia Systems, Vol. 1, No. 1, pp. 10–28, 1993.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Carrer, M., Ligresti, L., Ahanger, G. et al. An Annotation Engine for Supporting Video Database Population. Multimedia Tools and Applications 5, 233–258 (1997). https://doi.org/10.1023/A:1009647624347

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/A:1009647624347

Navigation