Visual Attention Guided Seed Selection for Color Image Segmentation

  • Nabil Ouerhani
  • Neculai Archip
  • Heinz Hügli
  • Pierre-Jean Erard
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 2124)


The ”seeded region growing” (SRG) is a segmentation technique which performs an image segmentation with respect to a set of initial points, known as seeds. Given a set of seeds, SRG then grows the regions around each seed, based on the conventional region growing postulate of similarity of pixels within regions. The choice of the seeds is considered as one of the key steps on which the performance of the SRG technique depends. Thus, numerous knowledge-based and pure data-driven techniques have been already proposed to select these seeds. This paper studies the usefulness of visual attention in the seed selection process for performing color image segmentation. The purely data-driven visual attention model, considered in this paper, provides the required points of attention which are then used as seeds in a SRG segmentation algorithm using a color homogeneity criterion. A first part of this paper is devoted to the presentation of the multicue saliency-based visual attention model, which detects the most salient parts of a given scene. A second part discusses the possibility of using the so far detected regions as seeds to achieve the region growing task. The last part is dedicated to experiments involving a variety of color images.


color image segementation visual attention seed selection 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Itti, L., Koch, Ch., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), Vol. 20(11) (1998) 1254–1259CrossRefGoogle Scholar
  2. 2.
    Milanese, R.: Detecting salient regions in an image: from biological evidence to computer implementation. Ph.D. Thesis, Dept. of Computer Science, University of Geneva, Switzerland (Dec 1993)Google Scholar
  3. 3.
    Culhane, S.M., Tsotsos, J.K.: A prototype for data-driven visual attention. ICPR92, Vol. 1. Hague, Netherland (1992) 36–40Google Scholar
  4. 4.
    Treisman, A.M., Gelade, G.: A feature-integration theory of attention. Cognitive Psychology 12 (1980) 97–136CrossRefGoogle Scholar
  5. 5.
    Ratan, A.L.: The role of fixation and visual attention in object recognition. MIT AI-Technical Report 1529 (1995)Google Scholar
  6. 6.
    Todt, E., Torras, C.: Detection of natural landmarks through multi-scale opponent features. ICPR 2000, Vol. 3. Barcelona (Sep 2000) 988–1001Google Scholar
  7. 7.
    Ouerhani, N., Hugli, H.: Computing visual attention from scene depth. ICPR 2000, Vol. 1. Barcelona (Sep 2000) 375–378Google Scholar
  8. 8.
    Adams, R., Bischof, L.: Seeded region growing. IEEE Trans. on Pattern Analysis and Maschine Intelligence, vol. 16, no. 6 (1994)Google Scholar
  9. 9.
    Koethe, U.: Primary image segmentation. 17. DAGM-Symposium. Springer (1995)Google Scholar
  10. 10.
    O’Gorman, L., Sanderson, A.C.: The converging squares algorithm: An efficient method for locating peaks in multidimensions. IEEE Trans. Pattern Analysis and Machine Intelligence, PAMI, Vol. 6 (1984) 280–288CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2001

Authors and Affiliations

  • Nabil Ouerhani
    • 1
  • Neculai Archip
    • 2
  • Heinz Hügli
    • 1
  • Pierre-Jean Erard
    • 2
  1. 1.Institute of MicrotechnologyUniversity of NeuchâtelNeuchâtelSwitzerland
  2. 2.Institute of Computer SciencesUniversity of NeuchâtelNeuchâtelSwitzerland

Personalised recommendations