Dov Sagi

Dov Sagi ++

 

From Images to Visual Perception

 

Our brains continuously transform sensory information into meaningful perceptions. We try to understand the brain processes involved in visual transformations, such as encoding sequences of line points into curves, curves into shapes, and shapes into recognizable images. Though these processes are mostly visual, we find they also make use of more general brain functions, such as information chunking, learning and memory, that are employed by other brain faculties. Transforming light into images may not be a very different task from transforming sound waves into music or ideas into thoughts.

Toward achieving the goal of understanding human vision we use psychophysical methods, in an attempt to quantify perceptual and cognitive abilities. Though human brain is not yet readily accessible to direct activity measurements, much of its logic can be uncovered by measuring human performance in well controlled settings. For example, our inability to discriminate between some color mixtures (red+green = yellow) puts constraints on models of color processing, and detailed experiments can be carried out to further understand the way we see colors. In a similar way, we try to understand processes involved in pattern vision, by manipulating specific shape components. We design computer generated displays aimed at testing human performance on well defined detection and discrimination tasks, with visual targets carefully designated to probe brain processes such as image segmentation, perceptual organization, learning and mental imagery.

As our visual system is constructed from many interacting modules, each dealing with a different aspect of the visual task, we made a strategic decision to start from relatively simple low level processes involved in image segmentation. These processes, probably residing at the entrance stage of the visual cortex, were believed to be devoid of cognitive intervention, and indeed we could successfullymodel performance on texture segmentation and perceptual grouping tasks by using simple localized image-analyzers with lateral excitatory and inhibitory interactions. The architecture of these interactions was explored using contrast detection tasks. However, it became evident that performance on these texture tasks improves with time, pointing toward learning effects. Further experiments provided evidence of a genuine learning process, probably governed by associative rules, occuring at an early stage of visual processing. Our extended knowledge of segmentation processes contributed here to an understanding of the learning process and to quantify some learning abilities. Recently we have also demonstrated cognitive modulation of lateral interactions by usingmental imagery (as in trying to imagine a visual object).

The research described above is being carried out in collaboration with young scientists, working toward their doctoral degree. The Weizmann Institute and the Center for the study of Higher Brain Function produce an excellent environment for them to express their original ideas and to mature as excellent scientists. Other collaborative efforts are made with Neurologists at the Sheba Medical Center (Tel-Hashomer) and at the Loewenstein Rehabilitation Hospital (Raanana) to better understand brain damage affecting vision. I received my Ph.D. in Neurobiology from the Hebrew University (Jerusalem) and further training at the AT&T Bell Laboratories and at the California Institute of Technology.


Rubenstein, B. S. and Sagi, D. (1990): Spatial variability as a limiting factor in texture discrimination tasks: Implication for performance asymmetries. Journal of the Optical Society of America A, 7, 1632-1643.

Karni, A. and Sagi, D. (1993): The time course of learning a visual skill. Nature, 365, 250-252.

Polat, U. and Sagi, D. (1994): Spatial interactions in human vision: from near to far via experience dependent cascades of connections.Proceed. of the Nat. Acad. of Sci. USA, 91, 1206-1209.

Barchilon Ben-Av, M. and Sagi, D. (1995): Perceptual grouping by similarity and proximity: Experimental results can be predicted by intensity autocorrelations. Vision Research, 35, 853-866.

Ishai A. and Sagi, D. (1995) Common Mechanisms of Visual Imagery and Perception. Science, 268, 1772-1774.

  • Click here for more publications.
  • Click here for the Psychophysics Lab members .  

    Tel: +972(8)934-3747

    Fax: +972(8)934-4131

    e-mail: Dov.Sagi@Weizmann.ac.il

     

     

    Images used to explore perceptual organization. The display elements are arranged to produce vertical grouping based on elements similarity in A, horizontal grouping based on elements proximity in B, and ambiguous organization when similarity cues are put in conflict with proxmity cues in C. Human performance on such grouping tasks can be predicted by intensity autocorrelations.


    This site was developed by Chaipi Wijnbergen using Accent Multilingual Publisher on 2/4/96

    For more information send mail to Ariela Sharfi

    [counter]