Brain Modularity

Brain Modularity

Using the your school Online Library; find two peer-reviewed journal articles on brain modularity, with a focus on visual sensation and perception. In your synopsis, you will include:

  • A summary of each of the journal articles
  • The main points discussed in each of the journal articles and how they relate to the week’s course and text readings
  • Your thoughts and perspectives regarding the concepts covered in each of the journal articles

Submission Details:

  • Name your document: SU_PSY3400_W2_Project_LastName_FirstInitial
  • Submit your report in a Microsoft Word document to the Submissions Area by the due date assigned.
  • Using APA format, cite sources appropriately throughout your assignment, and reference on a separate page.

    Brain Topography, Volume 18, Number 2, Winter 2005 (©2005) 67 DOI: 10.1007/s10548-005-0276-8

    Save your time - order a paper!

    Get your paper written from scratch within the tight deadline. Our service is a reliable solution to all your troubles. Place an order on any task and we will take care of it. You won’t have to worry about the quality and deadlines

    Order Paper Now

    Borowsky et al.68

    Modularity and Intersection 69

    Borowsky et al.70

    Modularity and Intersection 71

    Borowsky et al.72

    Modularity and Intersection 73

    Borowsky et al.74

    Modularity and Intersection 75

    Reproduced with permission of the copyright owner. Further reproduction prohibited without permission.

    218

    Perceiving visually presented objects: recognition, awareness, and modularity Anne M Treisman* and Nancy G Kanwisherf

    Object perception may involve seeing, recognition,

    preparation of actions, and emotional responses-functions

    that human brain imaging and neuropsychology suggest are

    localized separately. Perhaps because of this specialization,

    object perception is remarkably rapid and efficient.

    Representations of componential structure and interpolation

    from view-dependent images both play a part in object

    recognition. Unattended objects may be implicitly registered,

    but recent experiments suggest that attention is required to

    bind features, to represent three-dimensional structure, and to

    mediate awareness.

    Addresses *Department of Psychology, Princeton University, Princeton, New Jersey 08544-1010, USA; e-mail: [email protected] tDepartment of Brain and Cognitive Sciences, El O-243, Massachusetts Institute of Technology, Cambridge, Massachusetts 02138, USA; e-mail: [email protected]

    Current Opinion in Neurobiology 1998, 8:218-226

    http://biomednet.com/elecref/0959438800800218

    0 Current Biology Ltd ISSN 0959-4388

    Abbreviations

    ERP event-related potential fMRl functional magnetic resonance imaging IT inferotemporal cortex

    Introduction It is usually assumed that perception is mediated by specific patterns of neural activity that encode a selective

    description of what is seen, distinguishing it from other

    similar sights. When we perceive an object, we may form

    multiple representations, each specialized for a different

    purpose and therefore selecting different properties to

    encode at different levels of detail. There is empirical

    evidence supporting the existence of six different types

    of object representation. First, representation as an ‘object

    token’-a conscious viewpoint-dependent representation

    of the object as currently seen. Second, as a ‘structural de-

    scription’- a non-visually-conscious object-centered rep-

    resentation from which the object’s appearance from other

    angles and distances can be predicted. Third, as an

    ‘object type’-a recognition of the object’s identity (e.g. a

    banana) or membership in one or more stored categories.

    Fourth, a representation based on further knowledge

    associated with the category (such as the fact that the

    banana can be peeled and what it will taste like). Fifth, a

    representation that includes a specification of its emotional

    and motivational significance to the observer. Sixth, an

    ‘action-centered description’, specifying its “affordances”

    [l], that is, the properties we need in order to program

    appropriate motor responses to it, such as its location,

    size and shape relative to our hands. These different

    representations are probably formed in an interactive

    fashion, with prior knowledge facilitating the extraction of

    likely features and structure, and vice versa.

    Evidence suggests that the first four types of encoding

    depend primarily on the ventral (occipitotemporal) path-

    way, the fifth on connections to the amygdala, and the

    sixth on the dorsal (occipitoparietal) pathway; however,

    object tokens have also been equated with action-centered

    descriptions [PI. Dorsal representations appear to be

    distinct from those that mediate conscious perception;

    for example, grasping is unaffected by the Titchener

    size illusion [3]. Emotional responses can also be evoked

    without conscious recognition (e.g. see [4**]). Object

    recognition models differ over whether the type or identity

    of objects is accessed from the view-dependent token or

    from a structural description; in some cases, it may also be

    accessed directly from simpler features.

    The goal of perception is to account for systematic

    patterning of the retinal image, attributing features to their

    real world sources in objects and in the current viewing

    conditions. In order to achieve these representations,

    multiple sources of information are used, such as color,

    luminance, texture, relative size, dynamic cues from mo-

    tion and transformations, and stereo depth; however, the

    most important is typically shape. Many challenges arise in

    solving the inverse problem of retrieving the likely source

    of the retinal image: information about object boundaries

    is often incomplete and noisy; and three-dimensional

    objects are seen from multiple views, producing different

    two-dimensional projections on the retina, and objects in

    normal scenes are often partially occluded. The visual

    system has developed many heuristics for solving these

    problems. Continuity is assumed rather than random varia-

    tion. Regularities in the image are attributed to regularities

    in the real world rather than to accidental coincidences.

    Different types of objects and different levels of specificity

    require diverse discriminations, making it likely that

    specialized modules have evolved, or developed through

    learning, to cope with the particular demands of tasks

    such as face recognition, reading, finding our way through

    places, manipulating tools, and identifying animals, plants,

    minerals and artifacts.

    Research on object perception over the past year has made

    progress on a number of issues. Here, we will discuss

    recent advances in our understanding of the speed of

    object recognition, object types and tokens, and attention

    and awareness in object recognition. In addition, we will

    Perceiving visually presented objects Treisman and Kanwisher 219

    review evidence for cortical specializations for particular

    components of visual recognition.

    The speed of object recognition Evolutionary pressures have given high priority to speed

    of visual recognition, and there is both psychological and

    neuroscientific evidence that objects are discriminated

    within one or two hundred milliseconds. Behavioral

    studies have demonstrated that we can recognize up to

    eight or more objects per second, provided they are

    presented sequentially at fixation, making eye movements

    unnecessary [S]. Although rate measurements cannot tell

    us the absolute amount of time necessary for an individual

    object to be recognized, physiological recordings reveal

    the latency at which the two stimulus classes begin to

    be distinguished. Thorpe et al. [6”] have demonstrated significant differences in event-related brain potential

    (ERP) waveforms for viewing scenes containing animals

    versus scenes not containing animals at 150 ms after stim-

    ulus onset. Several other groups [7,8*,9-111 have found

    face-specific ERPs and magnetoencephalography (MEG)

    waveforms with latencies of 155-190 ms. DiGirolamo and

    Kanwisher (G DiGirolamo, NG Kanwisher, abstract in

    Psychonom Sot 1995, 305) found ERP differences for line drawings of familiar versus unfamiliar three-dimensional

    objects at 170 ms (see also [S]).

    Parallel results were found in the stimulus selectivity

    of early responses of cells in inferotemporal (IT) cortex

    in macaques, initiated at latencies of 80-looms. On

    the basis that IT cells are selective for particular faces

    even in the first 50ms of their response, Wallis and

    Rolls [12] conclude that “visual recognition can occur

    with largely feed-forward processing”. The duration of

    responses by these face-selective cells was reduced from

    250ms to 25 ms by a backward mask appearing 20ms

    after the onset of the face, a stimulus onset asynchrony

    at which human observers can still just recognize the

    face. The data suggest that “a cortical area can perform

    the computation necessary for the recognition of a visual

    stimulus in ZO-30ms”. Thus, a consensus is developing

    that the critical processes involved in object recognition

    are remarkably fast, occurring within lOO-200ms of

    stimulus presentation. However, it may take another

    1OOms for subsequent processes to bring this information

    into awareness.

    Object tokens How then does the visual system solve the problems of

    object perception with such impressive speed and accu-

    racy? A first stage must be a preliminary segregation of the

    sensory data that form separate candidate objects. Even

    at this early level, familiarity can override bottom-up cues

    such as common region and connectedness, supporting

    an interactive cascade process in which “partial results of

    the segmentation process are sent to higher level object

    representations”, which, in turn, guide the segmentation

    process [ 13.1.

    Kahneman, Treisman, and Gibbs [14] have proposed

    that conscious seeing is mediated by episodic ‘object

    files’ within which the object tokens defined earlier

    are constructed. Information about particular instances

    currently being viewed is selected from the sensory

    array, accumulates over time, and is ‘bound’ together in

    structured relations. Evidence for this claim came partly

    from the observation of ‘object-specific’ priming- that

    is, priming that occurs only, or more strongly, when the

    prime and probe are seen as a single object. This occurs

    even when they appear in different locations, if the

    object is seen in real or apparent motion between the

    two. Object-specific priming occurs between pictures and

    names when these are perceptually linked through the

    frames in which they appear (RD Gordon, DE Irwin,

    personal communication), suggesting that object files

    accumulate information not only about sensory features

    but also about more abstract identities. However, priming

    between synonyms or semantic associates is not object

    specific [15], that is, it occurs equally whether they

    are presented in the same perceptual object or in

    different objects. It appears that object files integrate

    object representations with their names, but maintain

    a distinct identity from other semantically associated

    objects. Priming at this level would be between object

    types rather than tokens. Irwin [ 161 has reviewed evidence on transsaccadic integration, suggesting that it is limited to

    about four object files.

    A similar distinction between tokens and types has

    emerged from the study of repetition blindness, a failure

    to see a second token of the same type, which was

    attributed to refractoriness in attaching a new token to

    a recently instantiated type [17]. Recent research has

    further explored this idea. One role of object tokens is

    to maintain spatiotemporal continuity of objects across

    motion and change. Chun and Cavanagh [18”] confirmed

    that repetition blindness is greater when repeated items

    are seen to occur within the same apparent motion

    sequence and hence are integrated as the same perceived

    object. They suggest that perception is biased to minimize

    the number of different tokens formed to account for the

    sensory data. Objects that appear successively are linked

    whenever the spatial and temporal separations make

    this physically plausible. This generally gives veridical

    perception because in the real world, objects seldom

    appear from nowhere or suddenly vanish. Arnell and

    Jolicoeur [ 191 have demonstrated repetition blindness for novel objects for which no pre-existing representations

    existed. According to Kanwisher’s account [ 171, this implies that a single presentation is sufficient to establish

    an object type to which new tokens will be matched.

    The ‘attentional blink’ [ZO] describes a failure to de-

    tect the second of two different targets when it is

    presented soon after the first. Chun (21’1 sees both

    repetition blindness and the attentional blink as failures

    of tokenization, although for different reasons, because

    220 Cognitive neuroscience

    they can be dissociated experimentally. Attentional blinks

    (reduced by target-distractor discriminability) reflect a

    Di I,ollo, JT Enns, personal communication). The account proposed

    is that awareness depends on a match between re-entrant

    information and the current sensory input at early

    visual levels. A mismatch erases the initial tentative

    representation. “It is as though the visual system treats the

    trailing configuration as a transformation or replacement

    of the earlier one.” Conversely, repetition blindness for

    locations (R Epstein, NG Kanwisher, abstract in Psychononz

    Sot 1996, 593) may result when the representation of an

    earlier-presented letter prevents the stable encoding of

    a subsequently presented letter appearing at the same

    location.

    Attention and awareness in object perception Attention seems, then, to be necessary for object tokens

    to mediate awareness. However, there is evidence (see

    [Z-l’]) that objects can be identified without attention

    and awareness. If this is so, do the representations differ

    from those formed with attention? Activation (shown

    by brain-imaging) in specialized regions of cortex for

    processing faces [26] and visual motion [27] is reduced

    when subjects direct attention away from the faces or

    moving objects (respectively), even when eye movements

    are controlled to guarantee identical retinal stimulation

    (see also [28]), consistent with the effects of attention

    on single units in macaque visual cortex. Unattended

    objects are seldom reportable. However, priming studies

    suggest that their shapes can be implicitly registered

    [?.9,30**], although there are clear limits to the number of

    unattended objects that will prime [31]. Representations

    formed without attention may differ from those that

    receive attention: they appear to be viewpoint-dependent

    [32’], two-dimensional, with no interpretation of occlusion

    or amodal completion [30”]. On the other hand, in

    clinical neglect, the ‘invisible’ representations formed in

    a patient’s neglected field include illusory contours and

    filled-in surfaces [33-l, suggesting that neglect arises at

    stages of processing beyond those that are suppressed in

    normal selective attention. With more extreme inattention,

    little explicit information is available beyond simple

    features such as location, color, size, and gross numerosity;

    even these simple features may not be available, produc-

    ing ‘inattentional blindness’ [34’]. Again, however, some

    implicit information is registered: unseen words may prime

    word fragment completion, and there is clear selectivity

    for emotionally important objects such as the person’s own

    name and happy (but not sad) faces.

    Binding of features to objects is often inaccurate unless

    attention is focused on the relevant locations [35].

"Get 15% discount on your first 3 orders with us"
Use the following coupon
"FIRST15"

Order Now