ARTICLES

Role for Human Posterior Parietal Cortex in Visual Processing of Aversive Objects in Peripersonal Space

Published Online:https://doi.org/10.1152/jn.00614.2005

Abstract

The posterior parietal cortex of both human and non-human primates is known to play a crucial role in the early integration of visual information with somatosensory, proprioceptive and vestibular signals. However, it is not known whether in humans this region is further capable of discriminating if a stimulus poses a threat to the body. In this functional magnetic resonance imaging (fMRI) study, we tested the hypothesis that the posterior parietal cortex of humans is capable of modulating its response to the visual processing of noxious threat representation in the absence of tactile input. During fMRI, participants watched while we “stimulated” a visible rubber hand, placed over their real hand with either a sharp (painful) or a blunt (nonpainful) probe. We found that superior and inferior parietal regions (BA5/7 and BA40) increased their activity in response to observing a painful versus nonpainful stimulus. However, this effect was only evident when the rubber hand was in a spatially congruent (vs. incongruent) position with respect to the participants’ own hand. In addition, areas involved in motivational-affective coding such as mid-cingulate (BA24) and anterior insula also showed such relevance-dependent modulation, whereas premotor areas known to receive multisensory information about limb position did not. We suggest these results reveal a human anatomical-functional homologue to monkey inferior parietal areas that respond to aversive stimuli by producing nocifensive muscle and limb movements.

INTRODUCTION

The brain is organized to support the complex spatial and sensorimotor representations required for interactions with objects in the world, such as navigation and fine manipulation. However, for objects that pose a direct threat to the body, visuospatial and sensorimotor mechanisms may “flag” potentially noxious stimuli relatively early in the visual-processing stream, for example, within structures of the posterior parietal cortex. Such threats are directly relevant to the body and are likely to be most imminent when they occur within the space surrounding a particular body part. In the current study, we investigated whether visual processing of noxious threat-related objects in peripersonal space would activate posterior parietal cortex more than the visual response to a nonthreatening object in the same space.

There is evidence that visuo-tactile receptive fields in monkey inferior parietal and intraparietal cortex are sensitive to both tactile and visual information about noxious stimuli (Dong et al. 1994). Cells in this region have also been associated with nocifensive movements of aversion (Cooke and Graziano 2003). Apart from these findings, networks classically implicated in the evaluation of the behavioral relevance of aversive stimuli have not included posterior parietal cortex. Rather in the context of pain processing, medial frontal and limbic regions (such as the anterior cingulate and anterior insula cortices) are associated with the evaluation of the motivational and behavioral relevance of the stimulus on the basis of visual or nociceptive information (Botvinick et al. 2005; Jackson et al. 2005; Morrison et al. 2004; Singer et al. 2004), although these studies have largely been concerned with the processing of empathy for pain.

In many situations, the organism must track potential threats in terms of their spatial proximity to particular body parts, and in such circumstances, it would be advantageous for underlying visuomotor representations to be dynamically sensitive to events in the space surrounding that body part. The aim of the current study was to investigate the visual processing of aversive objects in peripersonal space, specifically that surrounding the hand. One possible way of pursuing this would be to stimulate participants’ hand with an aversive (i.e., noxious) stimulus. However, because limb representation involves the multisensory integration of visual, tactile, and proprioceptive cues, this would include the contribution of nociceptive tactile input. An alternative way of isolating the visual component of noxious threat would be to present visual stimuli close to a realistic and aligned artificial limb in peripersonal space that could then be “stimulated” in the absence of actual tactile input. This manipulation has been shown in several studies to result in shifts of the felt location of the limb/and biases in proprioception and or reaching (Armel and Ramachandran 2003; Ehrsson et al. 2004; Farnè et al. 2000; Graziano 1999; Holmes et al. 2005; Pavani et al. 2000; Tastevin 1937; Tsakiris and Haggard 2005; Walton and Spence 2004).

Artificial hands have also recently been used to investigate the neural correlates of subjective limb ownership (Ehrsson et al. 2004). Previous behavioral research has shown that when both an artificial hand and the person’s own hand (which is hidden either beneath or at the side of the person’s own hand) are stroked repeatedly and synchronously by the experimenter, some participants can have the experience that the touch they feel on their own hand is located where they see the rubber hand being touched. This sensation is often accompanied by a sense of ownership of the rubber hand (Botvinick and Cohen 1998). Using functional magnetic resonance imaging (fMRI), Ehrsson and colleagues provided evidence that the subjective experience of ownership of the rubber hand correlates significantly with premotor cortex activation. In a separate analysis, this activation was observed as an interaction between the synchronicity of the stroking and the anatomical plausibility of the hand’s orientation. The impression of ownership of the rubber hand can be substantially reduced or even eliminated by placing the rubber hand in an anatomically implausible position with respect to the participant’s real hand, asynchronously touching the real and rubber hand and/or allowing vision of the real hand (see Maravita et al. 2003 for a review).

There is also a multitude of evidence from both animal electrophysiological recordings and human brain imaging studies to suggest that activity in the premotor and posterior parietal cortex (particularly the ventral-intraparietal area; VIP) represents both the seen and felt position of the hand (for a recent review, see Graziano et al. 2004). Multisensory cells within these regions fire both when the hand is touched or when a visual stimulus is presented near the hand (Rizzolatti et al. 1981) and when a fake hand is seen in place of the real hand (Ehrsson et al. 2004; Graziano 1999). In the monkey, area VIP has direct reciprocal connections with part of the ventral premotor cortex (F4 in PMv), the human homologue of which is the inferior frontal gyrus (BA44) (Rizzolatti et al. 1998), forming a circuit known as the VIP-F4 circuit (Luppino et al. 1999). Further evidence exists of a similar circuit in fronto-parietal regions of the human brain responsive to the multisensory representation of limb position (Lloyd et al. 2003).

A study by Armel and Ramachandran (2003) indicates that not only can an artificial hand be incorporated into the subject’s own body representation but that autonomic nervous system activity can occur according to the perceived threat of an object in contact with the artificial hand. They measured skin conductance responses (SCRs) while subjects experienced simultaneous and synchronous tactile stimulation of their own and a rubber hand. They found that if the rubber hand was suddenly and unexpectedly “injured” after this simultaneous tactile stimulation, subjects displayed a strong SCR even though they were aware that their real hand was never in danger.

To discover whether a potentially noxious visual stimulus, perceived within peripersonal hand space, would influence hemodynamic responses in the brain, especially within regions implicated in multisensory limb representation, we biased the integration of participants’ visual, tactile, and position senses by manipulating the position of the rubber hand over the participants’ own hand. Noxious (sharp) or innocuous (blunt) stimulation of the rubber hand was preceded by simultaneous stroking of the real and rubber hand to facilitate participants’ perception of the rubber hand as within body-part centered space (although visual capture of limb proprioception can also occur in the absence of synchronous stroking) (see Farnè et al. 2000; Holmes et al. 2005; Pavani et al. 2000; Rorden et al. 1999; Walton and Spence 2004). Specifically, areas previously shown to play a role in coding the space surrounding the hand—such as posterior parietal and premotor cortex—are predicted to distinguish a sharp probe versus a blunt probe striking the rubber hand. Importantly, we expect any such discrimination to depend on the anatomical plausibility of the false limb’s real position influencing the apparent position of the invisible real limb position, responding more when the rubber hand is oriented compatibly with the person’s own hand (and thus proprioceptively aligned). A secondary hypothesis is that activity in anterior cingulate and anterior insula, which respond to pain-related visual information and are involved in the motivational-affective aspect of pain processing, will likewise increase for the sharp versus the blunt probe. If these predictions are borne out, the results will provide the first neuroimaging evidence that regions supporting visuo-spatial representations of peripersonal space are capable of discriminating threatening stimuli near the hand.

METHODS

Participants

Twenty-eight participants (9 males, 19 females) aged between 22 and 50 yr (with a mean age of 29 yr) gave fully informed written consent of their willingness to participate in this study, which had local ethics committee approval. Fourteen participants took part in experiment 1 (rubber hand in a spatially congruent position with respect to the participant’s own hand), and a further 14 naïve participants took part in experiment 2 (rubber hand in a spatially incongruent position with respect to the participant’s own hand). All participants were strongly right-handed as assessed by the Edinburgh Handedness Inventory (Oldfield 1971) and in good health with no past history of psychiatric or neurological disease. Participants had normal or corrected-to-normal (with contact lenses) visual acuity and normal tactile sensation.

Apparatus and materials

A realistic-looking rubber (right) hand was placed on top of the participant’s own right hand (seeFig. 1 ). A piece of semi-circular plastic piping was placed in between the rubber hand and the participant’s own hand to ensure that touching the rubber hand did not inadvertently tactually stimulate the person’s real hand and was covered with a cloth to enable continuity of the perception of the fake hand and arm extending and occupying the position of the participant’s real hand and arm. A cotton bud (or Q-tip) and a syringe with a cocktail stick (also called a toothpick) in place of the steel lancet were used as the innocuous tactile and painful tactile probes, respectively. The cocktail stick caused a moderately painful pin-prick sensation when applied to the skin (as established on all participants prior to scanning) but does not contain metal and so can be used within the MRI environment. We were able to emulate the experience of a real syringe being applied to the skin by placing it in the plastic holder of a real syringe.

FIG. 1.

FIG. 1.Photograph of the experimental set-up. Inside the scanner, the rubber hand (representing a right hand) was placed above the participant’s real right hand in either an anatomically plausible (experiment 1) or implausible (experiment 2) position. A piece of semi-circular plastic tubing was placed between the participant’s real hand and the rubber hand to prevent secondary contact of the participants hand during stimulation of the rubber hand.


Design and procedure

Prior to scanning, all participants were exposed to the effect of the rubber hand by placing their own hand underneath the plastic tubing with the rubber hand on top (in the congruent position) and a blanket covering their arm to maintain the perception of the rubber hand as an extension of their own arm. We then began simultaneous stroking of the rubber hand and the participant’s own hand until they reported such statements to suggest that they could feel as if the rubber hand was their own hand or that they could feel “touch” on the rubber hand. This typically occurred after several seconds. We then stopped the procedure and showed the participants the cotton bud/Q-tip and modified syringe (with the cocktail stick in place instead of the steel lancet) and encouraged them to feel the tactile qualities of these two objects. All participants acknowledged the Q-tip as innocuous and the cocktail stick as moderately painful.

Stimuli were presented within a modified blocked design. At the start of each block, participants experienced 30 s of rest followed by 15 s of simultaneous stroking of the rubber (right) hand and their own right hand in a temporally synchronized and spatially compatible way. Specifically, both the rubber hand and the participant’s real hand were stroked by the experimenter (using their index finger) in a unidirectional way on the middle finger of the right hand starting at the finger nail and ending at the proximal interphalangeal joint (mid-way down the finger) at an approximate rate of 1 Hz. In studies where only visual cues of the fake/rubber hand were available (i.e., no simultaneous touch occurred), participants failed to incorporate the fake hand into the body image (Farnè et al. 2000; Pavani et al. 2000). We reasoned that an investigation of the visuo-nociceptive response to threatening objects in peripersonal space would benefit from including this simultaneous touch condition. However, we do not have any formal subjective measures (i.e., via questionnaires) of whether participants experienced the “rubber-hand illusion” per se as this was not the focus of the current study. This manipulation also ensured that in the absence of an explicit task, participants were aware of the rubber hand throughout the 7-min scan. After this time, either the cotton bud or the syringe was administered to the rubber hand for 15 s (the order of which was randomized between subjects), which the participants could see touching the rubber hand but not approaching the rubber hand given the confines of the scanner. Three blocks of each condition (i.e., noxious touch and innocuous touch) were presented. All stimuli were delivered at a steady rate of 1 Hz (practiced outside of the scanner), the duration of the stimulation was timed by a clock on the wall of the scanner, visible to the experimenter but not to the participant inside the scanner. During the prescan setup and during the experiment, participants were instructed to look only at the rubber hand (which they saw indirectly through a mirror positioned in the head coil of the scanner). Although we did not formally track their eye movements during the scan, we are confident participants were looking at the stimuli throughout this novel and stimulating experiment through postscan interviews as to their subjective impression of the effect of having the rubber hand touched with the different stimuli. Their own (right) hand was hidden from view underneath the rubber hand throughout the experiment, and participants were unable to see the experimenter touching their hand as the narrow bore of the magnet restricts the field of view of the participant such that he/she can see the rubber hand (placed over his/her own right hand) and the hands of the experimenter touching the rubber hand but not the body of the experimenter which is hidden from view at the side of the magnet. In experiment 1, the rubber hand was in an anatomically plausible (congruent) position with respect to the participant’s real arm (Fig. 1). In experiment 2, the position of the rubber hand was rotated 180° such that it faced toward the participant in what was deemed by all to be an anatomically implausible (incongruent) position. This was the optimum position for the rubber hand given the narrow confines of the scanner. Data from experiment 2 were collected in a different scanning session from that of experiment 1 and in a different set of participants to eliminate any familiarity or priming effects.

Scanning procedure

MR data were acquired using a 1.5 T Signa LX/Nvi neuro-optimized system (General Electric, Milwaukee, WI). fMRI was performed with a blood oxygenation-level-dependent (BOLD)-sensitive T2*-weighted multislice gradient echo EPI sequence (TE = 40 ms, TR = 3 s, flip angle = 90°, FOV = 190 mm, 64 × 64 matrix, in-plane resolution: 3 mm); 135 volumes were collected in a single EPI run. Twenty-four contiguous 5-mm-thick axial slices were prescribed parallel to the AC-PC line and covered the whole brain. For the purpose of anatomical referencing and visualization of brain activation, a high-resolution T1-weighted three-dimensional (3D) inversion recovery prepared gradient echo (IRp-GRASS) sequence was acquired (TE = 5.4 ms, TR = 12.3 ms, TI = 450 ms, 1.6-mm slice thickness, FOV = 200 mm, 256 × 192 matrix), with 124 axial slices covering the whole brain (in-plane resolution: 1 mm).

Data analysis was carried out using FEAT5 software (FMRI expert analysis tool, Oxford Centre for Functional Magnetic Resonance Imaging of the Brain, FMRIB, University of Oxford), part of the FMRIB software library [FSL 3.2; www.fmrib.ox.ac.uk/fsl (Smith et al. 2004)]. The following prestatistics processing was applied: motion correction using MCFLIRT (Jenkinson and Smith 2001); spatial smoothing using a Gaussian kernel of FWHM 5 mm; mean-based intensity normalization of all volumes by the same factor; nonlinear high-pass temporal filtering (Gaussian-weighted LSF straight line fitting, with σ = 60s). Statistical analysis was carried out using FILM (FMRIB’s improved linear model) with local autocorrelation correction of the data (nonlinear spatial smoothing and prewhitening) (Smith and Brady 1997; Woolrich et al. 2001).

For each individual subject, we fitted a linear regression model (general linear model, GLM) to the data (1st level analysis). Four covariates were analyzed separately corresponding to the four experimental conditions: two covariates of interest, viewing the painful stimulus touching the rubber hand (view pain, VP) and viewing the innocuous stimulus touching the rubber hand (view neutral, VN) and two covariates of no interest; simultaneous touching of the rubber hand and real hand prior to viewing the painful stimulus (pain rub, PR) and simultaneous touching of the rubber hand and real hand prior to viewing the innocuous stimulus (neutral rub, NR). In addition, linear contrasts were also defined within the GLM framework to identify areas in which the activity relating to the painful stimulus touching the rubber hand was greater than the activity to the innocuous stimulus touching the rubber hand [i.e., (VP − VN)], both when the hand was in the congruent position [i.e., congruent (VP − VN)] and incongruent position [i.e., incongruent (VP − VN)]. Similarly, contrasts were defined to measure activity in voxels where activity to the innocuous stimulus was greater than the painful stimulus [i.e., (VN − VP)], both when the arm was in a spatially congruent and incongruent position. The results from this analysis were contrast estimates for each condition for each of the 28 subjects (contrast images). To accommodate intersubject variability, the contrast images from all subjects were entered into a mixed effects group analysis (a second level analysis also known as random effects) carried out using FEAT5 software (Beckmann et al. 2003; Woolrich et al. 2004). z (Gaussianized T/F) statistic images were thresholded using clusters determined by Gaussian random field theory Resel-based correction, which gives the probability of a cluster, given its spatial extent and z threshold, under the null hypothesis and is therefore less conservative than the Bonferoni correction see (Friston et al. 1992; Worsley et al. 1992). Statistic images were thresholded using clusters determined by z > 2.3, P = 0.05 (corrected for multiple comparisons across the whole brain) and transformed into the stereotaxic space of the Montreal Neurological Institute (MNI) using FLIRT (FMRIB’s linear image registration tool) (Jenkinson and Smith 2001).

RESULTS

Analysis of main effects

Viewing a painful tactile probe touching the rubber hand (representing a right hand) with the hand placed in an anatomically plausible (congruent) position (vs. rest) revealed significant activation across a number of parietal and prefrontal regions as predicted. Activation contralateral to the position of the stimulated rubber hand (i.e., covering the person’s own right hand) was seen in inferior parietal cortex (BA40), premotor cortex (BA6), and inferior frontal gyrus (BA44) as well as extrastriate cortex (BA18) and right superior and middle temporal gyri. Bilateral activation was seen across middle and superior frontal gyri (BA6/8) and anterior cingulate cortex (see Table 1, Fig. 2). Similar sites of activity were observed when viewing a painful tactile stimulus touching the rubber hand with the hand in a spatially incongruent position with respect to participants’ own hands (vs. rest; see Table 1, Fig. 3). Contralateral activity was again observed in premotor and middle frontal gyrus (BA6/46), medial intraparietal sulcus and superior parietal lobe (BA7), extrastriate cortex (BA18/19), and anterior insula and posterior cingulate cortex (BA31). Ipsilateral activation was seen within parietal operculum (BA40) and precentral sulcus with bilateral activation occurring within superior temporal and inferior frontal gyri (BA45/46) and anterior cingulate cortex.

FIG. 2.

FIG. 2.Activation map illustrating a significantly greater response to viewing the painful tactile probe applied to the rubber hand when the hand was in an anatomically plausible position (with respect to the participants’ real hand) vs. rest (no stimulation). Maps were cluster-based thresholded at z > 2.3, P = 0.05 (corrected for multiple comparisons) and are shown in axial sections across the whole brain starting at the level of the brain stem (moving from left to right across the page in 4-mm slices) in radiological convention (right side of the brain on the left side of the picture). A, premotor cortex; B, superior parietal lobe; C, temporal-occipital cortex; D, inferior parietal lobe; E, cerebellum; F, anterior cingulate cortex.


FIG. 3.

FIG. 3.Activation map illustrating a significantly greater response to viewing the painful tactile probe applied to the rubber hand when the hand was in an anatomically implausible position (with respect to the participants’ real hand) vs. rest (no stimulation). Maps were cluster-based thresholded at z > 2.3, P = 0.05 (corrected for multiple comparisons) and are shown in axial sections across the whole brain starting at the level of the brain stem (moving from left to right across the page in 4-mm slices) in radiological convention (right side of the brain on the left side of the picture).


TABLE 1. Activation in response to a painful tactile probe touching the rubber hand

RegionMNI Coordinates (x,y,z mm)Maximum z scoreR/L
Congruent Arm Position
    Premotor cortex (BA6)−542345.70L
    Inferior parietal lobe (BA40)−56−28265.48L
    Extrastriate cortex (BA18)−44−80−25.44L
    Middle temporal gyrus56−5605.40R
    Middle frontal gyrus (BA6/8)5010444.40R
    Superior frontal gyrus (BA6)212604.18R
    Middle frontal gyrus (BA9)5818324.11R
    Inferior frontal gyrus (BA44)4418243.99R
    Superior temporal gyrus4824−223.81R
    Medial frontal gyrus (BA6)−40583.45L
    Anterior cingulate cortex (BA32)024383.41
Incongruent Arm Position
    Extrastriate cortex (BA18)−42−74−66.54L
    Medial intraparietal sulcus−38−48525.94L
    Parietal operculum (BA40)60−20225.93R
    Superior parietal cortex (BA7)−22−60625.82L
    Premotor cortex (BA6)−500305.28L
    Middle frontal gyrus (BA6)−26−8565.07L
    Superior temporal gyrus−40−8−64.90L
40−2−144.85R
    Precentral sulcus34−2504.86R
    Inferior frontal gyrus (BA45/46)4828124.57R
−403244.28L
    Anterior cingulate cortex (BA24)2−4344.38R
    Posterior cingulate gyrus (BA31)−14−32344.17L
    Anterior insula−362283.55L
    Middle frontal gyrus (BA46)−3442163.12L

Activation in response to a painful tactile stimulus touching the rubber hand (representing a right hand) when the hand is in a spatially congruent and incongruent position with respect to the participant’s own right hand (activations determined by clusters greater than z > 2.3, P = 0.05). MNI coordinate and peak z score of the maximum activating voxel in each cluster are shown.

Viewing an innocuous tactile probe touching the rubber hand with the hand placed in an anatomically plausible (congruent) position (vs. rest) revealed contralateral activation of premotor cortex (BA6), superior parietal, and temporal lobes and extrastriate visual cortex. Ipsilateral activation was seen in inferior and middle frontal gyrus (BA44), and bilateral activation in inferior parietal lobe (BA40; see Table 2, Fig. 4). With the rubber hand in a spatially incongruent position with respect to participants’ own right hand (vs. rest), significant activation was seen in the contralateral hemisphere in inferior frontal gyrus (BA44/45) and inferior and superior parietal lobes, extrastriate visual cortex (BA18), temporal lobes including the hippocampus and parahippocampal gyrus while premotor cortex activated bilaterally extending into the precentral sulcus, sylvian fissure and inferior frontal gyrus (BA47; see Table 2, Fig. 5). Ipsilateral activation was seen within middle frontal gyrus (BA6/9/10/11).

FIG. 4.

FIG. 4.Activation map illustrating a significantly greater response to viewing the innocuous tactile probe applied to the rubber hand when the hand was in an anatomically plausible position (with respect to the participants’ real hand) versus rest (no stimulation). Maps were cluster-based thresholded at z >2.3, P = 0.05 (corrected for multiple comparisons) and are shown in axial sections across the whole brain starting at the level of the brain stem (moving from left to right across the page in 4 mm slices) in radiological convention (right side of the brain on the left side of the picture). A = Premotor cortex; B = Superior parietal lobe; C = Temporal-occipital cortex; D = Inferior parietal lobe; E = Cerebellum.


FIG. 5.

FIG. 5.Activation map illustrating a significantly greater response to viewing the innocuous tactile probe applied to the rubber hand when the hand was in an anatomically implausible position (with respect to the participants’ real hand) versus rest (no stimulation). Maps were cluster-based thresholded at z >2.3, P = 0.05 (corrected for multiple comparisons) and are shown in axial sections across the whole brain starting at the level of the brain stem (moving from left to right across the page in 4 mm slices) in radiological convention (right side of the brain on the left side of the picture). A = Premotor cortex; B = Superior parietal lobe; C = Temporal-occipital cortex; D = Inferior parietal lobe; E = Cerebellum.


TABLE 2. Activation in response to an innocuous tactile probe touching the rubber hand

RegionMNI Co-ordinates (x,y,z mm)Maximum z ScoreR/L
Congruent Arm Position
    Extrastriate cortex (BA18)−46−76−25.91L
    Inferior parietal lobe (BA40)−56−30245.07L
68−22263.71R
    Superior parietal lobe (BA7)−32−56545.02L
    Premotor cortex (BA6)−502364.87L
    Middle frontal gyrus (BA6/8/9)488504.66R
    Superior temporal gyrus−54824.17L
    Inferior frontal gyrus (BA44)5816323.93R
Incongruent Arm Position
    Extrastriate cortex (BA18)−42−76−26.26L
    Inferior parietal lobe (BA40)−60−26285.74L
    Superior parietal lobe (BA7)−28−54505.71L
    Premotor cortex (BA6)−542345.13L
42−6423.78R
    Precentral sulcus−500404.90L
400423.96R
    Sylvian fissure−54444.70L
40−4−83.67R
    Middle frontal gyrus (BA6)36−4484.33R
    Inferior frontal gyrus (BA44/45)−444024.22L
    Superior temporal gyrus−40−12−84.04L
    Middle frontal gyrus (BA9/10/11)3264−83.94R
    Hippocampus−16−22−103.89L
    Inferior frontal gyrus (BA47)2632−183.83R
−4442−43.41L
    Inferior temporal gyrus−34−8−283.67L
    Parahippocampal gyrus−24−32−103.61L

Activation in response to an innocuous tactile stimulus touching the rubber hand (representing a right hand) when the hand is in a spatially congruent and incongruent position with respect to the participants own right hand (activations determined by clusters greater than z > 2.3, P = 0.05). MNI coordinate and peak z score of the maximum activating voxel in each cluster are shown with the laterality of response.

Contrast of main effects

A contrast of the main effects revealed those regions which activated significantly more in response to a painful tactile probe touching the rubber hand versus an innocuous tactile probe (and vice versa) either with the rubber hand in a spatially compatible (experiment 1) or incompatible (experiment 2) position with respect to participants’ own right hands. Viewing a painful versus innocuous tactile stimulus touching the rubber hand with the hand in a spatially congruent position [i.e., congruent (VP –VN)] revealed significant contralateral activation of superior and inferior parietal cortices, superior temporal gyrus, and sulcus and fusiform gyrus (BA19). Bilateral activation was observed in the VI lobe of the cerebellum, anterior cingulate cortex (BA24) and medial and superior frontal gyri (BA6) (Table 3; Fig. 6). With the rubber hand in a spatially incongruent position with respect to participants’ own hands [i.e., incongruent (VP –VN)] significantly greater activation was seen in response to the painful tactile stimulus versus the innocuous stimulus ipsilaterally in right putamen, superior temporal gyrus and anterior insula (Table 3). At the cluster-based threshold tested (z > 2.3, P = 0.05), no regions demonstrated significantly more activation to the innocuous tactile stimulus versus the painful stimulus applied to the rubber hand in either a spatially congruent or incongruent position.

FIG. 6.

FIG. 6.Activation map illustrating a significantly greater response to viewing the painful tactile probe applied to the rubber hand when the hand was in an anatomically plausible position (with respect to the participants’ real hand) versus the innocuous tactile probe. Maps were cluster-based thresholded at Z >2.3, P = 0.05 (corrected for multiple comparisons) and are shown in axial sections across the whole brain starting at the level of the brain stem (moving from left to right across the page in 4 mm slices) in radiological convention (right side of the brain on the left side of the picture).


TABLE 3. Regions showing significantly more activation in response to a painful vs. innocuous tactile probe touching the rubber hand

RegionMNI Coordinates (x,y,z mm)Maximum z ScoreR/L
Congruent Arm Position
    Superior parietal lobe (BA5)−30−40683.95L
    Anterior cingulate cortex (BA24)2−4363.53R
−8−6443.13L
    Sylvian fissure−64−34203.47L
    Cerebellum (VI lobe)28−62−263.41R
−18−72−223.48L
    Medial frontal gyrus (BA6)0−2643.43
−4−14702.94L
    Superior parietal lobe (BA7)−12−46623.38L
    Superior temporal sulcus−48−30123.27L
    Superior frontal gyrus (BA6)−84723.05L
8−2722.96R
    Fusiform gyrus (BA19)−34−68−123.00L
    Superior temporal gyrus−44−40122.78L
    Inferior parietal lobe (BA40)−60−36382.54L
Incongruent Arm Position
    Anterior Insula341225.03R
    Putamen261023.41R
    Superior temporal gyrus48203.14R

Areas showing significantly greater activation in response to a painful tactile stimulus touching the rubber hand (vs. an innocuous stimulus) both with the hand in a spatially congruent and incongruent position with respect to the participants own hand (activations determined by clusters greater than z > 2.3, P = 0.05). MNI coordinate and peak Z score of the maximum activating voxel in each cluster are shown with the laterality of response.

DISCUSSION

The major outcome of the current study was that regions of the contralateral posterior parietal cortex (surrounding the intraparietal sulcus) discriminated between painful and nonpainful stimulation of a rubber hand in participants’ peripersonal hand space. Preferential activation in response to viewing the sharp (painful) stimulus compared with the blunt (nonpainful) stimulus contacting the fingers of the rubber hand (in the absence of actual tactile input to the real hand and with the rubber hand in an anatomically plausible position) was seen in superior parietal (BA5/7) and inferior parietal (BA40) cortices. Other areas showing similar BOLD signal modulation included mid-cingulate and superior-medial frontal lobe, the cerebellum, and the fusiform and superior temporal gyri. On the basis of these results, we propose that the response of the posterior parietal cortex points to its role in the visuospatial encoding of noxious threats, operating alongside other motivational, proprioceptive, and movement-related areas in representing motivationally significant aversive events.

Posterior parietal cortex and noxious threat representation

A primary function of posterior parietal cortex is the integration of visuospatial and somatosensory information to shape an appropriate motor response (for a recent account, see Grefkes and Fink 2005). In the monkey, the inferior and superior posterior parietal areas chiefly receive visual inputs from striate cortex but are also the first regions along the dorsal visual stream to integrate these retinally derived signals with other sensory signals (such as somatosensory and proprioceptive afferents) to form a higher-order representation of visual space (Driver and Mattingley 1998). Rizzolatti (Rizzolatti and Matelli 2003) has proposed a separation of the dorsal-stream parietal areas into two distinct “sub-streams,” ventral and dorsal. In particular, the ventral part of the dorsal stream is comprised of inferior parietal regions (PF and PG) and supports visual representations of space for the purposes of organizing action. These regions have also been associated with action intention (Andersen and Buneo 2002) and are extensively connected with frontal premotor areas (Rizzolatti and Matelli 2003; Shipp et al. 1998).

The inferior parietal area observed in our study is in the region of the human homologue of monkey areas PF and PG, which play just such a role in the organization of action with respect to objects in space. Although posterior parietal processing is mainly insulated from semantic information about objects from the ventral visual stream, studies of human-neglect patients have indicated that inferior parietal cortex is itself involved in implicit visual awareness of objects in the context of movement planning (Marshall and Halligan 1988; Rizzolatti and Berti 1990). We therefore suggest a role for the inferior parietal area in the motivational response to threatening stimuli visually encoded in peripersonal space.

Further to its role in integrating visual responses to objects in peripersonal space, in humans, posterior parietal damage can alter pain sensation, sometimes resulting in hypoalgesia and asymbolia for pain (Berthier et al. 1988; Greenspan and Winfield 1992). Microstimulation in an epilepsy patient has been observed to evoke a painful somatosensory aura (Salanova et al. 1995). So far, however, very little is known about the role of the human posterior parietal cortex in processing visual information about pain, yet there is evidence that visuo-tactile receptive fields in monkey inferior parietal and intraparietal cortex are sensitive to both tactile and visual information about noxious stimuli (Dong et al., 1994). In these areas, the visual receptive field is bound to the space surrounding the tactile receptive field, e.g., the hand or face. In the macaque, one study showed that a proportion of cells in inferior parietal area PF fired both when skin on the face was stimulated with noxious heat and when the monkey viewed a threatening stimulus coming toward or hovering near that part of the skin (Dong et al. 1996). Furthermore, the responses of these cells closely matched the behavioral response curves for a tolerance-escape task the monkeys performed. Similarly, cells in nearby ventral intraparietal area (VIP) have also shown specifically nocifensive properties. Microstimulation here has produced eye, lip, and arm movements comparable to those elicited by an aversive airpuff into the eyes (Cooke and Graziano 2003).

Possible visuospatial network for motivational relevance

Here we postulate a possible network for the visuospatial encoding of aversive stimuli. The overarching feature of this network is the encoding of visuospatial information in body-part-centered terms for the ultimate purpose of organizing effective action away from the aversive stimulus. The main functional components of this network are body-part-centered encoding of the space surrounding the hand (in our study, the interpretation of “rubber-hand space” as peripersonal hand space), discriminating the motivational relevance of objects in that space (here, whether the probe was noxious or innocuous), and elaborating the motivational-affective sensorimotor representation of the stimulus in terms of appropriate motor responses.

The object in hand space becomes incorporated into the body schema in the sense that the space surrounding the rubber hand is interpreted as that surrounding the real hand. Under normal circumstances, proprioceptive and tactile input from one’s real hand would be in register with the visual information about hand position. In this study, we highlighted the visual, not the tactile, component of the representation of hand space by using a rubber hand that occluded the real hand from view. Because vision often dominates touch in cases of multisensory spatial conflict, it is very likely that objects within this surrogate hand space were interpreted by posterior parietal visual areas as being near the real, proprioceptively sensed hand and thus processed in hand-centered terms. Studies in monkeys and humans have shown this peripersonal space around the hand to be dynamic, extending a virtual body envelope around not only the hand but also non-body objects within it, such as tools (Maravita et al. 2001, 2002; for an alternative view, see Holmes et al. 2004), artificial limbs (Botvinick et al. 1998; Graziano et al. 2000; Pavani et al. 2000), and even the adjacent tabletop (Armel et al. 2003).

An element of this body-schema incorporation depends merely on there being an object within the peripersonal hand space. The main effects (Tables 1 and 2) revealed a contralateral parietal opercular cluster (BA40) and an ipsilateral superior parietal area (BA7) that did not differentiate between a hand that was oriented in an anatomically plausible way and one that was rotated 180°. However, other pain-preferring posterior parietal areas were sensitive to the orientation of the hand. This is consistent with monkey studies in which individual neurons’ responses to a plausible artificial limb decreased when the anatomical plausibility was violated or when the object did not resemble a hand (Graziano 1999; Graziano et al. 2000). The parietal opercular area also showed a significant response to viewing the nonpainful tactile stimulus, but both posterior parietal areas showed significantly higher responses to the painful as compared with the nonpainful stimulus (Table 3).

Another such posture-sensitive activation was seen in the cerebellum, which is, among other things, heavily implicated in position sense as well as in nociception (for a recent review, see Saab and Willis 2003). Major afferents to the cerebellum come from inferior parietal cortex and cingulate gyrus. It also receives visual projections from two ventral-stream areas that differentiated between painful and nonpainful stimuli: extrastriate area 19, and the suprasylvian fissure (Saab and Willis 2003). According to Blakemore and colleagues (2001), the cerebellum uses a forward model to compare the predicted consequences of an action to the actual result of an action and updates the prediction accordingly. To do this they suggest that activity in the cerebellum depends on the sensory-specific consequences of movement and signaling the sensory discrepancy between predicted and actual sensory feedback. For example, the cerebellum may be involved in signaling the discrepancy of seeing a painful probe touch the rubber hand, which the participant has incorporated into his/her own body representation, but not feeling any painful tactile sensation resulting from the probe touching the skin.

Relationship with classical motivational-affective networks

Beyond its initial visuospatial representation, the relevance of the object in hand space is evaluated in motivational terms. Our results suggest that posterior parietal regions are capable of discriminating between painful and nonpainful stimuli. Other areas conventionally associated with the motivational and affective evaluation of aversive stimuli were also preferentially active to the painful stimulus. It is interesting that two of these, the putamen and right anterior insula, preferred the implausible orientation. The putamen and anterior insula have often been reported in neuroimaging studies of pain (Bingel et al. 2004; Bowsher et al. 2004; Brooks et al. 2002, 2005), and the insula may have a unique role in empathy for pain (Decety and Jackson 2004; Morrison et al. 2004; Singer et al. 2004). If the representation of aversive events in the anterior insula and putamen is not strictly bound to visuospatial and -tactile coherence, it may not be “tricked” by seeing the rubber hand receiving noxious stimulation in this context.

Alongside the posterior parietal areas, the anterior cingulate activations revealed by the contrast between painful and nonpainful stimuli (Fig. 6) are of particular interest. The anterior cingulate cortex focus seen in this study is well-positioned to correspond to cingulate motor areas (CMAs) as determined by human functional and cytoarchitectonic studies (Vogt and Sikes 2000; Vogt et al. 2003). The CMAs receive dense projections from inferior parietal lobe (Isomura and Takada 2004; Matelli et al. 1991; Strick et al. 1998). The nociceptive function of these mid-cingulate areas is well established (Sewards and Sewards 2002; Vogt and Sikes 2000). The focus seen in this study was in a posterior mid-cingulate region that is thought to play a role in short-latency sensorimotor orienting to painful stimuli, perhaps using spatial information from its parietal inputs (Vogt et al. 2003). In the monkey, CMAs send fibers to premotor and primary motor cortices as well as having direct projections to dorsal horn motorneurons in the spinal cord (Isomura and Takada 2004; Matelli et al. 1991; Paus 2001). The motor areas show moderate somatotopic organization for trunk and distal and proximal limbs, and representations of cutaneous as well as skeletal muscles have been observed here in the macaque (Akazawa et al., 2000).

CMAs in posterior mid-cingulate are considerably more interconnected with rostral mid-cingulate regions, which also fall within the cluster significantly activated for the painful stimulus. Whereas CMAs have predominantly premotor properties, these nearby mid-cingulate areas are also associated with the motivational-affective dimension of pain processing (Price 2000; Rainville et al. 1997; Sewards and Sewards 2002) and response selection (Hoshi et al. 2005; Isomura et al. 2004). In that context, their role is tied to the ability to link events with outcomes, allowing the prediction and avoidance of noxious stimuli. Both CMAs and adjacent rostral mid-cingulate are characterized by dense fast excitatory (N-methyl-d-aspartate) and inhibitory (AMPA) receptor types (Bozkurt et al. 2005). It is therefore possible that the activity observed in the BOLD signal change in this region is due to the facilitation of an appropriate response to a threat in hand space, but by the same token, it may reflect inhibition related to sensorimotor response potentiation.

In summary, we propose that posterior parietal areas play a role in immediate, reactive nocifensive responses. These responses are tied to specific effectors and are coded in an egocentric spatial reference frame. Via direct cortical connections they provide initial information to cingulate motor and frontal premotor areas about the relevance of the stimulus. The mid-cingulate in particular may be involved in elaborating the representation of stimulus valence and is involved in more flexible motor response selection, learning, and regulation. Both mid-cingulate and prefrontal regions use contextual information and past experience to produce appropriate responses (Hadland et al. 2003; MacDonald et al. 2000; Matsumoto et al. 2003). The representation of pain in these areas is more flexible in the sense that they support a “generative” (Haggard 2001) representation of the noxious stimulus; that is, they incorporate factors such as current task constraints, motivational values, and past experience to produce behavioral outcomes that are not predictable from the nature of the stimulus alone. This representation is predictive, labile, and probably less immediately dependent on specific effectors or spatial information (Mesulam 1999).

Role of premotor cortex

Contrary to our initial hypothesis premotor cortex did not significantly change its response to viewing a painful tactile stimulus touching the rubber hand despite the fact that this region was activated consistently in all conditions (see Tables 1 and 2 of main effects). Posterior parietal and frontal premotor areas of the primate brain share dense inter-projections forming well-studied functional circuits for the planning and control of action (e.g., Rizzolatti and Luppino 2001). Despite several methodological differences, there is good correspondence between the sites of premotor cortex activation in the current study and those found by Ehrsson et al. (2004). Ehrsson and colleagues used fMRI to explore the neural correlates of the rubber-hand illusion with respect to the subjective sensation of limb ownership. The illusion was associated with activation along the left inferior precentral sulcus (BA44/6), the posterior bank of which corresponds to ventral premotor area 6 and the anterior bank to the posterior part of area 44. When the rubber hand was both plausibly oriented and synchronously stroked, the response in premotor cortex was superadditive, and correlated positively with subjective ratings about the strength of the illusion.

This region is well suited to the multisensory representation of one’s own body as it is anatomically connected to visual and somatosensory areas in the posterior parietal cortex and to frontal motor areas (Rizzolatti et al. 1998). Cells in the parietal and premotor cortex have been shown to represent both the seen and felt position of the hand in both humans and non-human primates, discharging when the hand is touched or when a visual stimulus is presented near the hand (Lloyd et al. 2003; for a recent review, see Graziano et al. 2004). However, despite its sensitivity to multisensory proprioceptive and tactile input, we did not see any evidence for a differential response to painful and nonpainful tactile probes touching the rubber hand using fMRI. However, the population response of cells in this region may have been too small to detect with fMRI, and it would be premature to say that the premotor cortex does not directly encode the biological relevance of stimuli touching or approaching the hand.

Conclusions

This study provides the first neuroimaging evidence of a role for human posterior parietal cortex in the visuospatial coding of the motivational relevance of events in hand space. We observed a significant increase in activation in this region in response to viewing painful (vs. innocuous) stimulation of the rubber hand (when in an anatomically plausible position) in the absence of tactile stimulation of the real hand. This is consistent with primate research that has shown similar preferential spatial encoding of noxious stimuli in posterior parietal areas, suggesting that homologous regions in the human brain may share similar functions in this regard. In our study, however, premotor areas associated with the subjective feeling of limb ownership did not discriminate between painful and nonpainful stimuli. These results provide compelling hints of an integrative network supporting visuospatial and sensorimotor aspects of aversive events in the primate brain that future research can explore.

FOOTNOTES

  • The costs of publication of this article were defrayed in part by the payment of page charges. The article must therefore be hereby marked “advertisement” in accordance with 18 U.S.C. Section 1734 solely to indicate this fact.

The authors thank N. Holmes, J. Parkinson, and G. di Pellegrino for reading an earlier version of this manuscript and the staff and radiographers at the Walton Centre for Neurology and Neurosurgery, Liverpool, UK.

REFERENCES

AUTHOR NOTES

  • Address for reprint requests and other correspondence: D. Lloyd, School of Psychological Sciences, The University of Manchester, Coupland Bldg., Coupland St., Oxford Rd., Manchester, M13 9PL, UK (E-mail: )