全部EyeLink出版物
以下按年份列出了截至2024年(2025年初)的所有13000多篇同行评审的EyeLink研究出版物。您可以使用视觉搜索、平滑追求、帕金森氏症等关键字搜索出版物库。您还可以搜索单个作者的姓名。按研究领域分组的眼动追踪研究可以在解决方案页面上找到。如果我们错过了任何EyeLink眼动追踪论文,请给我们发电子邮件!
2014 |
Wouter Kruijne; Stefan Van der Stigchel; Martijn Meeter A model of curved saccade trajectories: Spike rate adaptation in the brainstem as the cause of deviation away Journal Article In: Brain and Cognition, vol. 85, no. 1, pp. 259–270, 2014. @article{Kruijne2014, The trajectory of saccades to a target is often affected whenever there is a distractor in the visual field. Distractors can cause a saccade to deviate towards their location or away from it. The oculomotor mechanisms that produce deviation towards distractors have been thoroughly explored in behavioral, neurophysiological and computational studies. The mechanisms underlying deviation away, on the other hand, remain unclear. Behavioral findings suggest a mechanism of spatially focused, top-down inhibition in a saccade map, and deviation away has become a tool to investigate such inhibition. However, this inhibition hypothesis has little neuroanatomical or neurophysiological support, and recent findings go against it. Here, we propose that deviation away results from an unbalanced saccade drive from the brainstem, caused by spike rate adaptation in brainstem long-lead burst neurons. Adaptation to stimulation in the direction of the distractor results in an unbalanced drive away from it. An existing model of the saccade system was extended with this theory. The resulting model simulates a wide range of findings on saccade trajectories, including findings that have classically been interpreted to support inhibition views. Furthermore, the model replicated the effect of saccade latency on deviation away, but predicted this effect would be absent with large (400. ms) distractor-target onset asynchrony. This prediction was confirmed in an experiment, which demonstrates that the theory both explains classical findings on saccade trajectories and predicts new findings. |
James H. Kryklywy; Derek G. V. Mitchell Emotion modulates allocentric but not egocentric stimulus localization: implications for dual visual systems perspectives Journal Article In: Experimental Brain Research, vol. 232, no. 12, pp. 3719–3726, 2014. @article{Kryklywy2014, Considerable evidence suggests that emotional cues influence processing prioritization and neural representations of stimuli. Specifically, within the visual domain, emotion is known to impact ventral stream processes and ventral stream-mediated behaviours; it remains unclear, however, the extent to which emotion impacts dorsal stream processes. In the present study, participants localized a visual target stimulus embedded within a background array utilizing allocentric localization (requiring an object-centred representation of visual space to perform an action) and egocentric localization (requiring purely target-directed actions), which are thought to differentially rely on the ventral versus dorsal visual stream, respectively. Simultaneously, a task-irrelevant negative, positive or neutral sound was presented to produce an emotional context. In line with predictions, we found that during allocentric localization, response accuracy was enhanced in the context of negative compared to either neutral or positive sounds. In contrast, no significant effects of emotion were identified during egocentric localization. These results raise the possibility that negative emotional auditory contexts enhance ventral stream, but not dorsal stream, processing in the visual domain. Furthermore, this study highlights the complexity of emotion-cognition interactions, indicating how emotion can have a differential impact on almost identical overt behaviours that may be governed by distinct neurocognitive systems. |
Anuenue Kukona; Gerry T. M. Altmann; Yuki Kamide Knowing what, where, and when: Event comprehension in language processing Journal Article In: Cognition, vol. 133, no. 1, pp. 25–31, 2014. @article{Kukona2014, We investigated the retrieval of location information, and the deployment of attention to these locations, following (described) event-related location changes. In two visual world experiments, listeners viewed arrays with containers like a bowl, jar, pan, and jug, while hearing sentences like "The boy will pour the sweetcorn from the bowl into the jar, and he will pour the gravy from the pan into the jug. And then, he will taste the sweetcorn". At the discourse-final "sweetcorn", listeners fixated context-relevant "Target" containers most (jar). Crucially, we also observed two forms of competition: listeners fixated containers that were not directly referred to but associated with "sweetcorn" (bowl), and containers that played the same role as Targets (goals of moving events; jug), more than distractors (pan). These results suggest that event-related location changes are encoded across representations that compete for comprehenders' attention, such that listeners retrieve, and fixate, locations that are not referred to in the unfolding language, but related to them via object or role information. |
Richard Kunert; Christoph Scheepers Speed and accuracy of dyslexic versus typical word recognition: An eye-movement investigation Journal Article In: Frontiers in Psychology, vol. 5, pp. 1129, 2014. @article{Kunert2014, Developmental dyslexia is often characterized by a dual deficit in both word recognition accuracy and general processing speed. While previous research into dyslexic word recognition may have suffered from speed-accuracy trade-off, the present study employed a novel eye-tracking task that is less prone to such confounds. Participants (10 dyslexics and 12 controls) were asked to look at real word stimuli, and to ignore simultaneously presented non-word stimuli, while their eye-movements were recorded. Improvements in word recognition accuracy over time were modeled in terms of a continuous non-linear function. The words' rhyme consistency and the non-words' lexicality (unpronounceable, pronounceable, pseudohomophone) were manipulated within-subjects. Speed-related measures derived from the model fits confirmed generally slower processing in dyslexics, and showed a rhyme consistency effect in both dyslexics and controls. In terms of overall error rate, dyslexics (but not controls) performed less accurately on rhyme-inconsistent words, suggesting a representational deficit for such words in dyslexics. Interestingly, neither group showed a pseudohomophone effect in speed or accuracy, which might call the task-independent pervasiveness of this effect into question. The present results illustrate the importance of distinguishing between speed- vs. accuracy-related effects for our understanding of dyslexic word recognition. |
I. Kurki; Miguel P. Eckstein Template changes with perceptual learning are driven by feature informativeness Journal Article In: Journal of Vision, vol. 14, no. 11, pp. 1–18, 2014. @article{Kurki2014, Perceptual learning changes the way the human visual system processes stimulus information. Previous studies have shown that the human brain's weightings of visual information (the perceptual template) become better matched to the optimal weightings. However, the dynamics of the template changes are not well understood. We used the classification image method to investigate whether visual field or stimulus properties govern the dynamics of the changes in the perceptual template. A line orientation discrimination task where highly informative parts were placed in the peripheral visual field was used to test three hypotheses: (1) The template changes are determined by the visual field structure, initially covering stimulus parts closer to the fovea and expanding tward the periphery with learning; (2) the template changes are object centered, starting from the center and expanding toward edges; and (3) the template changes are determined by stimulus information, starting from the most informative parts and expanding to less informative parts. Results show that, initially, the perceptual template contained only the more peripheral, highly informative parts. Learning expanded the template to include less informative parts, resulting in an increase in sampling efficiency. A second experiment interleaved parts with high and low signal-to-noise ratios and showed that template reweighting through learning was restricted to stimulus elements that are spatially contiguous to parts with initial high template weights. The results suggest that the informativeness of features determines how the perceptual template changes with learning. Further, the template expansion is constrained by spatial proximity. |
Chigusa Kurumada; Meredith Brown; Sarah Bibyk; Daniel F. Pontillo; Michael K. Tanenhaus Is it or isn't it: Listeners make rapid use of prosody to infer speaker meanings Journal Article In: Cognition, vol. 133, no. 2, pp. 335–342, 2014. @article{Kurumada2014, A visual world experiment examined the time course for pragmatic inferences derived from visual context and contrastive intonation contours. We used the construction It looks like an X pronounced with either (a) a H* pitch accent on the final noun and a low boundary tone, or (b) a contrastive L + H* pitch accent and a rising boundary tone, a contour that can support contrastive inference (e.g., It LOOKSL+H* like a zebraL-H%...(but it is not)). When the visual display contained a single related set of contrasting pictures (e.g. a zebra vs. a zebralike animal), effects of LOOKSL+H*emerged prior to the processing of phonemic information from the target noun. The results indicate that the prosodic processing is incremental and guided by contextually-supported expectations. Additional analyses ruled out explanations based on context-independent heuristics that might substitute for online computation of contrast. |
MiYoung Kwon; Pinglei Bao; Rachel Millin; Bosco S. Tjan Radial-tangential anisotropy of crowding in the early visual areas Journal Article In: Journal of Neurophysiology, vol. 112, no. 10, pp. 2413–2422, 2014. @article{Kwon2014, Crowding, the inability to recognize an individual object in clutter (Bouma H. Nature 226: 177–178, 1970), is considered a major impediment to object recognition in peripheral vision. Despite its significance, the cortical loci of crowding are not well understood. In particular, the role of the primary visual cortex (V1) remains unclear. Here we utilize a diagnostic feature of crowding to identify the earliest cortical locus of crowding. Controlling for other factors, radially arranged flankers induce more crowding than tangentially arranged ones (Toet A, Levi DM. Vision Res 32: 1349–1357, 1992). We used functional magnetic resonance imaging (fMRI) to measure the change in mean blood oxygenation level-dependent (BOLD) response due to the addition of a middle letter between a pair of radially or tangen- tially arranged flankers. Consistent with the previous finding that crowding is associated with a reduced BOLD response [Millin R, Arman AC, Chung ST, Tjan BS. Cereb Cortex (July 5, 2013). doi:10.1093/cercor/bht159], we found that the BOLD signal evoked by the middle letter depended on the arrangement of the flankers: less BOLD response was associated with adding the middle letter between radially arranged flankers compared with adding it between tangen- tially arranged flankers. This anisotropy in BOLD response was present as early as V1 and remained significant in downstream areas. The effect was observed while subjects' attention was diverted away from the testing stimuli. Contrast detection threshold for the middle letter was unaffected by flanker arrangement, ruling out surround suppression of contrast response as a major factor in the observed BOLD anisotropy. Our findings support the view that V1 contributes to crowding. |
Nayoung Kwon; Patrick Sturt The use of control information in dependency formation: An eye-tracking study Journal Article In: Journal of Memory and Language, vol. 73, no. 1, pp. 59–80, 2014. @article{Kwon2014a, Recent research has shown much evidence that sentence comprehension can be extremely predictive. However, we currently know little about the limits of predictive processing. In the two eye-tracking experiments, we examined whether predictive information in dependency formation is inevitably given priority over a well-known structural preference in syntactic ambiguity resolution. Experiment 1 used sentences including control nouns like order (e.g. After Andrew's order to wash the kids came over to the house). If predictive dependency information is given priority over disambiguation preferences, then readers could immediately interpret the kids as the ones who have been ordered to wash, thus avoiding the garden path at the main verb came. However, garden path effects were found irrespective of control information, although the garden path difficulty was reduced when the lexical control information highlighted the globally correct analysis (as in the above example), relative to when it did not. Experiment 2 replicated these results with adjunct control, where the relevant dependency is obligatory (e.g. After refusing to wash the kids came over to the house). Again, control information did not influence initial disambiguation, but did affect the difficulty of garden path recovery. Overall, the results suggest that there are limitations on the influence of predictive dependency formation on on-line structural disambiguation. |
Wolfgang Jaschinski Pupil size affects measures of eye position in video eye tracking: Implications for recording vergence accuracy Journal Article In: Journal of eye movement research, vol. 9, no. 4, pp. 1–14, 2014. @article{Jaschinski2014, Video eye trackers rely on the position of the pupil centre. However, the pupil centre can shift when the pupil size changes. This pupillary artefact is investigated for binocular vergence accuracy (i.e. fixation disparity) in near vision where the pupil is smaller in the binocular test phase than in the monocular calibration. A regression between recordings of pupil size and fixation disparity allows correcting the pupillary artefact. This corrected fixation disparity appeared to be favourable with respect to reliability and validity, i.e. the correlation of fixation disparity versus heterophoria. The findings provide a quantitative estimation of the pupillary artefact on measured eye position as function of viewing disance and luminance, both for measures of monocular and binocular eye position. |
Pavitra Jayaramachandran; Frank A. Proudlock; Nita Odedra; Irene Gottlob; Rebecca J. McLean A randomized controlled trial comparing soft contact lens and rigid gas-permeable lens wearing in infantile nystagmus Journal Article In: Ophthalmology, vol. 121, no. 9, pp. 1827–1836, 2014. @article{Jayaramachandran2014, Objective: To perform the first randomized controlled trial comparing soft contact lens (SCL) with rigid gas-permeable lens (RGPL) wearing in infantile nystagmus (IN), using spectacle wear as a baseline. Design: Randomized, controlled cross-over trial with an intention-to-treat design. Participants and Controls: A total of 24 participants with IN (12 idiopathic, 12 with albinism). Methods: Participants were randomized into 1 of 2 treatment arms receiving the following sequence of treatments (2-3 weeks for each treatment): (A) spectacles, SCL, RGPL, and spectacle wear; or (B) spectacles, RGPL, SCL, and spectacle wear. Main Outcome Measures: The main outcome measure was mean intensity of nystagmus at the null region viewing at 1.2 m. Secondary outcome measures included the same measure at 0.4 m viewing and across the horizontal meridian (measured over a ±30° range at 3° intervals) for distance and near. The nystagmus foveation characteristics were similarly assessed over ±30° and at the null region at 1.2 m and 0.4 m viewing. Visual outcome measures included best-corrected visual acuity (BCVA) at 4 m and 0.4 m, gaze-dependent visual acuity (GDVA) (i.e., visual acuity when maintaining gaze angles over a ±30° range at 10° intervals) at 4 m, and reading performance at 0.4 m derived from the Radner reading chart. Results: There were no significant differences between SCL and RGPL wearing for any nystagmus characteristics or compared with spectacle wearing. The BCVA, reading acuity, and critical print size were significantly worse for SCL wearing compared with RGPL and baseline spectacle wear (P<0.05), although mean differences were less than 1 logarithm of the minimum angle of resolution (logMAR) line. Conclusions: Nystagmus was not significantly different during SCL and RGPL wearing in IN, and contact lens wearing does not significantly reduce nystagmus compared with baseline spectacle wearing. The wearing of SCL leads to a small but statistically significant deterioration in visual function compared with both RGPL and spectacle wearing at baseline, although mean effect sizes were not clinically relevant. |
Lisa N. Jefferies; Leon Gmeindl; Steven Yantis Attending to illusory differences in object size Journal Article In: Attention, Perception, & Psychophysics, vol. 76, no. 5, pp. 1393–1402, 2014. @article{Jefferies2014, Focused visual attention can be shifted between objects and locations (attentional orienting) or expanded and contracted in spatial extent (attentional focusing). Although orienting and focusing both modulate visual processing, they have been shown to be distinct, independent modes of attentional control. Objects play a central role in visual attention, and it is known that high-level object representations guide attentional orienting. It not known, however, whether attentional focusing is driven by low-level object representations (which code object size in terms of retinotopic extent) or by high-level representations (which code perceived size). We manipulated the perceived size of physically identical objects by using line drawings or photographs that induced the Ponzo illusion, in a task requiring the detection of a target within these objects. The distribution of attention was determined by the perceived size and not by the retinotopic size of an attended object, indicating that attentional focusing is guided by high-level object representations. |
Yu-Cin Jian; Hwa-Wei Ko Investigating the effects of background knowledge on Chinese word processing during text reading: Evidence from eye movements Journal Article In: Journal of Research in Reading, vol. 37, pp. S71–S86, 2014. @article{Jian2014a, This study investigates the effects of background knowledge on Chinese word processing during silent reading by monitoring adult readers' eye movements. Both higher knowledge (physics major) and lower knowledge (nonphysics major) graduate students were given physics texts to read. Higher knowledge readers spent less time rereading and had lower regression rates on unfamiliar physics words and common words in physics texts than did lower knowledge readers; they also had shorter gaze durations and fewer first-pass fixations on familiar physics words than on unfamiliar physics words. For unfamiliar physics words and common words, both groups predominantly fixated first on the beginnings of words when they made multiple fixations on a word and on a left-of-centre location when they fixated only once on a word. These findings suggest that both groups comprise mature readers with strong language concepts. However, differences in background knowledge led to different reading processes at different stages of reading. |
Yu-Cin Jian; Chao-Jung Wu; Jia-Han Su Learners' eye movements during construction of mechanical kinematic representations from static diagrams Journal Article In: Learning and Instruction, vol. 32, pp. 51–62, 2014. @article{Jian2014, We investigated the influence of numbered arrows on construction of mechanical kinematic representations by using static diagrams. Undergraduate participants viewed a two-stage diagram depicting a flushing cistern (with or without numbered arrows) and answered questions about its function, step-by-step. The arrow group demonstrated greater overall accuracy and made fewer errors on the measure of continuous relations than did the non-arrow group. The arrow group also spent more time looking at components relevant to the operational sequence and had longer first-pass fixation times and shorter saccade lengths. The non-arrow group made more saccades between the two diagrams. Analysis of transition probabilities indicated that both groups viewed components according to their continuous relations. The arrow group followed the numbered arrows whereas the unique pathway of the non-arrow group was to compare the two diagrams. These findings indicate that numbered arrows provide perceptual information but also facilitate cognitive processing. |
Zhenlan Jin; Scott N. J. Watamaniuk; Aarlenne Zein Khan; Elena Potapchuk; Stephen J. Heinen Motion integration for ocular pursuit does not hinder perceptual segregation of moving objects Journal Article In: Journal of Neuroscience, vol. 34, no. 17, pp. 5835–5841, 2014. @article{Jin2014, When confronted with a complex moving stimulus, the brain can integrate local element velocities to obtain a single motion signal, or segregate the elements to maintain awareness of their identities. The integrated motion signal can drive smooth-pursuit eye movements (Heinen and Watamaniuk, 1998), whereas the segregated signal guides attentive tracking of individual elements in multiple-object tracking tasks (MOT; Pylyshyn and Storm, 1988). It is evident that these processes can occur simultaneously, because we can effortlessly pursue ambulating creatures while inspecting disjoint moving features, such as arms and legs, but the underlying mechanism is unknown. Here, we provide evidence that separate neural circuits perform the mathematically opposed operations of integration and segregation, by demonstrating with a dual-task paradigm that the two processes do not share attentional resources. Human observers attentively tracked a subset of target elements composing a small MOT stimulus, while pursuing it ocularly as it translated across a computer display. Integration of the multidot stimulus yielded optimal pursuit. Importantly, performing MOT while pursuing the stimulus did not degrade performance on either task compared with when each was performed alone, indicating that they did not share attention. A control experiment showed that pursuit was not driven by integration of only the nontargets, leaving the MOT targets free for segregation. Nor was a predictive strategy used to pursue the stimulus, because sudden changes in its global velocity were accurately followed. The results suggest that separate neural mechanisms can simultaneously segregate and integrate the same motion signals. |
Kevin D. Johnston; Michael J. Koval; Stephen G. Lomber; Stefan Everling Macaque dorsolateral prefrontal cortex does not suppress saccade-related activity in the superior colliculus Journal Article In: Cerebral Cortex, vol. 24, no. 5, pp. 1373–1388, 2014. @article{Johnston2014, Of the many functions ascribed to the dorsolateral prefrontal cortex (DLPFC), the ability to override automatic stimulus-driven behavior is one of the most prominent. This ability has been investigated extensively with the antisaccade task, which requires suppression of saccades toward suddenly appearing visual stimuli. Convergent lines of evidence have supported a model in which the DLPFC suppresses unwanted saccades by inhibiting saccade-related activity in the ipsilateral superior colliculus (SC), a midbrain oculomotor structure. Here, we carried out a direct test of this inhibitory model using unilateral cryogenic deactivation of the DLPFC within the caudal principal sulcus (cPS) and simultaneous single-neuron recording of SC saccade-related neurons in monkeys performing saccades and antisaccades. Contrary to the inhibition model, which predicts that attenuation of inhibition effected by unilateral cPS deactivation should result in activity increases in ipsilateral and decreases in contralateral SC, we observed a delayed onset of saccade-related activity in the ipsilateral SC, and activity increases in the contralateral SC. These effects were mirrored by increased error rates of ipsiversive antisaccades, and reaction times of contraversive saccades. These data challenge the inhibitory model and suggest instead that the primary influence of the DLPFC on the SC is excitatory. |
Afsheen Khan; Sally A. McFadden; Mark R. Harwood; Josh Wallman Salient distractors can induce saccade adaptation Journal Article In: Journal of Ophthalmology, vol. 2014, pp. 1–11, 2014. @article{Khan2014, When saccadic eye movements consistently fail to land on their intended target, saccade accuracy is maintained by gradually adapting the movement size of successive saccades. The proposed error signal for saccade adaptation has been based on the distance between where the eye lands and the visual target (retinal error). We studied whether the error signal could alternatively be based on the distance between the predicted and actual locus of attention after the saccade. Unlike conventional adaptation experiments that surreptitiously displace the target once a saccade is initiated towards it, we instead attempted to draw attention away from the target by briefly presenting salient distractor images on one side of the target after the saccade. To test whether less salient, more predictable distractors would induce less adaptation, we separately used fixed random noise distractors. We found that both visual attention distractors were able to induce a small degree of downward saccade adaptation but significantly more to the more salient distractors. As in conventional adaptation experiments, upward adaptation was less effective and salient distractors did not significantly increase amplitudes. We conclude that the locus of attention after the saccade can act as an error signal for saccade adaptation. |
Roozbeh Kiani; Leah Corthell; Michael N. Shadlen Choice certainty is informed by both evidence and decision time Journal Article In: Neuron, vol. 84, no. 6, pp. 1329–1342, 2014. @article{Kiani2014, "Degree of certainty" refers to the subjective belief, prior to feedback, that a decision is correct. A reliable estimate of certainty is essential for prediction, learning from mistakes, and planning subsequent actions when outcomes are not immediate. It is generally thought that certainty is informed by a neural representation of evidence at the time of a decision. Here we show that certainty is also informed by the time taken to form the decision. Human subjects reported simultaneously their choice and confidence about the direction of a noisy display of moving dots. Certainty was inversely correlated with reaction times and directly correlated with motion strength. Moreover, these correlations were preserved even for error responses, a finding that contradicts existing explanations of certainty based on signal detection theory. We also contrived a stimulus manipulation that led to longer decision times without affecting choice accuracy, thus demonstrating that deliberation time itself informs the estimate of certainty. We suggest that elapsed decision time informs certainty because it serves as a proxy for task difficulty. |
Sunjung Kim; Linda J. Lombardino; Wind Cowles; Lori J. Altmann Investigating graph comprehension in students with dyslexia: An eye tracking study Journal Article In: Research in Developmental Disabilities, vol. 35, no. 7, pp. 1609–1622, 2014. @article{Kim2014, The purpose of this study was to examine graph comprehension in college students with developmental dyslexia. We investigated how graph types (line, vertical bar, and horizontal bar graphs), graphic patterns (single and double graphic patterns), and question types (point locating and comparison questions) differentially affect graph comprehension of students with and without dyslexia. Groups were compared for (1) reaction times for answering comprehension questions based on graphed data and (2) eye gaze times for specific graph subregions (x-axis, y-axis, pattern, legend, question, and answer). Dyslexic readers were significantly slower in their graph comprehension than their peers with group differences becoming more robust with the increasing complexity of graphs and tasks. In addition, dyslexic readers' initial eye gaze viewing times for linguistic subregions (question and answer) and total viewing times for both linguistic (question and answer) and nonlinguistic (pattern) subregions were significantly longer than their control peers' times. In spite of using elementary-level paragraphs for comprehension and simple graph forms, young adults with dyslexia needed more time to process linguistic and nonlinguistic stimuli. These findings are discussed relative to theories proposed to address fundamental processing deficits in individuals with dyslexia. |
Matthew O. Kimble; Mariam Boxwala; Whitney Bean; Kristin Maletsky; Jessica Halper; Kaleigh Spollen; Kevin Fleming The impact of hypervigilance: Evidence for a forward feedback loop Journal Article In: Journal of Anxiety Disorders, vol. 28, no. 2, pp. 241–524, 2014. @article{Kimble2014, A number of prominent theories suggest that hypervigilance and attentional bias play a central role in anxiety disorders and PTSD. It is argued that hypervigilance may focus attention on potential threats and precipitate or maintain a forward feedback loop in which anxiety is increased. While there is considerable data to suggest that attentional bias exists, there is little evidence to suggest that it plays this proposed but critical role. This study investigated how manipulating hypervigilance would impact the forward feedback loop via self-reported anxiety, visual scanning, and pupil size. Seventy-one participants were assigned to either a hypervigilant, pleasant, or control condition while looking at a series of neutral pictures. Those in the hypervigilant condition had significantly more fixations than those in the other two groups. These fixations were more spread out and covered a greater percentage of the ambiguous scene. Pupil size was also significantly larger in the hypervigilant condition relative to the control condition. Thus the study provided support for the role of hypervigilance in increasing visual scanning and arousal even to neutral stimuli and even when there is no change in self-reported anxiety. Implications for the role this may play in perpetuating a forward feedback loop are discussed. |
Kohitij Kar; Bart Krekelberg Transcranial alternating current stimulation attenuates visual motion adaptation Journal Article In: Journal of Neuroscience, vol. 34, no. 21, pp. 7334–7340, 2014. @article{Kar2014, Transcranial alternating current stimulation (tACS) is used in clinical applications and basic neuroscience research. Although its behavioral effects are evident from prior reports, current understanding of the mechanisms that underlie these effects is limited. We used motion perception, a percept with relatively well known properties and underlying neural mechanisms to investigate tACS mechanisms. Healthy human volunteers showed a surprising improvement in motion sensitivity when visual stimuli were paired with 10 Hz tACS. In addition, tACS reduced the motion-after effect, and this reduction was correlated with the improvement in motion sensitivity. Electrical stimulation had no consistent effect when applied before presenting a visual stimulus or during recovery from motion adaptation. Together, these findings suggest that perceptual effects of tACS result from an attenuation of adaptation. Important consequences for the practical use of tACS follow from our work. First, because this mechanism interferes only with adaptation, this suggests that tACS can be targeted at subsets of neurons (by adapting them), even when the applied currents spread widely throughout the brain. Second, by interfering with adaptation, this mechanism provides a means by which electrical stimulation can generate behavioral effects that outlast the stimulation. |
Hossein Karimi; Kumiko Fukumura; Fernanda Ferreira; Martin J. Pickering The effect of noun phrase length on the form of referring expressions Journal Article In: Memory & Cognition, vol. 42, no. 6, pp. 993–1009, 2014. @article{Karimi2014, The length of a noun phrase has been shown to influence choices such as syntactic role assignment (e.g., whether the noun phrase is realized as the subject or the object). But does length also affect the choice between different forms of referring expressions? Three experiments investigated the effect of antecedent length on the choice between pronouns (e.g., he) and repeated nouns (e.g., the actor) using a sentence-continuation paradigm. Experiments 1 and 2 found an effect of antecedent length on written continuations: Participants used more pronouns (relative to repeated nouns) when the antecedent was longer than when it was shorter. Experiment 3 used a spoken continuation task and replicated the effect of antecedent length on the choice of referring expressions. Taken together, the results suggest that longer antecedents increase the likelihood of pronominal reference. The results support theories arguing that length enhances the accessibility of the associated entity through richer semantic encoding. |
Koji Kashihara; Kazuo Okanoya; Nobuyuki Kawai Emotional attention modulates microsaccadic rate and direction Journal Article In: Psychological Research, vol. 78, no. 2, pp. 166–179, 2014. @article{Kashihara2014, Involuntary microsaccades and voluntary saccades reflect human brain activities during attention and cognitive tasks. Our eye movements can also betray our emotional state. However, the effects of attention to emotion on microsaccadic activity remain unknown. The present study was conducted in healthy volunteers to investigate the effects of devoting attention to exogenous emotional stimuli on microsaccadic response, with change in pupil size as an index of sympathetic nervous system activity. Event-related responses to unpleasant images significantly inhibited the rate of microsaccade appearance and altered pupil size (Experiment 1). Additionally, microsaccadic responses to covert orienting of attention to emotional stimuli appeared significantly in the anti-direction to a target, with a fast reaction time (Experiment 2). Therefore, we concluded that attentional shifts induced by exogenous emotional stimuli can modulate microsaccadic activities. Future studies of the interaction between miniature eye movements and emotion may be beneficial in the assessment of pathophysiological responses in mental disorders. |
Ioanna Katidioti; Jelmer P. Borst; Niels A. Taatgen What happens when we switch tasks: Pupil Dilation in Multitasking Journal Article In: Journal of Experimental Psychology: Applied, vol. 20, no. 4, pp. 380–396, 2014. @article{Katidioti2014, Interruption studies typically focus on external interruptions, even though self-interruptions occur at least as often in real work environments. In this article, we therefore contrast external interruptions with self-interruptions. Three multitasking experiments were conducted, in which we examined changes in pupil size when participants switched from a primary to a secondary task. Results showed an increase in pupil dilation several seconds before a self-interruption, which we could attribute to the decision to switch. This indicates that the decision takes a relatively large amount of time. This was supported by the fact that in Experiment 2, participants were significantly slower on the self-interruption blocks than on the external interruption blocks. These findings suggest that the decision to switch is costly, but may also be open for modification through appropriate training. In addition, we propose that if one must switch tasks, it can be more efficient to implement a forced switch after the completion of a subtask instead of leaving the decision to the user. |
Lisandro N. Kaunitz; Juan E. Kamienkowski; Alexander Varatharajah; Mariano Sigman; Rodrigo Quian Quiroga; Matias J. Ison Looking for a face in the crowd: Fixation-related potentials in an eye-movement visual search task Journal Article In: NeuroImage, vol. 89, pp. 297–305, 2014. @article{Kaunitz2014, Despite the compelling contribution of the study of event related potentials (ERPs) and eye movements to cognitive neuroscience, these two approaches have largely evolved independently. We designed an eye-movement visual search paradigm that allowed us to concurrently record EEG and eye movements while subjects were asked to find a hidden target face in a crowded scene with distractor faces. Fixation event-related potentials (fERPs) to target and distractor stimuli showed the emergence of robust sensory components associated with the perception of stimuli and cognitive components associated with the detection of target faces. We compared those components with the ones obtained in a control task at fixation: qualitative similarities as well as differences in terms of scalp topography and latency emerged between the two. By using single trial analyses, fixations to target and distractors could be decoded from the EEG signals above chance level in 11 out of 12 subjects. Our results show that EEG signatures related to cognitive behavior develop across spatially unconstrained exploration of natural scenes and provide a first step towards understanding the mechanisms of target detection during natural search. |
Kerry Kawakami; Amanda Williams; David M. Sidhu; Becky L. Choma; Rosa Rodriguez-Bailón; Elena Cañadas; Derek Chung; Kurt Hugenberg An eye for the I: Preferential attention to the eyes of ingroup members. Journal Article In: Journal of Personality and Social Psychology, vol. 107, no. 1, pp. 1–20, 2014. @article{Kawakami2014, Human faces, and more specifically the eyes, play a crucial role in social and nonverbal communication because they signal valuable information about others. It is therefore surprising that few studies have investigated the impact of intergroup contexts and motivations on attention to the eyes of ingroup and outgroup members. Four experiments investigated differences in eye gaze to racial and novel ingroups using eye tracker technology. Whereas Studies 1 and 3 demonstrated that White participants attended more to the eyes of White compared to Black targets, Study 2 showed a similar pattern of attention to the eyes of novel ingroup and outgroup faces. Studies 3 and 4 also provided new evidence that eye gaze is flexible and can be meaningfully influenced by current motivations. Specifically, instructions to individuate specific social categories increased attention to the eyes of target group members. Furthermore, the latter experiments demonstrated that preferential attention to the eyes of ingroup members predicted important intergroup biases such as recognition of ingroup over outgroup faces (i.e., the own-race bias; Study 3) and willingness to interact with outgroup members (Study 4). The implication of these findings for general theorizing on face perception, individuation processes, and intergroup relations are discussed. |
Barrie P. Klein; Ben M. Harvey; Serge O. Dumoulin Attraction of position preference by spatial attention throughout human visual cortex Journal Article In: Neuron, vol. 84, no. 1, pp. 227–237, 2014. @article{Klein2014, Voluntary spatial attention concentrates neural resources at the attended location. Here, we examined the effects of spatial attention on spatial position selectivity in humans. We measured population receptive fields (pRFs) using high-field functional MRI (fMRI) (7T) while subjects performed an attention-demanding task at different locations. We show that spatial attention attracts pRF preferred positions across the entire visual field, not just at the attended location. This global change in pRF preferred positions systematically increases up the visual hierarchy. We model these pRF preferred position changes as an interaction between two components: an attention field and a pRF without the influence of attention. This computational model suggests that increasing effects of attention up the hierarchy result primarily from differences in pRF size and that the attention field is similar across the visual hierarchy. A similar attention field suggests that spatial attention transforms different neural response selectivities throughout the visual hierarchy in a similar manner. |
Elise Klein; S. Huber; Hans-Christoph Nuerk; Korbinian Moeller Operational momentum affects eye fixation behaviour Journal Article In: Quarterly Journal of Experimental Psychology, vol. 67, no. 8, pp. 1614–1625, 2014. @article{Klein2014a, The operational momentum effect (OM) indicates an association of mental addition with a rightward spatial bias, whereas subtraction is associated with a leftward bias. To evaluate the assumed attentional origin of the OM effect, we evaluated not only participants' relative estimation error in a task requiring them to locate addition and subtraction results on a given number line but also their eye-fixation behaviour. Furthermore, to investigate the situatedness of spatial-numerical associations, the orientation of the number line (left-to-right vs. right-to left) was manipulated. OM biases in participants' explicit number line estimations and more implicit eye-fixation behaviour are integrated into a two-process hypothesis of the OM effect suggesting a first rough spatial anticipation followed by an evaluation/correction process. This account not only is capable of accounting for the results observed for participants' relative estimation error but is also corroborated by the eye-fixation results. Importantly, the fact that all effects were found independent of the orientation of the number line indicates that spatial-numerical associations such as the OM effect may not be hard-wired associations of spatial and numerical representations but rather reflect influences of situatedness on numerical cognition. |
Nadine Kloth; Susannah E. Shields; Gillian Rhodes On the other side of the fence: Effects of social categorization and spatial grouping on memory and attention for own-race and other-race faces Journal Article In: PLoS ONE, vol. 9, no. 9, pp. e105979, 2014. @article{Kloth2014, The term ‘‘own-race bias'' refers to the phenomenon that humans are typically better at recognizing faces from their own than a different race. The perceptual expertise account assumes that our face perception system has adapted to the faces we are typically exposed to, equipping it poorly for the processing of other-race faces. Sociocognitive theories assume that other-race faces are initially categorized as out-group, decreasing motivation to individuate them. Supporting sociocognitive accounts, a recent study has reported improved recognition for other-race faces when these were categorized as belonging to the participants' in-group on a second social dimension, i.e., their university affiliation. Faces were studied in groups, containing both own-race and other-race faces, half of each labeled as in-group and out-group, respectively. When study faces were spatially grouped by race, participants showed a clear own-race bias. When faces were grouped by university affiliation, recognition of other-race faces from the social in-group was indistinguishable from own- race face recognition. The present study aimed at extending this singular finding to other races of faces and participants. Forty Asian and 40 European Australian participants studied Asian and European faces for a recognition test. Faces were presented in groups, containing an equal number of own-university and other-university Asian and European faces. Between participants, faces were grouped either according to race or university affiliation. Eye tracking was used to study the distribution of spatial attention to individual faces in the display. The race of the study faces significantly affected participants' memory, with better recognition of own-race than other-race faces. However, memory was unaffected by the university affiliation of the faces and by the criterion for their spatial grouping on the display. Eye tracking revealed strong looking biases towards both own-race and own-university faces. Results are discussed in light of the theoretical accounts of the own-race bias. |
Zuzanna Klyszejko; Masih Rahmati; Clayton E. Curtis Attentional priority determines working memory precision Journal Article In: Vision Research, vol. 105, pp. 70–76, 2014. @article{Klyszejko2014, Visual working memory is a system used to hold information actively in mind for a limited time. The number of items and the precision with which we can store information has limits that define its capacity. How much control do we have over the precision with which we store information when faced with these severe capacity limitations? Here, we tested the hypothesis that rank-ordered attentional priority determines the precision of multiple working memory representations. We conducted two psychophysical experiments that manipulated the priority of multiple items in a two-alternative forced choice task (2AFC) with distance discrimination. In Experiment 1, we varied the probabilities with which memorized items were likely to be tested. To generalize the effects of priority beyond simple cueing, in Experiment 2, we manipulated priority by varying monetary incentives contingent upon successful memory for items tested. Moreover, we illustrate our hypothesis using a simple model that distributed attentional resources across items with rank-ordered priorities. Indeed, we found evidence in both experiments that priority affects the precision of working memory in a monotonic fashion. Our results demonstrate that representations of priority may provide a mechanism by which resources can be allocated to increase the precision with which we encode and briefly store information. |
Pia Knoeferle Conjunction meaning can modulate parallelism facilitation: Eye-tracking evidence from German clausal coordination Journal Article In: Journal of Memory and Language, vol. 75, pp. 140–158, 2014. @article{Knoeferle2014, In and-coordinated clauses, the second conjunct elicits faster reading times when it parallels (vs. does not parallel) the first in constituent order. This paper examined whether such parallelism facilitation results from simple constituent order priming from the first to the second clause, or whether it can be modulated through the linguistic context (the conjunction and clausal relations). Three eye-tracking experiments on German assessed this issue by manipulating conjunction meaning and type within subjects (resemblance: 'and' vs. adversative: 'but' or 'while'; coordinating: 'and' and 'but'; subordinating: 'while'), and by varying the clausal relations between experiments. Clausal parallelism facilitation was reduced when syntactic dependence of the clauses from a superordinate verb reinforced their coherence, and semantic expectations for 'but' and 'while' were violated through the parallel constituent order and thematic role relations of noun phrases. By contrast, it was not reduced when the same expectations were satisfied through other sentence constituents (temporally contrastive adverbs) and when the coordination involved matrix clauses. The contextual modulation of parallelism facilitation rules out simple priming as the only underlying mechanism. The observed facilitation rather reflects compositional processing of the coordinands and the conjunction in the linguistic context. |
Kathryn Koehler; Fei Guo; Sheng Zhang; Miguel P. Eckstein What do saliency models predict? Journal Article In: Journal of Vision, vol. 14, no. 3, pp. 1–27, 2014. @article{Koehler2014, Saliency models have been frequently used to predict eye movements made during image viewing without a specified task (free viewing). Use of a single image set to systematically compare free viewing to other tasks has never been performed. We investigated the effect of task differences on the ability of three models of saliency to predict the performance of humans viewing a novel database of 800 natural images. We introduced a novel task where 100 observers made explicit perceptual judgments about the most salient image region. Other groups of observers performed a free viewing task, saliency search task, or cued object search task. Behavior on the popular free viewing task was not best predicted by standard saliency models. Instead, the models most accurately predicted the explicit saliency selections and eye movements made while performing saliency judgments. Observers' fixations varied similarly across images for the saliency and free viewing tasks, suggesting that these two tasks are related. The variability of observers' eye movements was modulated by the task (lowest for the object search task and greatest for the free viewing and saliency search tasks) as well as the clutter content of the images. Eye movement variability in saliency search and free viewing might be also limited by inherent variation of what observers consider salient. Our results contribute to understanding the tasks and behavioral measures for which saliency models are best suited as predictors of human behavior, the relationship across various perceptual tasks, and the factors contributing to observer variability in fixational eye movements. |
Thorsten Kolling; Gabriella Óturai; Monika Knopf Is selective attention the basis for selective imitation in infants: An eye-tracking study of deferred imitation with 12-month-olds. Journal Article In: Journal of Experimental Child Psychology, vol. 124, no. 1, pp. 18–35, 2014. @article{Kolling2014, Infants and children do not blindly copy every action they observe during imitation tasks. Research demonstrated that infants are efficient selective imitators. The impact of selective perceptual processes (selective attention) for selective deferred imitation, however, is still poorly described. The current study, therefore, analyzed 12-month-old infants' looking behavior during demonstration of two types of target actions: arbitrary versus functional actions. A fully automated remote eye tracker was used to assess infants' looking behavior during action demonstration. After a 30-min delay, infants' deferred imitation performance was assessed. Next to replicating a memory effect, results demonstrate that infants do imitate significantly more functional actions than arbitrary actions (functionality effect). Eye-tracking data show that whereas infants do not fixate significantly longer on functional actions than on arbitrary actions, amount of fixations and amount of saccades differ between functional and arbitrary actions, indicating different encoding mechanisms. In addition, item-level findings differ from overall findings, indicating that perceptual and conceptual item features influence looking behavior. Looking behavior on both the overall and item levels, however, does not relate to deferred imitation performance. Taken together, the findings demonstrate that, on the one hand, selective imitation is not explainable merely by selective attention processes. On the other hand, notwithstanding this reasoning, attention processes on the item level are important for encoding processes during target action demonstration. Limitations and future studies are discussed. |
Oleg V. Komogortsev; Corey D. Holland; Alex Karpov; Larry R. Price Biometrics via oculomotor plant characteristics: Impact of parameters in oculomotor plant model Journal Article In: ACM Transactions on Applied Perception, vol. 11, no. 4, pp. 1–17, 2014. @article{Komogortsev2014, This article proposes and evaluates a novel biometric approach utilizing the internal, nonvisible, anatomical structure of the human eye. The proposed method estimates the anatomical properties of the human oculomotor plant from the measurable properties of human eye movements, utilizing a two-dimensional linear homeomorphic model of the oculomotor plant. The derived properties are evaluated within a biometric framework to determine their efficacy in both verification and identification scenarios. The results suggest that the physical properties derived from the oculomotor plant model are capable of achieving 20.3% equal error rate and 65.7% rank-1 identification rate on high-resolution equipment involving 32 subjects, with biometric samples taken over four recording sessions; or 22.2% equal error rate and 12.6% rank-1 identification rate on low-resolution equipment involving 172 subjects, with biometric samples taken over two recording sessions. |
Agnieszka E. Konopka; Antje S. Meyer Priming sentence planning Journal Article In: Cognitive Psychology, vol. 73, pp. 1–40, 2014. @article{Konopka2014, Sentence production requires mapping preverbal messages onto linguistic structures. Because sentences are normally built incrementally, the information encoded in a sentence-initial increment is critical for explaining how the mapping process starts and for predicting its timecourse. Two experiments tested whether and when speakers prioritize encoding of different types of information at the outset of formulation by comparing production of descriptions of transitive events (e.g., A dog is chasing the mailman) that differed on two dimensions: the ease of naming individual characters and the ease of apprehending the event gist (i.e., encoding the relational structure of the event). To additionally manipulate ease of encoding, speakers described the target events after receiving lexical primes (facilitating naming; Experiment 1) or structural primes (facilitating generation of a linguistic structure; Experiment 2). Both properties of the pictured events and both types of primes influenced the form of target descriptions and the timecourse of formulation: character-specific variables increased the probability of speakers encoding one character with priority at the outset of formulation, while the ease of encoding event gist and of generating a syntactic structure increased the likelihood of early encoding of information about both characters. The results show that formulation is flexible and highlight some of the conditions under which speakers might employ different planning strategies. |
Christof Körner; Verena Braunstein; Matthias Stangl; Alois Schlögl; Christa Neuper; Anja Ischebeck In: Psychophysiology, vol. 51, no. 4, pp. 385–395, 2014. @article{Koerner2014, To search for a target in a complex environment is an everyday behavior that ends with finding the target. When we search for two identical targets, however, we must continue the search after finding the first target and memorize its location. We used fixation-related potentials to investigate the neural correlates of different stages of the search, that is, before and after finding the first target. Having found the first target influenced subsequent distractor processing. Compared to distractor fixations before the first target fixation, a negative shift was observed for three subsequent distractor fixations. These results suggest that processing a target in continued search modulates the brain's response, either transiently by reflecting temporary working memory processes or permanently by reflecting working memory retention. |
Christof Körner; Margit Höfler; Barbara Tröbinger; Iain D. Gilchrist Eye movements indicate the temporal organisation of information processing in graph comprehension Journal Article In: Applied Cognitive Psychology, vol. 28, no. 3, pp. 360–373, 2014. @article{Koerner2014a, Hierarchical graphs (e.g. file system browsers and preference trees) represent objects (e.g. files and folders) as graph nodes and relations between them (e.g. sub-folder relations) as lines. We investigated the temporal organisation of two processes that are necessary for comprehending such graphs—search for the graph nodes and reasoning about their relation. We tracked eye movements to change graphs while participants interpreted them. In Experiment 1, we masked the graph at a time when search processes had finished but reasoning was hypothetically ongoing. We observed a dramatic deterioration in comprehension compared with unmasked graphs. In Experiment 2, we changed the relation between critical graph nodes after search for them had finished, unbeknownst to participants. Participants mostly based their response on the graph as presented after the change. These results suggest that comprehension processes are organised in a sequential manner, an observation that can potentially be applied to the interactive presentation of graphs. |
Yoshito Kosai; Yasmine El-shamayleh; Amber M. Fyall; Anitha Pasupathy The role of visual area V4 in the discrimination of partially occluded shapes Journal Article In: Journal of Neuroscience, vol. 34, no. 25, pp. 8570–8584, 2014. @article{Kosai2014, The primate brain successfully recognizes objects, even when they are partially occluded. To begin to elucidate the neural substrates of this perceptual capacity, we measured the responses of shape-selective neurons in visual area V4 while monkeys discriminated pairs of shapes under varying degrees of occlusion. We found that neuronal shape selectivity always decreased with increasing occlusion level, with some neurons being notably more robust to occlusion than others. The responses of neurons that maintained their selectivity across a wider range of occlusion levels were often sufficiently sensitive to support behavioral performance. Many of these same neurons were distinctively selective for the curvature of local boundary features and their shape tuning was well fit by a model of boundary curvature (curvature-tuned neurons). A significant subset of V4 neurons also signaled the animal's upcoming behavioral choices; these decision signals had short onset latencies that emerged progressively later for higher occlusion levels. The time course of the decision signals in V4 paralleled that of shape selectivity in curvature-tuned neurons: shape selectivity in curvature-tuned neurons, but not others, emerged earlier than the decision signals. These findings provide evidence for the involvement of contour-based mechanisms in the segmentation and recognition of partially occluded objects, consistent with psychophysical theory. Furthermore, they suggest that area V4 participates in the representation of the relevant sensory signals and the generation of decision signals underlying discrimination. |
Anna A. Kosovicheva; Benjamin A. Wolfe; David Whitney Visual motion shifts saccade targets Journal Article In: Attention, Perception, & Psychophysics, vol. 76, no. 6, pp. 1778–1788, 2014. @article{Kosovicheva2014, Saccades are made thousands of times a day and are the principal means of localizing objects in our environment. However, the saccade system faces the challenge of accurately localizing objects as they are constantly moving relative to the eye and head. Any delays in processing could cause errors in saccadic localization. To compensate for these delays, the saccade system might use one or more sources of information to predict future target locations, including changes in position of the object over time, or its motion. Another possibility is that motion influences the represented position of the object for saccadic targeting, without requiring an actual change in target position. We tested whether the saccade system can use motion-induced position shifts to update the represented spatial location of a saccade target, by using static drifting Gabor patches with either a soft or a hard aperture as saccade targets. In both conditions, the aperture always remained at a fixed retinal location. The soft aperture Gabor patch resulted in an illusory position shift, whereas the hard aperture stimulus maintained the motion signals but resulted in a smaller illusory position shift. Thus, motion energy and target location were equated, but a position shift was generated in only one condition. We measured saccadic localization of these targets and found that saccades were indeed shifted, but only with a soft-aperture Gabor patch. Our results suggest that motion shifts the programmed locations of saccade targets, and this remapped location guides saccadic localization. |
Christopher K. Kovach; Matthew J. Sutterer; Sara N. Rushia; Adrianna Teriakidis; Rick L. Jenison Two systems drive attention to rewards Journal Article In: Frontiers in Psychology, vol. 5, pp. 46, 2014. @article{Kovach2014, How options are framed can dramatically influence choice preference. While salience of information plays a central role in this effect, precisely how it is mediated by attentional processes remains unknown. Current models assume a simple relationship between attention and choice, according to which preference should be uniformly biased towards the attended item over the whole time-course of a decision between similarly valued items. To test this prediction we considered how framing alters the orienting of gaze during a simple choice between two options, using eye movements as a sensitive online measure of attention. In one condition participants selected the less preferred item to discard and in the other, the more preferred item to keep. We found that gaze gravitates towards the item ultimately selected, but did not observe the effect to be uniform over time. Instead, we found evidence for distinct early and late processes that guide attention according to preference in the first case and task demands in the second. We conclude that multiple time-dependent processes govern attention during choice, and that these may contribute to framing effects in different ways. |
Michael J. Koval; R. Matthew Hutchison; Stephen G. Lomber; Stefan Everling Effects of unilateral deactivations of dorsolateral prefrontal cortex and anterior cingulate cortex on saccadic eye movements Journal Article In: Journal of Neurophysiology, vol. 111, no. 4, pp. 787–803, 2014. @article{Koval2014, The dorsolateral prefrontal cortex (dlPFC) and anterior cingulate cortex (ACC) have both been implicated in the cognitive control of saccadic eye movements by single neuron recording studies in nonhuman primates and functional imaging studies in humans, but their relative roles remain unclear. Here, we reversibly deactivated either dlPFC or ACC subregions in macaque monkeys while the animals performed randomly interleaved pro- and antisaccades. In addition, we explored the whole-brain functional connectivity of these two regions by applying a seed-based resting-state functional MRI analysis in a separate cohort of monkeys. We found that unilateral dlPFC deactivation had stronger behavioral effects on saccades than unilateral ACC deactivation, and that the dlPFC displayed stronger functional connectivity with frontoparietal areas than the ACC. We suggest that the dlPFC plays a more prominent role in the preparation of pro- and antisaccades than the ACC. |
Eileen Kowler; Cordelia D. Aitkin; Nicholas M. Ross; Elio M. Santos; Min Zhao Davida Teller Award Lecture 2013: The importance of prediction and anticipation in the control of smooth pursuit eye movements Journal Article In: Journal of Vision, vol. 14, no. 5, pp. 1–16, 2014. @article{Kowler2014, The ability of smooth pursuit eye movements to anticipate the future motion of targets has been known since the pioneering work of Dodge, Travis, and Fox (1930) and Westheimer (1954). This article reviews aspects of anticipatory smooth eye movements, focusing on the roles of the different internal or external cues that initiate anticipatory pursuit.We present new results showing that the anticipatory smooth eye movements evoked by different cues differ substantially, even when the cues are equivalent in the information conveyed about the direction of future target motion. Cues that convey an easily interpretable visualization of the motion path produce faster anticipatory smooth eye movements than the other cues tested, including symbols associated arbitrarily with the path, and the same target motion tested repeatedly over a block of trials. The differences among the cues may be understood within a common predictive framework in which the cues differ in the level of subjective certainty they provide about the future path. Pursuit may be driven by a combined signal in which immediate sensory motion, and the predictions about future motion generated by sets of cues, are weighted according to their respective levels of certainty. Anticipatory smooth eye movements, an overt indicator of expectations and predictions, may not be operating in isolation, but may be part of a global process in which the brain analyzes available cues, formulates predictions, and uses them to control perceptual, motor, and cognitive processes. |
Jens Kremkow; Jianzhong Jin; Stanley J. Komban; Yushi Wang; Reza Lashgari; Xiaobing Li; Michael Jansen; Qasim Zaidi; Jose-Manuel Alonso Neuronal nonlinearity explains greater visual spatial resolution for darks than lights Journal Article In: Proceedings of the National Academy of Sciences, vol. 111, no. 8, pp. 3170–3175, 2014. @article{Kremkow2014, Astronomers and physicists noticed centuries ago that visual spatial resolution is higher for dark than light stimuli, but the neuronal mechanisms for this perceptual asymmetry remain unknown. Here we demonstrate that the asymmetry is caused by a neuronal nonlinearity in the early visual pathway. We show that neurons driven by darks (OFF neurons) increase their responses roughly linearly with luminance decrements, independent of the background luminance. However, neurons driven by lights (ON neurons) saturate their responses with small increases in luminance and need bright backgrounds to approach the linearity of OFF neurons. We show that, as a consequence of this difference in linearity, receptive fields are larger in ON than OFF thalamic neurons, and cortical neurons are more strongly driven by darks than lights at low spatial frequencies. This ON/OFF asymmetry in linearity could be demonstrated in the visual cortex of cats, monkeys, and humans and in the cat visual thalamus. Furthermore, in the cat visual thalamus, we show that the neuronal nonlinearity is present at the ON receptive field center of ON-center neurons and ON receptive field surround of OFF-center neurons, suggesting an origin at the level of the photoreceptor. These results demonstrate a fundamental difference in visual processing between ON and OFF channels and reveal a competitive advantage for OFF neurons over ON neurons at low spatial frequencies, which could be important during cortical development when retinal images are blurred by immature optics in infant eyes. |
Joanne Ingram; Christopher J. Hand; Linda M. Moxey Processing inferences drawn from the logically equivalent frames half full and half empty Journal Article In: Journal of Cognitive Psychology, vol. 26, no. 7, pp. 799–817, 2014. @article{Ingram2014, Choice-based experiments indicate that readers draw sophisticated inferences from logically equivalent frames. Readers may infer that a glass was previously full if described as currently half empty, and previously empty if described as currently half full. The information leakage framework suggests these inferences are made because information about a previous state is leaked from speaker' s choice of frame. We examine if similar inferences are made during reading in two eye-tracking experiments. In Experiment 1, participants read a passage where a character describes a glass as currently half full or half empty before making a statement about the previous volume. We hypothesised that participants would infer that the glass was previously empty or full, respectively. Results suggest processing a previous volume of full is simpler regardless of the frame provided. In Experiment 2, materials were constructed to ensure inferences were based on participants' beliefs as opposed to characters'. Results support the information leakage framework; previous volumes of full and empty were processed more easily after current volumes of half empty and half full, respectively. We suggest that processing discrepancies between the two experiments are driven by word-related factors (e.g., markedness) or by participants' integration of characters' expectations. |
Albrecht W. Inhoff; Ralph Radach Parafoveal preview benefits during silent and oral reading: Testing the parafoveal information extraction hypothesis Journal Article In: Visual Cognition, vol. 22, no. 3-4, pp. 354–376, 2014. @article{Inhoff2014, The preview of a parafoveally visible word conveys benefits when it is subsequently fixated. The current study examined whether these benefits are determined by the effectiveness of parafoveal information extraction, as implied by current models of eye movement control during reading, or by the effectiveness with which extracted information is integrated when a previewed word is fixated. For this, the boundary technique was used to manipulate the extent to which parafoveal information could be extracted, and text was read silently or orally. Consistent with prior work, a parafoveal target word preview conveyed fewer benefits when less parafoveal information could be extracted, target viewing durations were longer during oral than during silent reading, and the two factors interacted in the target fixation data, with smaller preview benefits during oral than during silent reading. Survival analyses indicated that this occurred because parafoveal information use occurred at later point in time during oral reading. Diminished opportunity for parafoveal information extraction also diminished target skipping rate, and it resulted in smaller saccades to target words, but these effects were not influenced by reading mode. Parafoveally extracted information was thus used less effectively during oral reading only when it involved the integration of parafoveally extracted information during subsequent target viewing. The dissociation of extraction from integration challenges current models of eye movement control.$backslash$nThe preview of a parafoveally visible word conveys benefits when it is subsequently fixated. The current study examined whether these benefits are determined by the effectiveness of parafoveal information extraction, as implied by current models of eye movement control during reading, or by the effectiveness with which extracted information is integrated when a previewed word is fixated. For this, the boundary technique was used to manipulate the extent to which parafoveal information could be extracted, and text was read silently or orally. Consistent with prior work, a parafoveal target word preview conveyed fewer benefits when less parafoveal information could be extracted, target viewing durations were longer during oral than during silent reading, and the two factors interacted in the target fixation data, with smaller preview benefits during oral than during silent reading. Survival analyses indicated that this occurred because parafoveal information use occurred at later point in time during oral reading. Diminished opportunity for parafoveal information extraction also diminished target skipping rate, and it resulted in smaller saccades to target words, but these effects were not influenced by reading mode. Parafoveally extracted information was thus used less effectively during oral reading only when it involved the integration of parafoveally extracted information during subsequent target viewing. The dissociation of extraction from integration challenges current models of eye movement control. |
Kazuya Inoue; Yuji Takeda The properties of object representations constructed during visual search in natural scenes Journal Article In: Visual Cognition, vol. 22, no. 9-10, pp. 1135–1153, 2014. @article{Inoue2014, To investigate properties of object representations constructed during a visual search task, we manipulated the proportion of trials/task within a block: In a search-frequent block, 80% of trials were search tasks; remaining trials presented a memory task; in a memory-frequent block, this proportion was reversed. In the search task, participants searched for a toy car (Experiments 1 and 2) or a T-shape object (Experiment 3). In the memory task, participants had to memorize objects in a scene. Memory performance was worse in the search-frequent block than in the memory-frequent block in Experiments 1 and 3, but not in Experiment 2 (token change in Experiment 1; type change in Experiments 2 and 3). Experiment 4 demonstrated that lower performance in the search-frequent block was not due to eye-movement behaviour. Results suggest that object representations constructed during visual search are different from those constructed during memorization and they are modulated by type of target. |
David E. Irwin Short-term memory across eye blinks Journal Article In: Memory, vol. 22, no. 8, pp. 898–906, 2014. @article{Irwin2014, The effect of eye blinks on short-term memory was examined in two experiments. On each trial, participants viewed an initial display of coloured, oriented lines, then after a retention interval they viewed a test display that was either identical or different by one feature. Participants kept their eyes open throughout the retention interval on some blocks of trials, whereas on others they made a single eye blink. Accuracy was measured as a function of the number of items in the display to determine the capacity of short-term memory on blink and no-blink trials. In separate blocks of trials participants were instructed to remember colour only, orientation only, or both colour and orientation. Eye blinks reduced short-term memory capacity by approximately 0.6-0.8 items for both feature and conjunction stimuli. A third, control, experiment showed that a button press during the retention interval had no effect on short-term memory capacity, indicating that the effect of an eye blink was not due to general motoric dual-task interference. Eye blinks might instead reduce short-term memory capacity by interfering with attention-based rehearsal processes. |
David E. Irwin; Maria M. Robinson Perceiving stimulus displacements across saccades Journal Article In: Visual Cognition, vol. 22, no. 3, pp. 548–575, 2014. @article{Irwin2014a, The visual world appears stable despite frequent retinal image movements caused by saccades. Many theories of visual stability assume that extraretinal eye position information is used to spatially adjust perceived locations across saccades, whereas others have proposed that visual stability depends upon coding of the relative positions of objects. McConkie and Currie (1996) proposed a refined combination of these views (called the Saccade Target Object Theory) in which the perception of stability across saccades relies on a local evaluation process centred on the saccade target object rather than on a remapping of the entire scene, with some contribution from memory for the relative positions of objects as well. Three experiments investigated the saccade target object theory, along with an alternative hypothesis that proposes that multiple objects are updated across saccades, but with variable resolution, with the saccade target object (by virtue of being the focus of attention before the saccade and residing near the fovea after the saccade) having priority in the perception of displacement. Although support was found for the saccade target object theory in Experiment 1, the results of Experiments 2 and 3 found that multiple objects are updated across saccades and that their positions are evaluated to determine perceived stability. There is an advantage for detecting displacements of the saccade target, most likely because of visual acuity or attentional focus being better near the fovea, but it is not the saccade target alone that determines the perception of stability and of displacements across saccades. Rather, multiple sources of information appear to contribute. |
Tomohiro Ishizu; Semir Zeki Varieties of perceptual instability and their neural correlates Journal Article In: NeuroImage, vol. 91, pp. 203–209, 2014. @article{Ishizu2014, We report experiments designed to learn whether different kinds of perceptually unstable visual images engage different neural mechanisms. 21 subjects viewed two types of bi-stable images while we scanned the activity in their brains with functional magnetic resonance imaging (fMRI); in one (intra-categorical type) the two percepts remained within the same category (e.g. face-face) while in the other (cross-categorical type) they crossed categorical boundaries (e.g. face-body). The results showed that cross- and intra-categorical reversals share a common reversal-related neural circuitry, which includes fronto-parietal cortex and primary visual cortex (area V1). Cross-categorical reversals alone engaged additional areas, notably anterior cingulate cortex and superior temporal gyrus, which have been posited to be involved in conflict resolution. |
Leyla Isik; Ethan M. Meyers; Joel Z. Leibo; Tomaso Poggio The dynamics of invariant object recognition in the human visual system Journal Article In: Journal of Neurophysiology, vol. 111, no. 1, pp. 91–102, 2014. @article{Isik2014, The human visual system can rapidly recognize objects despite transformations that alter their appearance. The precise timing of when the brain computes neural representations that are invariant to particular transformations, however, has not been mapped in humans. Here we employ magnetoencephalography decoding analysis to measure the dynamics of size- and position-invariant visual information development in the ventral visual stream. With this method we can read out the identity of objects beginning as early as 60 ms. Size- and position-invariant visual information appear around 125 ms and 150 ms, respectively, and both develop in stages, with invariance to smaller transformations arising before invariance to larger transformations. Additionally, the magnetoencephalography sensor activity localizes to neural sources that are in the most posterior occipital regions at the early decoding times and then move temporally as invariant information develops. These results provide previously unknown latencies for key stages of human-invariant object recognition, as well as new and compelling evidence for a feed-forward hierarchical model of invariant object recognition where invariance increases at each successive visual area along the ventral stream. |
Elena I. Ivleva; Amanda F. Moates; Jordan P. Hamm; Ira H. Bernstein; Hugh B. O'Neill; Darwynn Cole; Brett A. Clementz; Gunvant K. Thaker; Carol A. Tamminga In: Schizophrenia Bulletin, vol. 40, no. 3, pp. 642–652, 2014. @article{Ivleva2014, BACKGROUND: This study examined smooth pursuit eye movement (SPEM), prepulse inhibition (PPI), and auditory event-related potentials (ERP) to paired stimuli as putative endophenotypes of psychosis across the schizophrenia-bipolar disorder dimension. METHODS: Sixty-four schizophrenia probands (SZP), 40 psychotic bipolar I disorder probands (BDP), 31 relatives of SZP (SZR), 26 relatives of BDP (BDR), and 53 healthy controls (HC) were tested. Standard clinical characterization, SPEM, PPI, and ERP measures were administered. RESULTS: There were no differences between either SZP and BDP or SZR and BDR on any of the SPEM, PPI, or ERP measure. Compared with HC, SZP and BDP had lower SPEM maintenance and predictive pursuit gain and ERP theta/alpha and beta magnitudes to the initial stimulus. PPI did not differ between the psychosis probands and HC. Compared with HC, SZR and BDR had lower predictive pursuit gain and ERP theta/alpha and beta magnitudes to the first stimulus with differences ranging from a significant to a trend level. Neither active symptoms severity nor concomitant medications were associated with neurophysiological outcomes. SPEM, PPI, and ERP scores had low intercorrelations. CONCLUSION: These findings support SPEM predictive pursuit and lower frequency auditory ERP activity in a paired stimuli paradigm as putative endophenotypes of psychosis common to SZ and BD probands and relatives. PPI did not differ between the psychosis probands and HC. Future studies in larger scale psychosis family samples targeting putative psychosis endophenotypes and underlying molecular and genetic mediators may aid in the development of biology-based diagnostic definitions. |
Anshul Jain; Stuart Fuller; Benjamin T. Backus Cue-recruitment for extrinsic signals after training with low information stimuli Journal Article In: PLoS ONE, vol. 9, no. 5, pp. e96383, 2014. @article{Jain2014, Cue-recruitment occurs when a previously ineffective signal comes to affect the perceptual appearance of a target object, in a manner similar to the trusted cues with which the signal was put into correlation during training. Jain, Fuller and Backus reported that extrinsic signals, those not carried by the target object itself, were not recruited even after extensive training. However, recent studies have shown that training using weakened trusted cues can facilitate recruitment of intrinsic signals. The current study was designed to examine whether extrinsic signals can be recruited by putting them in correlation with weakened trusted cues. Specifically, we tested whether an extrinsic visual signal, the rotary motion direction of an annulus of random dots, and an extrinsic auditory signal, direction of an auditory pitch glide, can be recruited as cues for the rotation direction of a Necker cube. We found learning, albeit weak, for visual but not for auditory signals. These results extend the generality of the cue-recruitment phenomenon to an extrinsic signal and provide further evidence that the visual system learns to use new signals most quickly when other, long-trusted cues are unavailable or unreliable. |
David C. Jangraw; Ansh Johri; Meron Gribetz; Paul Sajda NEDE: An open-source scripting suite for developing experiments in 3D virtual environments Journal Article In: Journal of Neuroscience Methods, vol. 235, pp. 245–251, 2014. @article{Jangraw2014, Background: As neuroscientists endeavor to understand the brain's response to ecologically valid scenarios, many are leaving behind hyper-controlled paradigms in favor of more realistic ones. This movement has made the use of 3D rendering software an increasingly compelling option. However, mastering such software and scripting rigorous experiments requires a daunting amount of time and effort. New method: To reduce these startup costs and make virtual environment studies more accessible to researchers, we demonstrate a naturalistic experimental design environment (NEDE) that allows experimenters to present realistic virtual stimuli while still providing tight control over the subject's experience. NEDE is a suite of open-source scripts built on the widely used Unity3D game development software, giving experimenters access to powerful rendering tools while interfacing with eye tracking and EEG, randomizing stimuli, and providing custom task prompts. Results: Researchers using NEDE can present a dynamic 3D virtual environment in which randomized stimulus objects can be placed, allowing subjects to explore in search of these objects. NEDE interfaces with a research-grade eye tracker in real-time to maintain precise timing records and sync with EEG or other recording modalities. Comparison with existing methods: Python offers an alternative for experienced programmers who feel comfortable mastering and integrating the various toolboxes available. NEDE combines many of these capabilities with an easy-to-use interface and, through Unity's extensive user base, a much more substantial body of assets and tutorials. Conclusions: Our flexible, open-source experimental design system lowers the barrier to entry for neuroscientists interested in developing experiments in realistic virtual environments. |
David C. Jangraw; Jun Wang; Brent J. Lance; Shih Fu Chang; Paul Sajda Neurally and ocularly informed graph-based models for searching 3D environments Journal Article In: Journal of Neural Engineering, vol. 11, no. 4, pp. 1–13, 2014. @article{Jangraw2014a, OBJECTIVE: As we move through an environment, we are constantly making assessments, judgments and decisions about the things we encounter. Some are acted upon immediately, but many more become mental notes or fleeting impressions-our implicit 'labeling' of the world. In this paper, we use physiological correlates of this labeling to construct a hybrid brain-computer interface (hBCI) system for efficient navigation of a 3D environment. APPROACH: First, we record electroencephalographic (EEG), saccadic and pupillary data from subjects as they move through a small part of a 3D virtual city under free-viewing conditions. Using machine learning, we integrate the neural and ocular signals evoked by the objects they encounter to infer which ones are of subjective interest to them. These inferred labels are propagated through a large computer vision graph of objects in the city, using semi-supervised learning to identify other, unseen objects that are visually similar to the labeled ones. Finally, the system plots an efficient route to help the subjects visit the 'similar' objects it identifies. MAIN RESULTS: We show that by exploiting the subjects' implicit labeling to find objects of interest instead of exploring naively, the median search precision is increased from 25% to 97%, and the median subject need only travel 40% of the distance to see 84% of the objects of interest. We also find that the neural and ocular signals contribute in a complementary fashion to the classifiers' inference of subjects' implicit labeling. SIGNIFICANCE: In summary, we show that neural and ocular signals reflecting subjective assessment of objects in a 3D environment can be used to inform a graph-based learning model of that environment, resulting in an hBCI system that improves navigation and information delivery specific to the user's interests. |
Benedetta Heimler; Francesco Pavani; Mieke Donk; Wieske Zoest Stimulus-and goal-driven control of eye movements: Action videogame players are faster but not better Journal Article In: Attention, Perception, & Psychophysics, vol. 76, no. 8, pp. 2398–2412, 2014. @article{Heimler2014, Action videogame players (AVGPs) have been shown to outperform nongamers (NVGPs) in covert visual attention tasks. These advantages have been attributed to improved top-down control in this population. The time course of visual selection, which permits researchers to highlight when top-down strategies start to control performance, has rarely been investigated in AVGPs. Here, we addressed specifically this issue through an oculomotor additional-singleton paradigm. Participants were instructed to make a saccadic eye movement to a unique orientation singleton. The target was presented among homogeneous nontargets and one additional orientation singleton that was more, equally, or less salient than the target. Saliency was manipulated in the color dimension. Our results showed similar patterns of performance for both AVGPs and NVGPs: Fast-initiated saccades were saliency-driven, whereas later-initiated saccades were more goal-driven. However, although AVGPs were faster than NVGPs, they were also less accurate. Importantly, a multinomial model applied to the data revealed comparable underlying saliency-driven and goal-driven functions for the two groups. Taken together, the observed differences in performance are compatible with the presence of a lower decision bound for releasing saccades in AVGPs than in NVGPs, in the context of comparable temporal interplay between the underlying attentional mechanisms. In sum, the present findings show that in both AVGPs and NVGPs, the implementation of top-down control in visual selection takes time to come about, and they argue against the idea of a general enhancement of top-down control in AVGPs. |
Nils Heise; Ulrich Ansorge The roles of scene priming and location priming in object-scene consistency effects Journal Article In: Frontiers in Psychology, vol. 5, pp. 520, 2014. @article{Heise2014, Presenting consistent objects in scenes facilitates object recognition as compared to inconsistent objects. Yet the mechanisms by which scenes influence object recognition are still not understood. According to one theory, consistent scenes facilitate visual search for objects at expected places. Here, we investigated two predictions following from this theory: If visual search is responsible for consistency effects, consistency effects could be weaker (1) with better-primed than less-primed object locations, and (2) with less-primed than better-primed scenes. In Experiments 1 and 2, locations of objects were varied within a scene to a different degree (one, two, or four possible locations). In addition, object-scene consistency was studied as a function of progressive numbers of repetitions of the backgrounds. Because repeating locations and backgrounds could facilitate visual search for objects, these repetitions might alter the object-scene consistency effect by lowering of location uncertainty. Although we find evidence for a significant consistency effect, we find no clear support for impacts of scene priming or location priming on the size of the consistency effect. Additionally, we find evidence that the consistency effect is dependent on the eccentricity of the target objects. These results point to only small influences of priming to object-scene consistency effects but all-in-all the findings can be reconciled with a visual-search explanation of the consistency effect. |
Karen S. Helfer; Adrian Staub Competing speech perception in older and younger adults: Behavioral and eye-movement evidence Journal Article In: Ear & Hearing, vol. 35, no. 2, pp. 161–170, 2014. @article{Helfer2014, OBJECTIVES: The purpose of this study was to examine eye-movement patterns in older and younger adults to identify differences in how they respond to both to-be-attended and to-be-ignored speech. DESIGN: The study described in this article used an eye-tracking paradigm to provide insight into the factors underlying competing speech understanding in older (n = 23) and younger (n = 22) listeners. Participants attended to a sentence presented in one ear and were instructed to click on a visually displayed word that was heard in that ear while their eye movements were monitored. A foil word also was shown on the screen. Either no sound, steady state noise, or competing speech was presented to the other ear. RESULTS: Comparisons between younger and older listeners on all three types of indicators measured in this study (percent correct, response time, and eye movement patterns) demonstrated that older adults were more greatly affected by competing speech than were younger adults. Differences between the groups could not be attributed to the presence of hearing loss in the older participants, as performance for all subjects was at ceiling in quiet and none of the performance metrics was significantly associated with degree of hearing loss. CONCLUSIONS: Results of this study support the idea that age-related changes other than lack of audibility or susceptibility to energetic masking negatively affect the ability to understand speech in the presence of a competing message. |
Randolph F. Helfrich; Hannah Knepper; Guido Nolte; Daniel Strüber; Stefan Rach; Christoph S. Herrmann; Till R. Schneider; Andreas K. Engel Selective modulation of interhemispheric functional connectivity by HD-tACS shapes perception Journal Article In: PLoS Biology, vol. 12, no. 12, pp. 1–15, 2014. @article{Helfrich2014, Oscillatory neuronal synchronization between cortical areas has been suggested to constitute a flexible mechanism to coordinate information flow in the human cerebral cortex. However, it remains unclear whether synchronized neuronal activity merely represents an epiphenomenon or whether it is causally involved in the selective gating of information. Here, we combined bilateral high-density transcranial alternating current stimulation (HD-tACS) at 40 Hz with simultaneous electroencephalographic (EEG) recordings to study immediate electrophysiological effects during the selective entrainment of oscillatory gamma-band signatures. We found that interhemispheric functional connectivity was modulated in a predictable, phase-specific way: In-phase stimulation enhanced synchronization, anti-phase stimulation impaired functional coupling. Perceptual correlates of these connectivity changes were found in an ambiguous motion task, which strongly support the functional relevance of long-range neuronal coupling. Additionally, our results revealed a decrease in oscillatory alpha power in response to the entrainment of gamma band signatures. This finding provides causal evidence for the antagonistic role of alpha and gamma oscillations in the parieto-occipital cortex and confirms that the observed gamma band modulations were physiological in nature. Our results demonstrate that synchronized cortical network activity across several spatiotemporal scales is essential for conscious perception and cognition. |
Daphna Heller; Craig G. Chambers Would a blue kite by any other name be just as blue? Effects of descriptive choices on subsequent referential behavior Journal Article In: Journal of Memory and Language, vol. 70, no. 1, pp. 53–67, 2014. @article{Heller2014, Using objects that contrast along multiple dimensions, we examined how the earlier description of an object using one dimension (size/color) influences reference to as-yet unmentioned objects, and how this depends on whether the two objects contrast with each other (i.e., whether they belong to the same nominal category). The dimensions of size and color were used because of their different sensitivity, with size adjectives being more closely tied to the presence of a contrasting object from the same category in the situational context. Experiment 1 elicited speakers' descriptions for an object following an earlier description of another object, and Experiment 2 investigated the real-time comprehension of the second description in a two-utterance sequence. Although the priming of linguistic forms may play a role in explaining some of the observed referential patterns, the full set of data suggests that precedence effects in referential descriptions are best explained in terms of a representation that maps those forms onto a mental representation of entities, namely, a discourse model that encodes relationships between entities. The results also highlight how color and size adjectives are processed differently from the earliest moments in comprehension. |
Andrea Helo; Sebastian Pannasch; Louah Sirri; Pia Rämä The maturation of eye movement behavior: Scene viewing characteristics in children and adults Journal Article In: Vision Research, vol. 103, pp. 83–91, 2014. @article{Helo2014, While the close link between eye movements and visual attention has often been demonstrated, recently distinct attentional modes have been associated with specific eye movement patterns. The ambient mode-serving the localization of objects and dominating early scene inspection-is expressed by short fixations and large saccade amplitudes. The focal mode-associated with the identification of object details and dominating later stages of scene exploration-is indicated by longer fixations embedded in short saccades. The relationship between these processing modes and eye movement characteristics has so far only been examined in adults. While studies in children revealed a maturation of oculomotor behavior up to adolescence, developmental aspects of the processing modes are still unknown. Here we explored these mechanisms by comparing eye movements during the inspection of naturalistic scenes. Therefore, gaze behavior from adults and children in four different age groups (2, 4-6, 6-8, 8-10. years old) was examined. We found a general effect of age, revealing that with age fixation durations decrease and saccade amplitudes increase. However, in all age groups fixations were shorter and saccades were longer at the beginning of scene inspection but fixations became longer and saccades became shorter over time. While saliency influenced eye guidance in the two youngest groups over the full inspection period, for the older groups this influence was found only at the beginning of scene inspection. The results reveal indications for ambient and focal processing strategies for as early as 2 years of age. |
Kasey S. Hemington; James N. Reynolds In: Clinical Neurophysiology, vol. 125, no. 12, pp. 2364–2371, 2014. @article{Hemington2014, Objective: Children with Fetal Alcohol Spectrum Disorder (FASD) exhibit cognitive deficits that can be probed using eye movement tasks. We employed a recently developed, single-sensor electroencephalographic (EEG) recording device in measuring EEG activity during the performance of an eye movement task probing working memory in this population. Methods: Children with FASD (n= 18) and typically developing children (n= 19) performed a memory-guided saccade task requiring the participant to remember the spatial location of one, two or three stimuli. We hypothesized that children with FASD would (i) exhibit performance deficits, particularly at greater mnemonic loads; and (ii) display differences in theta (4-8 Hz) and alpha (8-12 Hz) frequency band power compared with controls. Results: Children with FASD failed to perform the task correctly more often than controls when presented with two or three stimuli, and demonstrated related reductions in alpha and theta power. Conclusion: These data suggest that the memory-guided task is sensitive to working memory deficits in children with FASD. Significance: Simultaneous recording of EEG activity suggest differing patterns of underlying neural recruitment in the clinical group, consistent with previous literature indicating more cognitive resources are required by children with FASD in order to complete complex tasks correctly. |
John M. Henderson; Wonil Choi; Steven G. Luke Morphology of primary visual cortex predicts individual differences in fixation duration during text reading Journal Article In: Journal of Cognitive Neuroscience, vol. 26, no. 12, pp. 2880–2888, 2014. @article{Henderson2014a, In skilled reading, fixations are brief periods of time in which the eyes settle on words. E-Z Reader, a computational model of dynamic reading, posits that fixation durations are under realtime control of lexical processing. Lexical processing, in turn, requires efficient visual encoding. Here we tested the hypothesis that individual differences in fixation durations are related to individual differences in the efficiency of early visual encoding. To test this hypothesis, we recorded participantsʼ eye movements during reading. We then examined individual differences in fixation duration distributions as a function of individual differences in the morphology of primary visual cortex measured from MRI scans. The results showed that greater gray matter surface area and volume in visual cortex predicted shorter and less variable fixation durations in reading. These results suggest that individual differences in eye movements during skilled reading are related to initial visual encoding, consistent with models such as E-Z Reader that emphasize lexical control over fixation time. |
John M. Henderson; Steven G. Luke Stable individual differences in saccadic eye movements during reading, pseudoreading, scene viewing, and scene search Journal Article In: Journal of Experimental Psychology: Human Perception and Performance, vol. 40, no. 4, pp. 1390–1400, 2014. @article{Henderson2014, Mean fixation duration and mean saccade amplitude during active viewing tasks differ from person to person. Previous studies have shown that these individual differences tend to be stable across at least some tasks, suggesting that they may reflect underlying traits associated with individuals. However, whether these individual differences are also stable over time has not been established. The present study established stable individual differences in mean fixation duration and mean saccade amplitude across 4 viewing tasks, showed that the observed individual differences are stable over several days, and extended these results to standard deviations of fixation duration and saccade amplitude. The results have implications for theories of eye movement control and for using eye movement characteristics as individual difference measures. |
John M. Henderson; Jennifer Olejarczyk; Steven G. Luke; Joseph Schmidt Eye movement control during scene viewing: Immediate degradation and enhancement effects of spatial frequency filtering Journal Article In: Visual Cognition, vol. 22, no. 3-4, pp. 486–502, 2014. @article{Henderson2014b, What controls how long the eyes remain fixated during scene perception? We investigated whether fixation durations are under the immediate control of the quality of the current scene image. Subjects freely viewed photographs of scenes in preparation for a later memory test while their eye movements were recorded. Using the saccade-contingent display change method, scenes were degraded (Experiment 1) or enhanced (Experiment 2) via blurring (low-pass filtering) during predefined saccades. Results showed that fixation durations immediately after a display change were influenced by the degree of blur, with a monotonic relationship between degree of blur and fixation duration. The results also demonstrated that fixation durations can be both increased and decreased by changes in the degree of blur. The results suggest that fixation durations in scene viewing are influenced by the ease of processing of the image currently in view. The results are consistent with models of saccade generation in scenes in which moment-to-moment difficulty in visual and cognitive processing modulates fixation durations.$backslash$nWhat controls how long the eyes remain fixated during scene perception? We investigated whether fixation durations are under the immediate control of the quality of the current scene image. Subjects freely viewed photographs of scenes in preparation for a later memory test while their eye movements were recorded. Using the saccade-contingent display change method, scenes were degraded (Experiment 1) or enhanced (Experiment 2) via blurring (low-pass filtering) during predefined saccades. Results showed that fixation durations immediately after a display change were influenced by the degree of blur, with a monotonic relationship between degree of blur and fixation duration. The results also demonstrated that fixation durations can be both increased and decreased by changes in the degree of blur. The results suggest that fixation durations in scene viewing are influenced by the ease of processing of the image currently in view. The results are consistent with models of saccade generation in scenes in which moment-to-moment difficulty in visual and cognitive processing modulates fixation durations. |
Nora A. Herweg; Bernd Weber; Anna-Maria Kasparbauer; Inga Meyhöfer; Maria Steffens; Nikolaos Smyrnis; Ulrich Ettinger Functional magnetic resonance imaging of sensorimotor transformations in saccades and antisaccades Journal Article In: NeuroImage, vol. 102, pp. 848–860, 2014. @article{Herweg2014, Saccades to peripheral targets require a direct visuomotor transformation. In contrast, antisaccades, saccades in opposite direction of a peripheral target, require more complex transformation processes due to the inversion of the spatial vector. Here, the differential neural mechanisms underlying sensorimotor control in saccades and antisaccades were investigated using functional magnetic resonance imaging (fMRI) at 3. T field strength in 22 human volunteers. We combined a task factor (prosaccades: look towards target; antisaccades: look away from target) with a parametric factor of transformation demand (single vs. multiple peripheral targets) in a two-factorial block design. Behaviorally, a greater number of peripheral targets resulted in decreased spatial accuracy and increased reaction times in antisaccades. No effects were seen on the percentage of antisaccade direction errors or on any prosaccade measures. Neurally, a greater number of targets led to increased BOLD signal in the posterior parietal cortex (PPC) bilaterally. This effect was partially qualified by an interaction that extended into somatosensory cortex, indicating greater increases during antisaccades than prosaccades. The results implicate the PPC as a sensorimotor interface that is especially important in nonstandard mapping for antisaccades and point to a supportive role of somatosensory areas in antisaccade sensorimotor control, possibly by means of proprioceptive processes. |
Arvid Herwig; Werner X. Schneider Predicting object features across saccades: Evidence from object recognition and visual search Journal Article In: Journal of Experimental Psychology: General, vol. 143, no. 5, pp. 1903–1922, 2014. @article{Herwig2014, When we move our eyes, we process objects in the visual field with different spatial resolution due to the nonhomogeneity of our visual system. In particular, peripheral objects are only coarsely represented, whereas they are represented with high acuity when foveated. To keep track of visual features of objects across eye movements, these changes in spatial resolution have to be taken into account. Here, we develop and test a new framework proposing a visual feature prediction mechanism based on past experience to deal with changes in spatial resolution accompanying saccadic eye movements. In 3 experiments, we first exposed participants to an altered visual stimulation where, unnoticed by participants, 1 object systematically changed visual features during saccades. Experiments 1 and 2 then demonstrate that feature prediction during peripheral object recognition is biased toward previously associated postsaccadic foveal input and that this effect is particularly associated with making saccades. Moreover, Experiment 3 shows that during visual search, feature prediction is biased toward previously associated presaccadic peripheral input. Together, these findings demonstrate that the visual system uses past experience to predict how peripheral objects will look in the fovea, and what foveal search templates should look like in the periphery. As such, they support our framework based on ideomotor theory and shed new light on the mystery of why we are most of the time unaware of acuity limitations in the periphery and of our ability to locate relevant objects in the periphery. |
Alistair J. Harvey Some effects of alcohol and eye movements on cross-race face learning Journal Article In: Memory, vol. 22, no. 8, pp. 1126–1138, 2014. @article{Harvey2014, This study examines the impact of acute alcohol intoxication on visual scanning in cross-race face learning. The eye movements of a group of white British participants were recorded as they encoded a series of own-and different-race faces, under alcohol and placebo conditions. Intoxication reduced the rate and extent of visual scanning during face encoding, reorienting the focus of foveal attention away from the eyes and towards the nose. Differences in encoding eye movements also varied between own-and different-race face conditions as a function of alcohol. Fixations to both face types were less frequent and more lingering following intoxication, but in the placebo condition this was only the case for different-race faces. While reducing visual scanning, however, alcohol had no adverse effect on memory, only encoding restrictions associated with sober different-race face processing led to poorer recognition. These results support perceptual expertise accounts of own-race face processing, but suggest the adverse effects of alcohol on face learning published previously are not caused by foveal encoding restrictions. The implications of these findings for alcohol myopia theory are discussed. |
Hannah Harvey; Robin Walker Reading with peripheral vision: A comparison of reading dynamic scrolling and static text with a simulated central scotoma Journal Article In: Vision Research, vol. 98, pp. 54–60, 2014. @article{Harvey2014a, Horizontally scrolling text is, in theory, ideally suited to enhance viewing strategies recommended to improve reading performance under conditions of central vision loss such as macular disease, although it is largely unproven in this regard. This study investigated if the use of scrolling text produced an observable improvement in reading performed under conditions of eccentric viewing in an artificial scotoma paradigm. Participants (n=17) read scrolling and static text with a central artificial scotoma controlled by an eye-tracker. There was an improvement in measures of reading accuracy, and adherence to eccentric viewing strategies with scrolling, compared to static, text. These findings illustrate the potential benefits of scrolling text as a potential reading aid for those with central vision loss. |
Nabil Hasshim; Benjamin A. Parris Two-to-one color-response mapping and the presence of semantic conflict in the Stroop task Journal Article In: Frontiers in Psychology, vol. 5, pp. 1157, 2014. @article{Hasshim2014, A series of recent studies have utilized the two-to-one mapping paradigm in the Stroop task. In this paradigm, the word red might be presented in blue when both red and blue share the same-response key (same-response trials). This manipulation has been used to show the separate contributions of (within) semantic category conflict and response conflict to Stroop interference. Such results evidencing semantic category conflict are incompatible with models of the Stroop task that are based on response conflict only. However, the nature of same-response trials is unclear since they are also likely to involve response facilitation given that both dimensions of the stimulus provide evidence toward the sameresponse. In this study we explored this possibility by comparing them with three other trial types. We report strong (Bayesian) evidence for no statistical difference between sameresponse and non-color word neutral trials, faster responses to same-response trials than to non-response set incongruent trials, and no differences between same-response vs. congruent trials when contingency is controlled. Our results suggest that same-response trials are not different from neutral trials indicating that they cannot be used reliably to determine the presence or absence of semantic category conflict. In light of these results, the interpretation of a series of recent studies might have to be reassessed. |
Anna Hatzidaki; Manon W. Jones; M. Santesteban; H. P. Branigan It's not what you see: It's the language you say it in Journal Article In: Language, Cognition and Neuroscience, vol. 29, no. 10, pp. 1233–1239, 2014. @article{Hatzidaki2014, In an eye-tracking experiment, we investigated the interplay between visual and linguistic information processing during time-telling, and how this is affected by speaking in a non-native language. We compared time-telling in Greek and English, which differ in time-telling word order (hour vs. minute mentioned first), by contrasting Greek-English bilinguals speaking in their L1-Greek or their L2-English, and English monolingual speakers. All three groups were faster when telling the time for digital than for analogue clocks, and when telling the time for the first half-hour than the second half-hour. Critically, first fixation and gaze duration analyses for the hour and minute regions showed a different pattern for Greek-English bilinguals when speaking in their L1 versus L2, with the latter resembling that of English monolinguals. Our results suggest that bilingual speakers' eye-movement programming was influenced by the type of time-telling utterance specific to the language of production currently in use.$backslash$nIn an eye-tracking experiment, we investigated the interplay between visual and linguistic information processing during time-telling, and how this is affected by speaking in a non-native language. We compared time-telling in Greek and English, which differ in time-telling word order (hour vs. minute mentioned first), by contrasting Greek-English bilinguals speaking in their L1-Greek or their L2-English, and English monolingual speakers. All three groups were faster when telling the time for digital than for analogue clocks, and when telling the time for the first half-hour than the second half-hour. Critically, first fixation and gaze duration analyses for the hour and minute regions showed a different pattern for Greek-English bilinguals when speaking in their L1 versus L2, with the latter resembling that of English monolinguals. Our results suggest that bilingual speakers' eye-movement programming was influenced by the type of time-telling utterance specific to the language of production currently in use. |
Wei He; Jon Brock; Blake W. Johnson Face-sensitive brain responses measured from a four-year-old child with a custom-sized child MEG system Journal Article In: Journal of Neuroscience Methods, vol. 222, pp. 213–217, 2014. @article{He2014, Background: Previous magnetoencephalography (MEG) studies have failed to find a facesensitive, brain response-M170 in children. If this is the case, this suggests that the developmental trajectory of the M170 is different from that of its electrical equivalent, the N170. We investigated the alternative possibility that the child M170 may not be detectable in conventional adult-sized MEG systems. New method: Brain responses to pictures of faces and well controlled stimuli were measured from the same four-year-old child with a custom child MEG system and an adult-sized MEG system. Results: The goodness of fit of the child's head was about the same over the occipital head surface in both systems, but was much worse over all other parts of the head surface in the adult MEG system compared to the child MEG system. The face-sensitive M170 was measured from the child in both MEG systems, but was larger in amplitude, clearer in morphology, and had a more accurate source localization when measured in the child MEG system. Comparison with existing method: The custom-sized child MEG system is superior for measuring the face-sensitive M170 brain response in children than the conventional adult MEG system. Conclusions: The present results show that the face-sensitive M170 brain response can be elicited in a four-year-old child. This provides new evidence for early maturation of face processing brain mechanisms in humans, and offers new opportunities for the study of neurodevelopmental disorders that show atypical face processing capabilities, such as autism spectrum disorder. |
Kristin J. Heaton; Alexis L. Maule; Jun Maruta; Elisabeth M. Kryskow; Jamshid Ghajar Attention and visual tracking degradation during acute sleep deprivation in a military sample Journal Article In: Aviation Space and Environmental Medicine, vol. 85, no. 5, pp. 497–503, 2014. @article{Heaton2014, Background: Fatigue due to sleep restriction places individuals at elevated risk for accidents, degraded health, and impaired physical and mental performance. Early detection of fatigue-related performance decrements is an important component of injury prevention and can help to ensure optimal performance and mission readiness. This study used a predictive visual tracking task and a computer-based measure of attention to characterize fatigue-related attention decrements in healthy Army personnel during acute sleep deprivation. Methods: Serving as subjects in this laboratory-based study were 87 male and female service members between the ages of 18 and 50 with no history of brain injury with loss of consciousness, substance abuse, or significant psychiatric or neurologic diagnoses. Subjects underwent 26 h of sleep deprivation, during which eye movement measures from a continuous circular visual tracking task and attention measures (reaction time, accuracy) from the Attention Network Test (ANT) were collected at baseline, 20 h awake, and between 24 to 26 h awake. Results: Increases in the variability of gaze positional errors (46-47%), as well as reaction time-based ANT measures (9-65%), were observed across 26 h of sleep deprivation. Accuracy of ANT responses declined across this same period (11%). Discussion: Performance measures of predictive visual tracking accurately reflect impaired attention due to acute sleep deprivation and provide a promising approach for assessing readiness in personnel serving in diverse occupational areas, including flight and ground support crews. |
Jessica Heeman; Jan Theeuwes; Stefan Van der Stigchel The time course of top-down control on saccade averaging Journal Article In: Vision Research, vol. 100, pp. 29–37, 2014. @article{Heeman2014, When objects in a visual scene are positioned in close proximity, eye movements to these objects tend to land at an intermediate location between the objects (i.e. the global effect). This effect is most pronounced for short latency saccades and is therefore believed to be reflexive and dominantly controlled by bottom-up information. At longer latencies this effect can be modulated by top-down factors. The current study established the time course at which top-down information starts to have an influence on bottom-up averaging. In a standard global effect task two peripheral stimuli (a red and a green abrupt onset) were positioned within an angular distance of 20°. In the condition in which observers received no specific target instruction, the eyes landed in between the red and green element establishing the classic global effect. However, when observers were instructed to make a saccade to the red element during a whole block or when the target color varied from trial-to-trial (red or green), a clear effect of the target instruction on the accuracy of the landing position of the primary saccade was found. With increasing saccade latencies, the eyes landed closer to the instructed target. Crucially, however, this effect was even seen for the shortest saccade latencies (as early as 200ms), suggesting that saccade averaging is affected early on by top-down processes. |
Constanze Hesse; Keira Ball; Thomas Schenk Pointing in visual periphery: Is DF's dorsal stream intact? Journal Article In: PLoS ONE, vol. 9, no. 3, pp. e91420, 2014. @article{Hesse2014, Observations of the visual form agnosic patient DF have been highly influential in establishing the hypothesis that separate processing streams deal with vision for perception (ventral stream) and vision for action (dorsal stream). In this context, DF's preserved ability to perform visually-guided actions has been contrasted with the selective impairment of visuomotor performance in optic ataxia patients suffering from damage to dorsal stream areas. However, the recent finding that DF shows a thinning of the grey matter in the dorsal stream regions of both hemispheres in combination with the observation that her right-handed movements are impaired when they are performed in visual periphery has opened up the possibility that patient DF may potentially also be suffering from optic ataxia. If lesions to the posterior parietal cortex (dorsal stream) are bilateral, pointing and reaching deficits should be observed in both visual hemifields and for both hands when targets are viewed in visual periphery. Here, we tested DF's visuomotor performance when pointing with her left and her right hand toward targets presented in the left and the right visual field at three different visual eccentricities. Our results indicate that DF shows large and consistent impairments in all conditions. These findings imply that DF's dorsal stream atrophies are functionally relevant and hence challenge the idea that patient DF's seemingly normal visuomotor behaviour can be attributed to her intact dorsal stream. Instead, DF seems to be a patient who suffers from combined ventral and dorsal stream damage meaning that a new account is needed to explain why she shows such remarkably normal visuomotor behaviour in a number of tasks and conditions. |
Simon J. Hickman; Naz Raoof; Rebecca J. McLean; Irene Gottlob Vision and multiple sclerosis Journal Article In: Multiple Sclerosis and Related Disorders, vol. 3, no. 1, pp. 3–16, 2014. @article{Hickman2014, Multiple sclerosis can affect vision in many ways, including optic neuritis, chronic optic neuropathy, retrochiasmal visual field defects, higher order cortical processing, double vision, nystagmus and also by related ocular conditions such as uveitis. There are also side effects from recently introduced multiple sclerosis treatments that can affect vision. This review will discuss all these aspects and how they come together to cause visual symptoms. It will then focus on practical aspects of how to recognise when there is a vision problem in a multiple sclerosis patient and on what treatments are available to improve vision. |
Matthew D. Hilchey; Mahmoud Hashish; Gregory H. MacLean; Jason Satel; Jason Ivanoff; Raymond M. Klein On the role of eye movement monitoring and discouragement on inhibition of return in a go/no-go task Journal Article In: Vision Research, vol. 96, pp. 133–139, 2014. @article{Hilchey2014, Inhibition of return (IOR) most often describes the finding of increased response times to cued as compared to uncued targets in the standard covert orienting paradigm. A perennial question in the IOR literature centers on whether the effect of IOR is on motoric/decision-making processes (output-based IOR), attentional/perceptual processes (input-based IOR), or both. Recent data converge on the idea that IOR is an output-based effect when eye movements are required or permitted whereas IOR is an input-based effect when eye movements are monitored and actively discouraged. The notion that the effects of IOR may be fundamentally different depending on the activation state of the oculomotor system has been challenged by several studies demonstrating that IOR exists as an output-, or output- plus input-based effect in simple keypress tasks not requiring oculomotor responses. Problematically, experiments in which keypress responses are required to visual events rarely use eye movement monitoring let alone the active discouragement of eye movement errors. Here, we return to an experimental method implemented by Ivanoff and Klein (2001) whose results demonstrated that IOR affected output-based processes when, ostensibly, only keypress responses occurred. Unlike Ivanoff and Klein, however, we assiduously monitor and discourage eye movements. We demonstrate that actively discouraging eye movements in keypress tasks changes the form of IOR from output- to input-based and, as such, we strongly encourage superior experimental control over or consideration of the contribution of eye movement activity in simple keypress tasks exploring IOR. |
Matthew D. Hilchey; Raymond M. Klein; Jason Satel Returning to “inhibition of return” by dissociating long-term oculomotor IOR from short-term sensory effects Journal Article In: Journal of Experimental Psychology: Human Perception and Performance, vol. 40, no. 4, pp. 1603–1616, 2014. @article{Hilchey2014a, We explored the nature and time course of effects generated by spatially uninformative peripheral cues by measuring these effects with localization responses to peripheral onsets or central arrow targets. In Experiment 1, participants made saccadic eye movements to equiprobable peripheral and central targets. At short cue-target onset asynchronies (CTOAs), responses to cued peripheral stimuli suffered from slowed responding attributable to sensory adaptation while responses to central targets were transiently facilitated, presumably due to cue-elicited oculomotor activation. At the longest CTOA, saccadic responses to central and peripheral targets were indistinguishably delayed, suggesting a common, output/decision effect (inhibition of return; IOR). In Experiment 2, we tested the hypothesis that the generation of this output effect is dependent on the activation state of the oculomotor system by forbidding eye movements and requiring keypress responses to frequent peripheral targets, while probing oculomotor behavior with saccades to infrequent central arrow targets. As predicted, saccades to central arrow targets showed neither the early facilitation nor later inhibitory effects that were robust in Experiment 1. At the long CTOA, manual responses to cued peripheral targets showed the typical delayed responses usually attributed to IOR. We recommend that this late “inhibitory” cueing effect (ICE) be distinguished from IOR because it lacks the cause (oculomotor activation) and effect (response bias) attributed to IOR when it was named by Posner, Rafal, Choate, and Vaughan (1985). |
Renske S. Hoedemaker; Peter C. Gordon It takes time to prime: Semantic priming in the ocular lexical decision task Journal Article In: Journal of Experimental Psychology: Human Perception and Performance, vol. 40, no. 6, pp. 2179–2197, 2014. @article{Hoedemaker2014a, Two eye-tracking experiments were conducted in which the manual response mode typically used in lexical decision tasks (LDTs) was replaced with an eye-movement response through a sequence of 3 words. This ocular LDT combines the explicit control of task goals found in LDTs with the highly practiced ocular response used in reading text. In Experiment 1, forward saccades indicated an affirmative lexical decision (LD) on each word in the triplet. In Experiment 2, LD responses were delayed until all 3 letter strings had been read. The goal of the study was to evaluate the contribution of task goals and response mode to semantic priming. Semantic priming is very robust in tasks that involve recognition of words in isolation, such as LDT, but limited during text reading, as measured using eye movements. Gaze durations in both experiments showed robust semantic priming even though ocular response times were much shorter than manual LDs for the same words in the English Lexicon Project. Ex-Gaussian distribution fits revealed that the priming effect was concentrated in estimates of tau (tau), meaning that priming was most pronounced in the slow tail of the distribution. This pattern shows differential use of the prime information, which may be more heavily recruited in cases in which the LD is difficult, as indicated by longer response times. Compared with the manual LD responses, ocular LDs provide a more sensitive measure of this task-related influence on word recognition as measured by the LDT. |
Renske S. Hoedemaker; Peter C. Gordon Embodied language comprehension: Encoding-based and goal-driven processes Journal Article In: Journal of Experimental Psychology: General, vol. 143, no. 2, pp. 914–929, 2014. @article{Hoedemaker2014, Theories of embodied language comprehension have proposed that language is understood through perceptual simulation of the sensorimotor characteristics of its meaning. Strong support for this claim requires demonstration of encoding-based activation of sensorimotor representations that is distinct from task-related or goal-driven processes. Participants in 3 eye-tracking experiments were presented with triplets of either numbers or object and animal names. In Experiment 1, participants indicated whether the size of the referent of the middle object or animal name was in between the size of the 2 outer items. In Experiment 2, the object and animal names were encoded for an immediate recognition memory task. In Experiment 3, participants completed the same comparison task of Experiment 1 for both words and numbers. During the comparison tasks, word and number decision times showed a symbolic distance effect, such that response time was inversely related to the size difference between the items. A symbolic distance effect was also observed for animal and object encoding times in cases where encoding time likely reflected some goal-driven processes as well. When semantic size was irrelevant to the task (Experiment 2), it had no effect on word encoding times. Number encoding times showed a numerical distance priming effect: Encoding time increased with numerical difference between items. Together these results suggest that while activation of numerical magnitude representations is encoding-based as well as goal-driven, activation of size information associated with words is goal-driven and does not occur automatically during encoding. This conclusion challenges strong theories of embodied cognition which claim that language comprehension consists of activation of analog sensorimotor representations irrespective of higher level processes related to context or task-specific goals. |
Margit Hofler; Iain D. Gilchrist; Christof Korner Searching the same display twice: Properties of short-term memory in repeated search Journal Article In: Attention, Perception, & Psychophysics, vol. 76, no. 2, pp. 335–352, 2014. @article{Hofler2014, Consecutive search for different targets in the same display is supported by a short-term memory mechanism: Distractors that have recently been inspected in the first search are found more quickly in the second search when they become the target (Exp. 1). Here, we investigated the properties of this memory process. We found that this recency advantage is robust to a delay between the two searches (Exp. 2) and that it is only slightly disrupted by an interference task between the two searches (Exp. 3). Introducing a concurrent secondary task (Exp. 4) showed that the memory representations formed in the first search are based on identity as well as location information. Together, these findings show that the short-term memory that supports repeated visual search stores a complex combination of item identity and location that is robust to disruption by either time or interference. |
Sven Hohenstein; Reinhold Klieg Semantic preview benefit during reading Journal Article In: Journal of Experimental Psychology: Learning, Memory, and Cognition, vol. 40, no. 1, pp. 166–190, 2014. @article{Hohenstein2014, Word features in parafoveal vision influence eye movements during reading. The question of whether readers extract semantic information from parafoveal words was studied in 3 experiments by using a gaze-contingent display change technique. Subjects read German sentences containing 1 of several preview words that were replaced by a target word during the saccade to the preview (boundary paradigm). In the 1st experiment the preview word was semantically related or unrelated to the target. Fixation durations on the target were shorter for semantically related than unrelated previews, consistent with a semantic preview benefit. In the 2nd experiment, half the sentences were presented following the rules of German spelling (i.e., previews and targets were printed with an initial capital letter), and the other half were presented completely in lowercase. A semantic preview benefit was obtained under both conditions. In the 3rd experiment, we introduced 2 further preview conditions, an identical word and a pronounceable nonword, while also manipulating the text contrast. Whereas the contrast had negligible effects, fixation durations on the target were reliably different for all 4 types of preview. Semantic preview benefits were greater for pretarget fixations closer to the boundary (large preview space) and, although not as consistently, for long pretarget fixation durations (long preview time). The results constrain theoretical proposals about eye movement control in reading. |
Pawel Holas; Izabela Krejtz; Marzena Cypryańska; John B. Nezlek Orienting and maintenance of attention to threatening facial expressions in anxiety - An eye movement study Journal Article In: Psychiatry Research, vol. 220, no. 1-2, pp. 362–369, 2014. @article{Holas2014, Cognitive models posit that anxiety disorders stem in part from underlying attentional biases to threat. Consistent with this, studies have found that the attentional bias to threat-related stimuli is greater in high vs. low anxious individuals. Nevertheless, it is not clear if similar biases exist for different threatening emotions or for any facial emotional stimulus. In the present study, we used eye-tracking to measure orienting and maintenance of attention to faces displaying anger, fear and disgust as threats, and faces displaying happiness and sadness. Using a free viewing task, we examined differences between low and high trait anxious (HTA) individuals in the attention they paid to each of these emotional faces (paired with a neutral face). We found that initial orienting was faster for angry and happy faces, and high trait anxious participants were more vigilant to fearful and disgust faces. Our results for attentional maintenance were not consistent. The results of the present study suggest that attentional processes may be more emotion-specific than previously believed. Our results suggest that attentional processes for different threatening emotions may not be the same and that attentional processes for some negative and some positive emotions may be similar. |
Linbi Hong; Jennifer M. Walz; Paul Sajda Your eyes give you away: Prestimulus changes in pupil diameter correlate with poststimulus task-related EEG dynamics Journal Article In: PLoS ONE, vol. 9, no. 3, pp. e91321, 2014. @article{Hong2014, Pupillary measures have been linked to arousal and attention as well as activity in the brainstem's locus coeruleus norepinephrine (LC-NE) system. Similarly, there is evidence that evoked EEG responses, such as the P3, might have LC-NE activity as their basis. Since it is not feasible to record electrophysiological data directly from the LC in humans due to its location in the brainstem, an open question has been whether pupillary measures and EEG variability can be linked in a meaningful way to shed light on the nature of the LC-NE role in attention and arousal. We used an auditory oddball task with a data-driven approach to learn task-relevant projections of the EEG, for windows of data spanning the entire trial. We investigated linear and quadratic relationships between the evoked EEG along these projections and both prestimulus (baseline) and poststimulus (evoked dilation) pupil diameter measurements. We found that baseline pupil diameter correlates with early (175-200 ms) and late (350-400 ms) EEG component variability, suggesting a linear relationship between baseline (tonic) LC-NE activity and evoked EEG. We found no relationships between evoked EEG and evoked pupil dilation, which is often associated with evoked (phasic) LC activity. After regressing out reaction time (RT), the correlation between EEG variability and baseline pupil diameter remained, suggesting that such correlation is not explainable by RT variability. We also investigated the relationship between these pupil measures and prestimulus EEG alpha activity, which has been reported as a marker of attentional state, and found a negative linear relationship with evoked pupil dilation. In summary, our results demonstrate significant relationships between prestimulus and poststimulus neural and pupillary measures, and they provide further evidence for tight coupling between attentional state and evoked neural activity and for the role of cortical and subcortical networks underlying the process of target detection. |
Matthew Haigh; Heather J. Ferguson; Andrew J. Stewart An eye-tracking investigation into readers' sensitivity to actual versus expected utility in the comprehension of conditionals Journal Article In: Quarterly Journal of Experimental Psychology, vol. 67, no. 1, pp. 166–185, 2014. @article{Haigh2014, The successful comprehension of a utility conditional (i.e., an "if p, then q" statement where p and/or q is valued by one or more agents) requires the construction of a mental representation of the situation described by that conditional and integration of this representation with prior context. In an eye-tracking experiment, we examined the time course of integrating conditional utility information into the broader discourse model. Specifically, the experiment determined whether readers were sensitive, during rapid heuristic processing, to the congruency between the utility of the consequent clause of a conditional (positive or negative) and a reader's subjective expectations based on prior context. On a number of eye-tracking measures we found that readers were sensitive to conditional utility-conditionals for which the consequent utility mismatched the utility that would be anticipated on the basis of prior context resulted in processing disruption. Crucially, this sensitivity emerged on measures that are accepted to indicate early processing within the language comprehension system and suggests that the evaluation of a conditional's utility informs the early stages of conditional processing. |
Carlos M. Hamamé; Juan R. Vidal; Marcela Perrone-Bertolotti; Tomás Ossandón; Karim Jerbi; Philippe Kahane; Olivier Bertrand; Jean Philippe Lachaux Functional selectivity in the human occipitotemporal cortex during natural vision: Evidence from combined intracranial EEG and eye-tracking Journal Article In: NeuroImage, vol. 95, pp. 276–286, 2014. @article{Hamame2014, Eye movements are a constant and essential component of natural vision, yet, most of our knowledge about the human visual system comes from experiments that restrict them. This experimental constraint is mostly in place to control visual stimuli presentation and to avoid artifacts in non-invasive measures of brain activity, however, this limitation can be overcome with intracranial EEG (iEEG) recorded from epilepsy patients. Moreover, the high-frequency components of the iEEG signal (between about 50 and 150. Hz) can provide a proxy of population-level spiking activity in any cortical area during free-viewing. We combined iEEG with high precision eye-tracking to study fine temporal dynamics and functional specificity in the fusiform face (FFA) and visual word form area (VWFA) while patients inspected natural pictures containing faces and text. We defined the first local measure of visual (electrophysiological) responsiveness adapted to free-viewing in humans: amplitude modulations in the high-frequency activity range (50-150. Hz) following fixations (fixation-related high-frequency response). We showed that despite the large size of receptive fields in the ventral occipito-temporal cortex, neural activity during natural vision of realistic cluttered scenes is mostly dependent upon the category of the foveated stimulus - suggesting that category-specificity is preserved during free-viewing and that attention mechanisms might filter out the influence of objects surrounding the fovea. |
Yuko Hara; Justin L. Gardner Encoding of graded changes in spatial specificity of prior cues in human visual cortex Journal Article In: Journal of Neurophysiology, vol. 112, no. 11, pp. 2834–2849, 2014. @article{Hara2014, Prior information about the relevance of spatial locations can vary in specificity; a single location, a subset of locations, or all locations may be of potential importance. Using a contrast-discrimination task with four possible targets, we asked whether performance benefits are graded with the spatial specificity of a prior cue and whether we could quantitatively account for behavioral performance with cortical activity changes measured by blood oxygenation level-dependent (BOLD) imaging. Thus we changed the prior probability that each location contained the target from 100 to 50 to 25% by cueing in advance 1, 2, or 4 of the possible locations. We found that behavioral performance (discrimination thresholds) improved in a graded fashion with spatial specificity. However, concurrently measured cortical responses from retinotopically defined visual areas were not strictly graded; response magnitude decreased when all 4 locations were cued (25% prior probability) relative to the 100 and 50% prior probability conditions, but no significant difference in response magnitude was found between the 100 and 50% prior probability conditions for either cued or uncued locations. Also, although cueing locations increased responses relative to noncueing, this cue sensitivity was not graded with prior probability. Furthermore, contrast sensitivity of cortical responses, which could improve contrast discrimination performance, was not graded. Instead, an efficient-selection model showed that even if sensory responses do not strictly scale with prior probability, selection of sensory responses by weighting larger responses more can result in graded behavioral performance benefits with increasing spatial specificity of prior information. |
James J. Harrison; Tom C. A. Freeman; Petroc Sumner In: Journal of Experimental Psychology: General, vol. 143, no. 5, pp. 1923–1938, 2014. @article{Harrison2014, As a potential exemplar for understanding how volitional actions emerged from reflexes, we studied the relationship between an ancient reflexive gaze stabilization mechanism (optokinetic nystagmus [OKN]) and purposeful eye movements (saccades) that target an object. Traditionally, these have been considered distinct (except in the kinematics of their execution) and have been studied independently. We find that the fast-phases of OKN clearly show properties associated with saccade planning: (a) They are characteristically delayed by irrelevant distractors in an indistinguishable way to saccades (the saccadic inhibition effect), and (b) horizontal OKN fast-phases produce curvature in vertical targeting saccades, just like a competing saccade plan. Thus, we argue that the saccade planning network plays a role in the production of OKN fast-phases, and we question the need for a strict distinction between eye movements that appear to be automatic or volitional. We discuss whether our understanding might benefit from shifting perspective and considering the entire "saccade" system to have developed from an increasingly sophisticated OKN system. |
William J. Harrison; Peter J. Bex Integrating retinotopic features in spatiotopic coordinates Journal Article In: Journal of Neuroscience, vol. 34, no. 21, pp. 7351–7360, 2014. @article{Harrison2014a, The receptive fields of early visual neurons are anchored in retinotopic coordinates (Hubel and Wiesel, 1962). Eye movements shift these receptive fields and therefore require that different populations of neurons encode an object's constituent features across saccades. Whether feature groupings are preserved across successive fixations or processing starts anew with each fixation has been hotly debated (Melcher and Morrone, 2003; Melcher, 2005, 2010; Knapen et al., 2009; Cavanagh et al., 2010a,b; Morris et al., 2010). Here we show that feature integration initially occurs within retinotopic coordinates, but is then conserved within a spatiotopic coordinate frame independent of where the features fall on the retinas. With human observers, we first found that the relative timing of visual features plays a critical role in determining the spatial area over which features are grouped. We exploited this temporal dependence of feature integration to show that features co-occurring within 45 ms remain grouped across eye movements. Our results thus challenge purely feedforward models of feature integration (Pelli, 2008; Freeman and Simoncelli, 2011) that begin de novo after every eye movement, and implicate the involvement of brain areas beyond early visual cortex. The strong temporal dependence we quantify and its link with trans-saccadic object perception instead suggest that feature integration depends, at least in part, on feedback from higher brain areas (Mumford, 1992; Rao and Ballard, 1999; Di Lollo et al., 2000; Moore and Armstrong, 2003; Stanford et al., 2010). |
William J. Harrison; Roger W. Remington; Jason B. Mattingley Visual crowding is anisotropic along the horizontal meridian during smooth pursuit Journal Article In: Journal of Vision, vol. 14, no. 1, pp. 1–16, 2014. @article{Harrison2014b, Humans make smooth pursuit eye movements to foveate moving objects of interest. It is known that smooth pursuit alters visual processing, but there is currently no consensus on whether changes in vision are contingent on the direction the eyes are moving. We recently showed that visual crowding can be used as a sensitive measure of changes in visual processing, resulting from involvement of the saccadic eye movement system. The present paper extends these results by examining the effect of smooth pursuit eye movements on the spatial extent of visual crowding-the area over which visual stimuli are integrated. We found systematic changes in crowding that depended on the direction of pursuit and the distance of stimuli from the pursuit target. Relative to when no eye movement was made, the spatial extent of crowding increased for objects located contraversive to the direction of pursuit at an eccentricity of approximately 3°. By contrast, crowding for objects located ipsiversive to the direction of pursuit remained unchanged. There was no change in crowding during smooth pursuit for objects located approximately 7° from the fovea. The increased size of the crowding zone for the contraversive direction may be related to the distance that the fovea lags behind the pursuit target during smooth eye movements. Overall, our results reveal that visual perception is altered dynamically according to the intended destination of oculomotor commands. |
Lynn Huestegge; Iring Koch When two actions are easier than one: How inhibitory control demands affect response processing Journal Article In: Acta Psychologica, vol. 151, pp. 230–236, 2014. @article{Huestegge2014, Numerous studies showed that the simultaneous execution of multiple actions is associated with performance costs. Here, we demonstrate that when highly automatic responses are involved, performance in single-response conditions can actually be worse than in dual-response conditions. Participants responded to peripheral visual stimuli with an eye movement (saccade), a manual key press, or both. To manipulate saccade automaticity, a central fixation cross either remained present throughout the trial (overlap condition, lower automaticity) or disappeared 200. ms before visual target onset (gap condition, greater automaticity). Crucially, single-response conditions yielded more performance errors than dual-response conditions (i.e., dual-response benefit), especially in gap trials. This was due to difficulties associated with inhibiting saccades when only manual responses were required, suggesting that response inhibition (remaining fixated) can be even more resource-demanding than overt response execution (saccade to peripheral target). |
Lynn Huestegge; Aleksandra Pieczykolan; Iring Koch Talking while looking: On the encapsulation of output system representations Journal Article In: Cognitive Psychology, vol. 73, pp. 72–91, 2014. @article{Huestegge2014a, The idea that the human mind can be divided into distinct (but interacting) functional modules is an important presupposition in many theories of cognition. While previous research on modularity predominantly studied input domains (e.g., vision) or central processes, the present study focused on cognitive representations of output domains. Specifically, we asked to what extent output domain representations are encapsulated (i.e., immune to influence from other domains, representing a key feature of modularity) by studying determinants of interference between simultaneous action demands (oculomotor and vocal responses). To examine the degree of encapsulation, we compared single- vs. dual-response performance triggered by single stimuli. Experiment 1 addressed the role of stimulus modality under dimensionally overlapping response requirements (stimuli and responses were spatial and compatible throughout). In Experiment 2, we manipulated the presence of dimensional overlap across responses. Substantial performance costs associated with dual-response (vs. single-response) demands were observed across response modalities, conditions, and experiments. Dimensional overlap combined with shared spatial codes across responses enabled response-code priming (i.e., beneficial crosstalk between output domains). Overall, the results are at odds with the idea of strong encapsulation of output system representations and show how processing content determines the extent of interdependency between output domains in cognition. |
Stephanie Huette; Bodo Winter; Teenie Matlock; David H. Ardell; Michael J. Spivey Eye movements during listening reveal spontaneous grammatical processing Journal Article In: Frontiers in Psychology, vol. 5, pp. 410, 2014. @article{Huette2014, Recent research using eye-tracking typically relies on constrained visual contexts in particular goal-oriented contexts, viewing a small array of objects on a computer screen and performing some overt decision or identification. Eyetracking paradigms that use pictures as a measure of word or sentence comprehension are sometimes touted as ecologically invalid because pictures and explicit tasks are not always present during language comprehension. This study compared the comprehension of sentences with two different grammatical forms: the past progressive (e.g., was walking), which emphasizes the ongoing nature of actions, and the simple past (e.g., walked), which emphasizes the end-state of an action. The results showed that the distribution and timing of eye movements mirrors the underlying conceptual structure of this linguistic difference in the absence of any visual stimuli or task constraint: Fixations were shorter and saccades were more dispersed across the screen, as if thinking about more dynamic events when listening to the past progressive stories. Thus, eye movement data suggest that visual inputs or an explicit task are unnecessary to solicit analog representations of features such as movement, that could be a key perceptual component to grammatical comprehension. |
Heeju Hwang; Elsi Kaiser The role of the verb in grammatical function assignment in English and Korean Journal Article In: Journal of Experimental Psychology: Learning, Memory, and Cognition, vol. 40, no. 5, pp. 1363–1376, 2014. @article{Hwang2014, One of the central questions in speech production is how speakers decide which entity to assign to which grammatical function. According to the lexical hypothesis (e. g., Bock & Levelt, 1994), verbs play a key role in this process (e. g., "send" and "receive" result in different entities being assigned to the subject position). In contrast, according to the structural hypothesis (e. g., Bock, Irwin, & Davidson, 2004), grammatical functions can be assigned based on a speaker's conceptual representation of an event, even before a particular verb is chosen. In order to examine the role of the verb in grammatical function assignment, we investigated whether English and Korean speakers exhibit semantic interference effects for verbs during a scene description task. We also analyzed speakers' eye movements during production. We found that English speakers exhibited verb interference effects and also fixated the action/verb region before the subject region. In contrast, Korean speakers did not show any verb interference effects and did not fixate the action/verb region before the subject region. Rather, in Korean, looks to the action/verb region sharply increased following looks to the object region. The findings provide evidence for the lexical hypothesis for English and are compatible with the structural hypothesis for Korean. We suggest that whether the verb is retrieved before speech onset depends on the role that the verb plays in grammatical function assignment or structural choice in a particular language. |
Guilhem Ibos; David J. Freedman Dynamic integration of task-relevant visual features in posterior parietal cortex Journal Article In: Neuron, vol. 83, no. 6, pp. 1468–1480, 2014. @article{Ibos2014, The primate visual system consists of multiple hierarchically organized cortical areas, each specialized for processing distinct aspects of the visual scene. For example, color and form are encoded in ventralpathway areas such as V4 and inferior temporal cortex, while motion is preferentially processed in dorsal pathway areas such as the middletemporal area. Such representations often need to be integrated perceptually to solve tasks that depend on multiple features. We tested the hypothesis that the lateral intraparietal area (LIP) integrates disparate task-relevant visual features by recording from LIP neurons in monkeys trained to identify target stimuli composed of conjunctionsof color and motion features. We show that LIP neurons exhibit integrative representations of both color and motion features when they are taskrelevant and task-dependent shifts of both direction and color tuning. This suggests that LIP plays a role in flexibly integrating task-relevant sensory signals. |
Jörn M. Horschig; Ole Jensen; Martine R. Schouwenburg; Roshan Cools; Mathilde Bonnefond Alpha activity reflects individual abilities to adapt to the environment Journal Article In: NeuroImage, vol. 89, pp. 235–243, 2014. @article{Horschig2014, Recent findings suggest that oscillatory alpha activity (7-13. Hz) is associated with functional inhibition of sensory regions by filtering incoming information. Accordingly the alpha power in visual regions varies in anticipation of upcoming, predictable stimuli which has consequences for visual processing and subsequent behavior. In covert spatial attention studies it has been demonstrated that performance correlates with the adaptation of alpha power in response to explicit spatial cueing. However it remains unknown whether such an adaptation also occurs in response to implicit statistical properties of a task. In a covert attention switching paradigm, we here show evidence that individuals differ on how they adapt to implicit statistical properties of the task. Subjects whose behavioral performance reflects the implicit change in switch trial likelihood show strong adjustment of anticipatory alpha power lateralization. Most importantly, the stronger the behavioral adjustment to the switch trial likelihood was, the stronger the adjustment of anticipatory posterior alpha lateralization. We conclude that anticipatory spatial attention is reflected in the distribution of posterior alpha band power which is predictive of individual detection performance in response to the implicit statistical properties of the task. |
Dayana Hristova; Matej Guid; Ivan Bratko Assessing the difficulty of chess tactical problems Journal Article In: International Journal on Advances in Intelligent Systems, vol. 7, no. 3-4, pp. 728–738, 2014. @article{Hristova2014, We investigate experts' ability to assess the difficulty of a mental task for a human. The final aim is to find formalized measures of difficulty that could be used in automated assessment of the difficulty of a task. In experiments with tactical chess problems, the experts' estimations of difficulty are compared to the statistic-based difficulty ratings on the Chess Tempo website. In an eye tracking experiment, the subjects' solutions to chess problems and the moves that they considered are analyzed. Performance data (time and accuracy) are used as indicators of subjectively perceived difficulty. We also aim to identify the attributes of tactical positions that affect the difficulty of the problem. Understanding the connection between players' estimation of difficulty and the properties of the search trees of variations considered is essential, but not sufficient, for modeling the difficulty of tactical problems. Our findings include that (a) assessing difficulty is also very difficult for human experts, and (b) algorithms designed to estimate difficulty should interpret the complexity of a game tree in the light of knowledge-based patterns that human players are able to detect in a chess problem. |
S. Huber; Alon Mann; Hans-Christoph Nuerk; Korbinian Moeller Cognitive control in number magnitude processing: Evidence from eye-tracking Journal Article In: Psychological Research, vol. 78, no. 4, pp. 539–548, 2014. @article{Huber2014, The unit-decade compatibility effect describes longer response times and higher error rates for incompatible (e.g., 37_52) than compatible (e.g., 42_57) number comparisons. Recent research indicated that the effect depends on the percentage of same-decade filler items. In the present study, we further examined this relationship by recording participants' eye-fixation behaviour. In four conditions, participants had to compare item sets with different filler item types (i.e., same-decade and same-unit filler items) and different numbers of same-decade filler items (i.e., 25, 50, and 75 %). We found a weaker unit-decade compatibility effect with most fixations on tens in the condition with same-unit filler items. Moreover, the compatibility effect increased with the percentage of same-decade filler items which was accompanied by less fixations on tens and more fixations on units. Thus, our study provides first eye-tracking evidence for the influence of cognitive control in number processing. |
S. Huber; Korbinian Moeller; Hans-Christoph Nuerk Adaptive processing of fractions - Evidence from eye-tracking Journal Article In: Acta Psychologica, vol. 148, pp. 37–48, 2014. @article{Huber2014a, Recent evidence indicated that fraction pair type determined whether a particular fraction is processed holistically, componentially or in a hybrid manner. Going beyond previous studies, we investigated how participants adapt their processing of fractions not only to fraction type, but also to experimental context. To examine adaptation in fraction processing, we recorded participants' eye-fixation behaviour in a fraction magnitude comparison task.Participants' eye fixation behaviour indicated componential processing of fraction pairs with common components for which the decision-relevant components are easy to identify. Importantly, we observed that fraction processing was adapted to experimental context: Evidence for componential processing was stronger, when experimental context allowed valid expectations about which components are decision-relevant.Taken together, we conclude that fraction processing is adaptive beyond the comparison of different fraction types, because participants continuously adjust to the experimental context in which fractions are processed. |
Stefan Huber; Elise Klein; Klaus Willmes; Hans-Christoph Nuerk; Korbinian Moeller In: Frontiers in Human Neuroscience, vol. 8, pp. 172, 2014. @article{Huber2014b, Decimal fractions comply with the base-10 notational system of natural Arabic numbers. Nevertheless, recent research suggested that decimal fractions may be represented differently than natural numbers because two number processing effects (i.e., semantic interference and compatibility effects) differed in their size between decimal fractions and natural numbers. In the present study, we examined whether these differences indeed indicate that decimal fractions are represented differently from natural numbers. Therefore, we provided an alternative explanation for the semantic congruity effect, namely a string length congruity effect. Moreover, we suggest that the smaller compatibility effect for decimal fractions compared to natural numbers was driven by differences in processing strategy (sequential vs. parallel). To evaluate this claim, we manipulated the tenth and hundredth digits in a magnitude comparison task with participants' eye movements recorded, while the unit digits remained identical. In addition, we evaluated whether our empirical findings could be simulated by an extended version of our computational model originally developed to simulate magnitude comparisons of two-digit natural numbers. In the eye-tracking study, we found evidence that participants processed decimal fractions more sequentially than natural numbers because of the identical leading digit. Importantly, our model was able to account for the smaller compatibility effect found for decimal fractions. Moreover, string length congruity was an alternative account for the prolonged reaction times for incongruent decimal pairs. Consequently, we suggest that representations of natural numbers and decimal fractions do not differ. |
Stefanie Mueller; Katja Fiehler Effector movement triggers gaze-dependent spatial coding of tactile and proprioceptive-tactile reach targets Journal Article In: Neuropsychologia, vol. 62, no. 1, pp. 184–193, 2014. @article{Mueller2014, Reaching in space requires that the target and the hand are represented in the same coordinate system. While studies on visually-guided reaching consistently demonstrate the use of a gaze-dependent spatial reference frame, controversial results exist in the somatosensory domain. We investigated whether effector movement (eye or arm/hand) after target presentation and before reaching leads to gaze-dependent coding of somatosensory targets. Subjects reached to a felt target while directing gaze towards one of seven fixation locations. Touches were applied to the fingertip(s) of the left hand (proprioceptive-tactile targets) or to the dorsal surface of the left forearm (tactile targets). Effector movement was varied in terms of movement of the target limb or a gaze shift. Horizontal reach errors systematically varied as a function of gaze when a movement of either the target effector or gaze was introduced. However, we found no effect of gaze on horizontal reach errors when a movement was absent before the reach. These findings were comparable for tactile and proprioceptive-tactile targets. Our results suggest that effector movement promotes a switch from a gaze-independent to a gaze-dependent representation of somatosensory reach targets. |
Stefanie Mueller; Katja Fiehler Gaze-dependent spatial updating of tactile targets in a localization task Journal Article In: Frontiers in Psychology, vol. 5, pp. 66, 2014. @article{Mueller2014a, There is concurrent evidence that visual reach targets are represented with respect to gaze. For tactile reach targets, we previously showed that an effector movement leads to a shift from a gaze-independent to a gaze-dependent reference frame. Here we aimed to unravel the influence of effector movement (gaze shift) on the reference frame of tactile stimuli using a spatial localization task (yes/no paradigm). We assessed how gaze direction (fixation left/right) alters the perceived spatial location (point of subjective equality) of sequentially presented tactile standard and visual comparison stimuli while effector movement (gaze fixed/shifted) and stimulus order (vis-tac/tac-vis) were varied. In the fixed-gaze condition, subjects maintained gaze at the fixation site throughout the trial. In the shifted-gaze condition, they foveated the first stimulus, then made a saccade toward the fixation site where they held gaze while the second stimulus appeared. Only when an effector movement occurred after the encoding of the tactile stimulus (shifted-gaze, tac-vis), gaze similarly influenced the perceived location of the tactile and the visual stimulus. In contrast, when gaze was fixed or a gaze shift occurred before encoding of the tactile stimulus, gaze differentially affected the perceived spatial relation of the tactile and the visual stimulus suggesting gaze-dependent coding of only one of the two stimuli. Consistent with previous findings this implies that visual stimuli vary with gaze irrespective of whether gaze is fixed or shifted. However, a gaze-dependent representation of tactile stimuli seems to critically depend on an effector movement (gaze shift) after tactile encoding triggering spatial updating of tactile targets in a gaze-dependent reference frame. Together with our recent findings on tactile reaching, the present results imply similar underlying reference frames for tactile spatial perception and action. |