Armitage, James and Eerola, Tuomas (2022) 'Cross-modal Transfer of Valence or Arousal from Music to Word Targets in Affective Priming?', Auditory Perception & Cognition, 5 . pp. 192-210.
This registered report considers how emotion induced in an auditory modality (music) can influence affective evaluations of visual stimuli (words). Specifically, it seeks to determine which emotional dimension is transferred across modalities – valence or arousal – or whether the transferred dimension depends on the focus of attention (feature-specific attention allocation). Two experiments were carried out. The first was an affective priming paradigm that will allow for the orthogonal manipulation of valence and arousal in both the words and music, alongside a manipulation to direct participants’ attention to either the valence or the arousal dimension. Secondly, a lexical decision task allowed cross-modal transfer of valence and arousal to be probed without the focus of participants’ attention being manipulated. Congruence effects were present in the affective priming task – valence was transferred in both the valence and arousal tasks, whereas arousal was transferred in the arousal task only. Contrary to predictions, the lexical decision task did not exhibit any congruence effects.
|Full text:||(VoR) Version of Record|
Available under License - Creative Commons Attribution Non-commercial No Derivatives 4.0.
Download PDF (2928Kb)
|Publisher Web site:||https://doi.org/10.1080/25742442.2022.2087451|
|Publisher statement:||This is an Open Access article distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivatives License (http://creativecommons.org/licenses/by-nc-nd/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited, and is not altered, transformed, or built upon in any way.|
|Date accepted:||23 April 2022|
|Date deposited:||24 August 2022|
|Date of first online publication:||23 June 2022|
|Date first made open access:||24 August 2022|
Save or Share this output
|Look up in GoogleScholar|