Cookies on this website

We use cookies to ensure that we give you the best experience on our website. If you click 'Accept all cookies' we'll assume that you are happy to receive all cookies and you won't see this message again. If you click 'Reject all non-essential cookies' only necessary cookies providing core functionality such as security, network management, and accessibility will be enabled. Click 'Find out more' for information on how to change your cookie settings.

We examined the time-courses of the crossmodal semantic congruency effects elicited by naturalistic sounds or spoken words on the processing of visual pictures and printed words. Auditory primes were presented at seven stimulus onset asynchronies (SOAs) with respect to the visual targets, ranging from auditory leading by 1000 ms to auditory lagging by 250 ms. Participants made speeded categorization judgments (living vs. non-living) regarding the visual targets. Three common effects were observed for both types of visual targets: Both naturalistic sounds and spoken words induced a slow facilitatory effect when leading by 250 ms or more in the congruent condition, but induced a rapid inhibitory effect when leading by 250 ms or less in the incongruent condition. Only spoken words that did not match the visual targets elicited an additional inhibitory effect when leading by 100 ms or when presented simultaneously. As compared to non-linguistic stimuli, the priming effects associated with linguistic stimuli occurred over a wider range of SOAs, and the interactions occurred at a more specific level of the category hierarchy (i.e., the basic level) than required by the task. These results suggest different routes to semantic access: Linguistic stimuli accessed meaning mediated by the lexical representation that activated semantic representations at the corresponding level of the category hierarchy precisely, whereas non-linguistic stimuli accessed their meaning rapidly and directly, but first accessed meaning at a higher-level of semantic hierarchy. We propose a comprehensive framework to provide a dynamic view regarding how meaning is extracted during the processing of linguistic and non-linguistic stimuli presented in the visual and auditory modalities, therefore contributing to our understanding of the human semantic system.

Type

Conference paper