|
|
مروری بر پژوهشهای نوایِ گفتار عاطفی انجام شده با استفاده از تکنیک پتانسیل وابسته به رویداد و تصویربرداری تشدید مغناطیسی کارکردی و معرفی پایگاه دادۀ گفتارِ عاطفی زبان فارسی (persian esd)
|
|
|
|
|
نویسنده
|
جودی نیره ,کشتیاری نیلوفر
|
منبع
|
تازه هاي علوم شناختي - 1403 - دوره : 26 - شماره : 1 - صفحه:91 -108
|
چکیده
|
مقدمه: نوای گفتار ابزارِ انتقال اطلاعات زبانی و عاطفی است که بدون آنها ارتباط کلامیِ موثر امکانپذیر نیست. پژوهش حاضر میکوشد با مرور پژوهشهای انجام شده در زمینۀ نوای عاطفی با به کارگیری روش های تصویربرداری عصبی و ثبت سیگنالهای مغزی وابسته به رویداد و با توجه به یافتههای این پژوهشها گزارش کند که هنگام پردازش نوای عاطفی کدام مناطق مغز فعال هستند و سازوکار درک نوای عاطفی چگونه است.روش کار: برای دست یابی به این هدف، مطالعاتی که از سال 2005 تا 2021 در پایگاههای اطلاعاتی معتبر pubmed وgoogle scholar منتشر شده بودند و بر پایۀ دو روش نام برده و نیز با استفاده از جملات به عنوان محرک، آزمودنیهای سالم را بررسی کرده بودند، انتخاب و به صورت نظاممند مرور شدند. در بخش دوم این نوشتار برای نخستین بار به زبان فارسی مراحل طراحی، ساخت و رواسازی «پایگاه دادۀ گفتارِ عاطفی زبان فارسی (persian esd)» شرح داده شده است. یافتهها: فعالیت مغز هنگام پردازش نوای گفتار به منطقهای خاص در نیمکرۀ راست مغز محدود نمیشود. بلکه در نواحی مشخصی فعالیت همزمان دو نیمکره مشاهده میگردد. پردازش نوای گفتار زبانشناختی و عاطفی سازوکارهای متفاوتی دارند.نتیجه گیری: سازوکار عصبی مغز در هنگام پردازش نوای گفتار به عواملی مانند درجۀ دشواری تکلیف آزمایش، کیفیت محرکها و چگونگی طراحی و اجرای آزمایش بستگی دارد. دسترسی به ابزارهای پژوهشی رواسازی شده برای ارزیابی نوای گفتار از ملزومات مطالعۀ عصبشناختی گفتار عاطفی است.
|
کلیدواژه
|
نوای گفتار عاطفی، پتانسیل وابسته به رویداد (erp)، تصویربرداری تشدید مغناطیسی کارکردی (fmri)، پایگاه دادۀ گفتار عاطفی زبان فارسی (persian esd)
|
آدرس
|
دانشگاه علامه طباطبائی, ایران, دانشگاه آزاد اسلامی واحد برلین, آلمان
|
|
|
|
|
|
|
|
|
|
|
a review of emotional prosody research based on functional magnetic resonance imaging and event-related potential techniques and an introduction to the persian emotional speech database (persian esd)
|
|
|
Authors
|
joodi nayereh ,keshtiari niloofar
|
Abstract
|
introductioneffective verbal communication is impossible without prosody, which transmits both linguistic and emotional information. this study aims to identify the brain regions involved in emotional prosody processing and uncover the mechanisms involved in understanding emotional prosody. methodsusing databases such as pubmed and google scholar, we searched for keywords such as emotional prosody, speech processing, event-related potentials (erps), and functional magnetic resonance imaging (fmri) published between 2005 and 2021 on emotional prosody processing in healthy adult subjects’ brains. the search results for these keywords included more than 100 articles related to emotional prosody processing, an essential part of processing and understanding emotions. the subjects in all selected studies were healthy and right-handed regarding hearing, vision, and neuropsychology, and only studies using sentences as stimuli were included. the authors include only information regarding the present study questions in tables 1 and 2, and they avoid expressing technical details related to imaging and recording the brain signal, processing the images and signals, and avoiding issues related to statistical processing steps for the results of each study.resultsstudying emotional prosody provided models for the stages of emotional tone processing. as one of these models that has been considered in subsequent research, the model of schirmer and kotz (11) was used to obtain more information about the brain areas involved in processing emotional speech sounds (2, 19). in the ear, brain stem, thalamus, and primary auditory cortex, speech sounds are processed and decoded in three stages. the first step is to extract phonological characteristics from prosodic cues in the right auditory cortex. in the second step, the right hemisphere’s posterior/posterior parts (sts) represent meaningful sequences of phonological elements. evaluation and cognitive interpretation of expressed emotions constitute the third step, which produces simultaneous activity in two hemispheres of the frontal cortex (figures 1 and 2 show these steps). figure 1. model for the processing of emotional prosody (11, p. 25)notably, according to the socio-cultural characteristics of their native language, the speakers use these phonological characteristics (such as fundamental frequency, intensity or loudness of the voice, speed of speech, and voice quality (breathy, whispering, sharp, growling, and the like) to produce and use emotional and attitudinal understanding (5, 24, 25).figure 2. three successive steps for the processing of emotional prosody (20, p.261)conclusionprocessing speech sounds is confined to just one hemisphere of the brain. it is also crucial to understand that various factors, such as the difficulty level of the test, the quality of the stimuli, and the design and execution of the test, influence neural mechanisms. given the functional complexity of linguistic prosody, analyzing the neural structure of emotional prosody seems more straightforward. clinical and neurological research has shown that these two types of prosody process differently in the brain. therefore, speech prosody should be examined according to its multi-level division and application rather than viewed as a general concept (5).given the intricate and multifaceted nature of language, it appears essential to analyze the emotional prosody of speech independently, owing to its complexity and diverse characteristics. although it provides detailed instructions on how to create verbal-emotional communication, attention must also be paid to how syntax, meaning, and pragmatics influence the tone of speech (both emotional and linguistic).for linguistic studies, particularly neurological studies, linguistic data are examined outside their natural context due to technical limitations. in order to obtain accurate information about how language is processed and understood, it is essential to carefully design the appropriate task and select and formulate the appropriate stimuli. therefore, a standard form of emotional prosody in any language is necessary to research emotional prosody (8). as manipulated stimuli are usually used in these experiments, it is also necessary to identify the effective phonological characteristics (fundamental frequency, duration, intensity, and the like) associated with emotional, attitudinal, and motivational prosody in each language using appropriate perceptual and behavioral research.in order to support researchers conducting proper research on persian language processing, the steps involved in creating and maintaining the “persian emotional speech database” will be explained. these databases can help researchers design and conduct neurological experiments based on erp or fmri techniques to investigate how the emotional prosody of persian speech is processed and perceived in the brains of persian speakers. the database was previously designed and produced at freie universität berlin in collaboration with a research group (26). however, this information was only available in english. this study aimed to build an authentic database of emotional speech in persian.the database contains a set of 90 validated novel persian sentences classified into five basic emotional categories (anger, disgust, fear, happiness, and sadness), as well as a neutral category. these sentences were validated in two experiments by a group of 1,126 native persian speakers. the sentences were articulated by two native persian speakers (one male, one female) in three conditions: 1) congruent (emotional lexical content articulated in a congruent emotional voice), 2) incongruent (neutral sentences articulated in an emotional voice), and 3) baseline (all emotional and neutral sentences articulated in neutral voice). the speech materials comprise about 470 sentences. the validity of the database was evaluated by a group of 34 native speakers in a perception test. utterances recognized better than five times chance performance (71.4 %) were regarded as valid portrayals of the target emotions. acoustic analysis of the valid emotional utterances revealed differences in pitch, intensity, and duration, attributes that may help listeners to correctly classify the intended emotion. the database is designed to be used as a reliable material source (for both text and speech) in future cross-cultural or cross-linguistic studies of emotional speech, and it is available for academic research purposes free of charge. this tool can benefit research in various fields, including neurology of language, psychology of language, clinical linguistics, speech therapy, and speech synthesis. to access the database, please contact the second author.ethical considerations compliance with ethical guidelinesthe present study has a review nature. besides, there is no doubt that all the research introduced in this review had a valid code of ethics.authors’ contributionsfirst author: drafting the article, revising it, and being accountable for all aspects of the research. second author: designing, constructing, and validating the persian emotional speech database; writing the second part of the article; revising and correcting the entire article.fundingthe german research foundation (dfg) awarded a scholarship to the second author for designing the persian language emotional speech database. acknowledgmentsthanks to shahla raghibdoust for her assistance in writing this paper’s the first part (review).conflict of interestthe authors have no conflict of interest.
|
Keywords
|
emotional prosody ,event-related potential (erp) ,functional magnetic resonance imaging (fmri) ,persian emotional speech database (persian esd)
|
|
|
|
|
|
|
|
|
|
|