Affective Auditory Stimulus Database: An Expanded Version of The International Affective Digitized Sounds (IADS-E)
Affective Auditory Stimulus Database: An Expanded Version of The International Affective Digitized Sounds (IADS-E)
https://doi.org/10.3758/s13428-018-1027-6
Abstract
Using appropriate stimuli to evoke emotions is especially important for researching emotion. Psychologists have provided several
standardized affective stimulus databases—such as the International Affective Picture System (IAPS) and the Nencki Affective
Picture System (NAPS) as visual stimulus databases, as well as the International Affective Digitized Sounds (IADS) and the
Montreal Affective Voices as auditory stimulus databases for emotional experiments. However, considering the limitations of the
existing auditory stimulus database studies, research using auditory stimuli is relatively limited compared with the studies using
visual stimuli. First, the number of sample sounds is limited, making it difficult to equate across emotional conditions and
semantic categories. Second, some artificially created materials (music or human voice) may fail to accurately drive the intended
emotional processes. Our principal aim was to expand existing auditory affective sample database to sufficiently cover natural
sounds. We asked 207 participants to rate 935 sounds (including the sounds from the IADS-2) using the Self-Assessment Manikin
(SAM) and three basic-emotion rating scales. The results showed that emotions in sounds can be distinguished on the affective
rating scales, and the stability of the evaluations of sounds revealed that we have successfully provided a larger corpus of natural,
emotionally evocative auditory stimuli, covering a wide range of semantic categories. Our expanded, standardized sound sample
database may promote a wide range of research in auditory systems and the possible interactions with other sensory modalities,
encouraging direct reliable comparisons of outcomes from different researchers in the field of psychology.
Keywords Emotion . Affective auditory stimuli . Affective ratings . International Affective Digitized Sounds . SAM
certain types of emotion such as affective pictures (e.g., according to the Self-Assessment Manikin (SAM; Bradley
Buodo et al., 2017; Lang, Greenwald, Bradley, & Hamm, & Lang, 1994).
1993), emotional facial expressions (e.g., Esslen, Pascual- However, previous studies have exposed that IADS-2, as a
Marqui, Hell, Kochi, & Lehmann, 2004; Shuman, Clark- standardized international affective stimulus database, had
Polner, Meuleman, Sander, & Scherer, 2015), and written af- limitations and deficiencies (Choi et al., 2016; da Silva &
fective words (e.g., Schindler & Kissler, 2016). Among other Backs, 2015). First, as compared to most visual stimulus da-
senses (i.e., audition, touch, taste, and olfaction), auditory tabases, the number of stimuli in IADS-2 is insufficient (only
stimuli also profoundly provoke human emotions (Redondo, 167 sounds; the International Affective Picture System [IAPS]
Fraga, Padrón, & Piñeiro, 2008). However, research on audi- has 1,182 pictures: Lang, Bradley, & Cuthbert, 1997, 2008;
tory system is much less frequent than on visual system the Affective Norms for English Words [ANEW] has 1,034
(Gerdes et al., 2014). One of the potential reasons would be words: Bradley & Lang, 2007b). The small number of sound
that visual stimuli are easier to manipulate and control as com- stimuli means that the distribution of IADS-2 in the
pared with auditory stimuli (Bradley & Lang, 2000). For an bidimensional affective space of valence and arousal is un-
instance, in contrast with static pictures of facial expressions even as compared to IAPS (see Fig. 1a and b, respectively).
or written words, auditory stimuli are dynamic over time We counted the number of sounds for the IADS-2 in each
(Banse & Scherer, 1996; Droit-Volet, Ramos, Bueno, & quadrant of arousal–valence space, with the origin at ratings
Bigand, 2013; Juslin & Laukka, 2003), requiring continuous of 5. Consequently, the numbers of sounds from the first quad-
perception, integration, and interpretation for recognition rant to the fourth quadrant are 50, 31, 16, and 70, respectively,
(Schirmer & Kotz, 2006). whereas the numbers of pictures for the IAPS in each quadrant
In recent decades, researches on emotion using auditory are 225, 420, 225, and 312, respectively. The biased distribu-
stimuli have gradually increased following the advancement tion of sounds makes it difficult to choose the same number of
and growing popularity of digital information storage and sounds from different emotional conditions, and it is extreme-
software to conveniently store and manipulate digitized data; ly difficult to keep experimental conditions balanced like they
this has made experimental control of auditory stimuli increas- can be with the IAPS. Second, the number of sounds in each
ingly feasible (Bradley & Lang, 2000). For instance, Fabiani, semantic category is varied: animals (N = 18), people (N =
Kazmerski, Cycowicz, and Friedman (1996) developed a 52), nature (N = 7), daily routine sounds (N = 28), transport (N
large battery of brief (400 ms), seemingly neutral auditory = 15), electric sounds (N = 5), breaking sounds (N = 4), music
stimuli (i.e., animals, noises, artificial sounds, etc.) to investi- (N = 12), and scenarios (N = 26). The classification was eval-
gate novelty effects in the event-related potential oddball par- uated by ten participants in a pilot study. Considering the
adigm. Czigler, Cox, Gyimesi, and Horváth (2007) had irrelevant number of sounds, it is also difficult to select exper-
attempted to investigate brain processes related to unpleasant imental materials from each semantic category equally.
auditory stimuli. Armony, Aubé, Angulo-Perkins, Peretz, and In addition to the IADS, other sets of auditory stimulus
Concha (2015) performed an fMRI experiment to compare the databases, including the Montreal Affective Voices database
neural activities evoked by music and voice. Nevertheless, (Belin, Fillion-Bilodeau, & Gosselin, 2008), musical excerpts
there are flaws in these experimental materials. Most of the (Vieillard et al., 2008), Portuguese sentences and
auditory stimuli used in emotional research were selected by pseudosentences for research on emotional prosody (Castro
researchers a priori or from a limited number of participants’ & Lima, 2010), and vocal emotional stimuli in Mandarin
ratings in a preliminary experiment. Such a limited number Chinese (Liu & Pell, 2012) have also been developed.
and type of stimuli make it difficult to compare the outcomes However, the materials in these databases are limited to music
and to replicate them across research labs (Bradley & Lang, pieces or recorded vocalizations. Studies that are based on
2007a). these artificial databases may fail to accurately predict the
To address these problems and encourage research human emotion process, because the sounds we hear in daily
assessing basic and applied problems in psychological sci- life are much more extensive. Therefore, inventing a standard-
ence, the NIMH Center for Emotion and Attention at the ized natural emotional auditory stimulus database that con-
University of Florida has developed a standardized auditory tains sufficient stimuli for emotional research is urgently re-
stimulus database named the International Affective Digitized quired (Gerdes et al., 2014).
Sounds (Bradley & Lang, 1999 [IADS-1]; Bradley & Lang, In this study, we aimed to expand the IADS-2 (Bradley &
2007a [IADS-2]). The latest version, IADS-2 (Bradley & Lang, 2007a) and provide a high-quality standardized, emo-
Lang, 2007a), consists of 167 digitally recorded natural tionally evocative auditory stimuli database with contents
sounds that are common in daily life, such as a baby laughing, from across a wider range of semantic categories. We based
someone cooking, a thunderstorm, erotic sounds, and so on. our database on the IADS-2 because it is the only auditory
Each sound lasts 6 s and is characterized by the affective database that includes a wide range of semantic categories
dimensions of valence, arousal, and dominance or control and, as compared with other auditory databases, is also the
Behav Res (2018) 50:1415–1429 1417
Method
Participants
(e.g., rain, writing, laughing, and barking), and some automation. The software, running on a personal comput-
were composed by a composer. Depending on the dura- er, controlled the presentation of instructions on the com-
tion of the sound (sounds less than 1.5 s were excluded), puter display as well as the presentation of sounds via a
naturalness, and the emotional ratings from a preliminary speaker (GX-70HD II, ONKYO Corp., Japan). It also
experiment, after careful filtering, an initial observation registered the ratings that participants entered using the
revealed 1,226 sounds, from which 768 sounds were keyboard.
selected.
We manipulated the 768 newly collected sounds to 6 s with Emotional rating measurements
the SoundEngine Free Version 5.21 software (Coderium.,
Inc., Japan) as this is the IADS-2’s standard duration. (1) Self-Assessment Manikin (SAM): This system for non-
Because differences in physical properties may influence verbal pictorial assessment, devised by Lang (1980)
emotional ratings (Lolli, Lewenstein, Basurto, Winnik, & (also seen in Bradley & Lang, 1994) to measure
Loui, 2015), we edited three physical properties (frequency, emotional ratings in the three dimensions of valence,
44.1 kHz; bit rate, 1,411 kbps; and channels, two) equally for arousal, and dominance, was used. The SAM scale
all sounds (including those from IADS-2). we used contained five affective graphical figures,
with a space between each two consecutive pictures,
(2) Experimental materials: When the new sound stimuli defining a 9-point scale for each dimension (Fig. 2):
were added to those from the IADS-2, we had collected valence (ranging from 1, depicted by a frowning,
935 sound stimuli for this experiment. The sounds were unhappy figure, to 9, a smiling, happy figure), arous-
preliminarily divided into ten semantic categories—ani- al (ranging from 1, depicted by a relaxed, sleepy
mals (N = 54), people (N = 74), nature (N = 70), daily figure, to 9, an excited, wide-eyed figure), and dom-
routine sounds (N = 187), transport (N = 59), electric inance (ranging from 1, depicted by a small figure
sounds (N = 64), sound effects (N = 171), breaking representing being controlled, to 9, a large figure
sounds (N = 56), music (N = 170), and scenarios (N = representing being in control).
30)—by ten participants who did not participate in the (2) Basic emotion rating scales: Three basic emotion rating
main experiment. Each sound’s physical property values scales—happiness, sadness, and fear—were presented as
were calculated using the Amadeus Pro sound editing supplementary evaluations. The ratings were measured
version 2.2.2 software (HairerSoft., Inc. UK), shown in from 1 point, for not at all, to 9 points, for extremely, and
the fourth sheet of the supplementary materials were shown by words. Both the SAM scales and the
(Sounds2nd). The peak sound intensity at presentation basic emotion rating scales were presented in a pseudo-
ranged from 50.4 to 88 dB(A), as measured using a NL- random order, to control for order effects (Castonguay,
42 sound level meter (RION Co., Ltd., Japan); the max- Sabiston, Crocker, & Mack, 2014; Krosnick & Presser,
imum level (Lmax), the minimum level (Lmin), the equiv- 2010). All ratings were further analyzed using the statis-
alent continuous level (Leq), and the exposure level (LE) tical package IBM SPSS Version 20.0 software (IBM
are shown on the fifth sheet of the supplementary mate- Corp., US).
rials (Sounds3rd). All sounds were presented with a
prompt fixed volume and generally varied according to
the respective natural volumes in the environment.
Measurement of mental state conducted at the participant’s own pace. The average duration
of one trial was approximately 30 s. Participants were made to
To identify the participants’ psychological states when they rate all 52 sounds in one block at a stretch. The order of the two
participated in the experiment, the Japanese version (Shimizu blocks was counterbalanced among the participants who rated the
& Imae, 1981) of the State–Trait Anxiety Inventory (STAI; sounds in the same group. A 5-min break was inserted between
Spielberger, Gorsuch, & Lushene, 1970) was used. This ques- the two experimental blocks to reduce the participants’ fatigue.
tionnaire is used to diagnose and distinguish a participant’s Each experimental session lasted approximately 1 h 10 min.
level of anxiety at the time of filling out the questionnaire.
The STAI has two parts: state anxiety and trait anxiety. Our
study used the state anxiety part (STAI-S), which consists of Results and discussion
20 statements that measure anxiety according to the situation.
All items are rated on a four-point scale from almost never to Reliability
almost always. A higher score indicates a higher level of
anxiety. We calculated the STAI-S scores for all participants. The
means and standard deviations of the STAI-S scales by gender
Procedure were 40.20 (SD = 5.57) for female participants and 41.34 (SD
= 6.28) for male participants. Mean scores of STAI-S in this
The participants were guided into a soundproof experimental study were in the normal range (32–42 points for male partic-
booth and sat in front of a computer monitor at a distance of ipants and 31–41 points for female participants) of the mental
approximately 70 cm. After receiving a detailed briefing on state (Iwata & Higuchi, 2000). These results suggested that
the experimental purpose and providing informed consent, none of the participants showed extreme uneasiness on the
two online pre-questionnaire investigations, the STAI-S day of the experiment.
(Spielberger et al., 1970), and the Edinburgh Handedness Ratings for each sound of the IADS-E are presented in the
Inventory (Oldfield, 1971), were administered. The investiga- first sheet to third sheet of the supplementary materials. The
tions lasted approximately 10 min. supplementary materials list the mean values (M) and standard
On the affective rating task, first of all, six practice sounds deviations (SD) for arousal (Aro), valence (Val), dominance
that were excluded from the experimental stimuli but had the (Dom), fear (Fear), happiness (Hap), and sadness (Sad) for
same characteristics as the experimental stimuli were used to each of the 935 sounds, considering the total sample (All) and
illustrate the evaluation task and to allow practice making rat- the male and female subsamples separately. We calculated the
ings. After participants had listened to each sound, six indepen- Cronbach’s alpha for all sounds using ratings from 22 partic-
dent emotional rating scales (the SAM and the basic-emotion ipants (α = .95 for valence, α = .92 for arousal, α = .81 for
rating scales) were provided for ratings. The practice session dominance, α = .92 for fear, α = .92 for happiness, and α = .82
lasted approximately 3 min and was conducted to ensure that for sadness) to validate the internal consistency as the previous
the participants were fully familiar with the experiment task. studies did (i.e., Drače, Efendić, Kusturica, & Landzo, 2013;
During the experimental session, each participant was random- Dufey, Fernández, & Mayol, 2011; Suk & Irtel, 2010). The
ly assigned to one of the nine sound groups. The experimental results suggested that the affective ratings for each sound in
task was the same as in the practice session. Each trial was our study were reliable (Iacobucci & Duhachek, 2003). The
Fig. 3 Numbers of sound stimuli in the two-dimensional affective space defined by mean valence and arousal, for (a) IADS-E and (b) IADS-2
1420 Behav Res (2018) 50:1415–1429
Fig. 4 Distributions of overall mean values for the IADS-E sounds in the valence–arousal affective dimensions (a), valence–dominance affective
dimensions (b), and arousal–dominance affective dimensions (c)
coefficient of variation (CV) was also calculated, to evaluate SDmVal = 1.46, SDmDom = 2.05) were obviously lower than in
the variability in the affective ratings; the average values of the the previous research (SDm Aro = 1.92, SDm Val = 1.76,
CVs for the present study across different affective dimen- SDmDom = 1.91). Thus, even though the number of the partic-
sions were CVVal = 31.23%, CVAro = 21.04%, CVDom = ipants was smaller than in the original IADS-2 study, the rat-
19.95%, which revealed a common pattern with the IADS-2 ings in our study were relatively stable.
(CVVal = 36.60%, CVAro = 19.74%, CVDom = 24.66%). We
also compared the standard deviations of the ratings for the Affective Ratings for the IADS-E
IADS-2 sounds between this study and the original one
(Bradley & Lang, 2007a). The mean standard deviations (1) Correlation analyses of valence and arousal We calcu-
(SDms) of the IADS-2 sounds in our study (SDmAro = 1.58, lated the number of sounds in arousal–valence
Behav Res (2018) 50:1415–1429 1421
Table 1 Descriptive statistics, calculated separately for all affective ratings scales in each semantic category
All scales range from 1 to 9, in which 1 represents negative, sleepy, being controlled, no fear, no happiness, and no sadness. Min = minimal value; Max =
maximal value; SD = standard deviation; N = number of sounds
happiness (254), fear (391), and sadness (18)—154 as However, unfortunately, the number of sounds that elic-
blended emotion sounds—happiness and fear (87), fear ited sadness was particularly small, and we measured
and sadness (52), and happiness and sadness (15)—and only three basic emotions in this study. These issues need
118 as undifferentiated sounds. For details on the basic to be addressed in future research.
emotion classification for each sound, please refer to the (4) Relationship between the affective dimensions and
supplementary materials. The correlation coefficient basic-emotion rating scales To ascertain the relation-
analysis showed strong negative relationships between ships between the three dimensions (i.e., valence, arous-
fear and happiness (r = − .71, R2 = .51, p < .001) and al, and dominance) and the three basic-emotion rating
sadness and happiness (r = − .49, R2 = .24, p < .001), as scales (i.e., fear, happiness, and sadness), Pearson corre-
well as a positive relationship between fear and sadness lation analyses were performed. The results suggested
(r = .55, R2 = .30, p < .001). The distribution of the that the ratings of valence (rFear = − .82, p < .001; rSad
IADS-E ratings on the three basic-emotion rating scales = − .37, p < .001) and dominance (rFear = − .86, p < .001;
can be seen in Fig. 5 (a 3-D scatterplot). The ratings of rSad = − .43, p < .001) have negative correlations with
basic emotions could assist researchers in choosing emo- fear and sadness, but positive correlations with happi-
tional stimuli more accurately from a discrete categorical ness (rVal = .84, p < .001; rDom = .57, p < .001). The
perspective according to their research purpose. ratings of arousal showed a positive correlation with fear
Behav Res (2018) 50:1415–1429 1423
Table 2 Correlation coefficients resulting from the affective rating scales in each semantic category
(r = .40, p < .001) and a weak negative correlation with rating scales in each semantic category are shown in Table 2.
sadness (r = − .10, p < .01), but no correlation with Although the correlation values are consistent when using all
happiness. As has been demonstrated in previous studies sounds, some subtle differences exist in each semantic catego-
(Choi et al., 2015; Stevenson & James, 2008), valence ry. Considering these results, researchers could choose sounds
and dominance showed positive relationships with hap- from every semantic category in line with their purpose.
piness but negative relationships with the other basic According to these results, we considered that all 935
emotion rating scales (i.e., fear and sadness), whereas sounds (including those from the IADS-2) provide a new ver-
the opposite pattern was observed for arousal. sion of the standardized affective auditory stimulus database
that is capable of effectively evoking emotion. However, the
number of sounds in the low valence–arousal quadrant is par-
Table 1 shows the descriptive statistics for the sound ratings ticularly scarce relative to the other three quadrants, thus sug-
on all affective ratings scales in each semantic category. The gesting that negative stimuli would be rated as much more
ratings in each semantic category have their unique features; for arousing, irrespective of whether they are auditory or visual
example, sounds in music tend to have more positive ratings, stimuli.
whereas breaking sounds tend to have more negative and The ratings of basic emotions could assist researchers in
arousing ratings. The correlation coefficients for the affective choosing emotional stimuli more accurately, according to their
1424 Behav Res (2018) 50:1415–1429
Table 3 Descriptive statistics, calculated separately for each dimension and each semantic category and gender
Male Female
All scales range from 1 to 9, in which 1 represents negative, sleepy, being controlled, no fear, no happiness, and no sadness. Min = minimal value; Max =
maximal value; SD = standard deviation; N = number of sounds
research purpose. Unfortunately, because of the long experi- = .00, did not reach statistical significance here, but the ratings
mental time, we measured only three basic emotions in this of valence, F(1, 1868) = 8.27, p = .004, η2 = .004; happiness,
study. Further research should measure other basic emotions, F(1, 1868) = 29.49, p < .001, η2 = .02; and sadness, F(1, 1868)
such as disgust, anger, and surprise. = 31.47, p < .001, η2 = .02, did show significant differences
between the genders. Table 3 shows the descriptive statistics
Sex difference in sounds ratings for valence, arousal, and dominance for the female and male
subsamples in each semantic category separately. The results
A multivariate analysis of variance (MANOVA) was conduct- suggest that Japanese men assessed the sounds as being much
ed with between-subjects factors to determine whether differ- more emotional than did the women; however, the difference
ences existed between the genders (male and female) for all according to gender was particularly small.
affective rating scales with all sounds. The results suggested Pearson’s correlations were calculated in order to ex-
that the main effects of gender in the ratings of arousal, F(1, amine the relationship between the ratings of arousal, va-
1868) = 2.61, p = .11, η2 = .001; dominance, F(1, 1868) = lence, and dominance for each semantic category and gen-
0.28, p = .60, η2 = .00; and fear, F(1, 1868) = 0.23, p = .63, η2 der separately (Table 4). The relationship between every
Behav Res (2018) 50:1415–1429 1425
Table 4 Correlation coefficients resulting from correlations of ratings of valence, arousal, and dominance, listed for each category and gender
Male Female
Table 5 List of sounds that differ by more than 2.5 standard deviations between male and female participants’ ratings in the valence and arousal
dimensions for the IADS-E
Z-Distance is the distance between male and female participants’ emotional ratings on the valence and arousal dimensions, which is converted into a z-
score
1426 Behav Res (2018) 50:1415–1429
Table 6 Proportions of sound stimuli in each valence condition, defined by loudness of the peak sound intensity
IADS-E IADS-2
Valence Category Physical Intensity Low Moderate High Low Moderate High
(≤70 dB) (70–75 dB) (≥75) (≤70 dB) (70–75 dB) (≥75)
IADS-E is the sum for both the new sounds and the IADS-2 sounds (N = 935)
two affective dimensions showed differences in each se- − .31 (p < .001) for valence, r = .47 (p < .001) for arousal, r = −
mantic category between males and females. We also .39 (p < .001) for dominance, r = .32 (p < .001) for fear, r = −
identified differences in the affective ratings for each .10 (p = .003) for happiness, and r = .18 (p < .001) for sadness.
sound between the genders. The distance between male The relationship between valence and the peak intensity
and female participants’ affective ratings in the dimen- across the 935 sounds was inconsistent with Bradley and
sions of valence and arousal was measured for each of Lang’s (2000) research, which suggested that there is no cor-
the 935 sounds and converted into a z-score. A large z- relation between valence and peak sound intensity. On the
score indicates a greater difference in ratings between the other hand, arousal had a significant positive relationship with
male and female participants. Table 5 lists 16 sounds for peak sound intensity, and this result is consistent with Bradley
which the standardized distance was greater than 2.5 stan- and Lang’s (2000) study. To further clarify the relationship
dard deviations. between the affective ratings and physical intensity, quadratic
As we mentioned above, even though no pronounced correlations were calculated between peak sound intensity and
differences were found in gender, there are still some spe- affective ratings. The results showed that only valence (r =
cific individual sounds’ ratings showed big gaps in gen- .34, p < .001) and happiness (r = .17, p < .001) reached the
der. We should pay attention to these sounds when level of significance; however, these correlations accounted
selecting the experimental materials. for only 11.5% and 3%, respectively, of the valence and hap-
piness variance. These results revealed that physically intense
Physical intensity sounds were more likely to be highly arousing and to cause a
fear response, but the relationship between a sound’s physical
Correlation analyses were performed to identify the relation- intensity and valence looks more complex.
ship between sounds’ physical intensities and their affective To clarify this problem, we first divided the sounds into
ratings. The Pearson correlation coefficients between a three conditions according to the valence rating (ratings higher
sound’s peak intensity (Lmax) and affective ratings were r = than 6 for positive sounds, greater than or equal to 4 and less
than or equal to 6 for neutral sounds, and less than 4 for
negative sounds). Then we calculated the percentages of the
Table 7 Descriptive statistics, calculated separately for each affective
sounds in three different physical intensity ranges: low inten-
rating scale in our study and the original version, for all IADS-2 sounds
sity (peak intensity ≤ 70 dB), moderate intensity (70 dB <
Dimension Our Study Original Version peak intensity < 75 dB), and high intensity (peak intensity ≥
75 dB) under the three valence types (Table 6). For sounds
Valence Mean 4.40 4.78
categorized as having positive or neutral valence, there were
SD 1.54 1.76
more sounds with low and moderate than with high intensities,
Min–Max 1.18–7.86 1.57–7.90
whereas for negative valence, there were more sounds with
Arousal Mean 5.85 5.84 high and moderate than with low intensities. This clarifies that
SD 1.00 1.16 the positive sounds are usually mild, whereas the negative
Min–Max 2.48–8.00 2.88–8.16 sounds are usually perceived as being intense and strong.
Dominance Mean 5.04 4.71 However, the numbers of sounds in the IADS-2 for each in-
SD 1.04 1.17 tensity were distributed relatively uniformly. On the basis of
Min–Max 2.00–7.24 2.29–6.86 the results in Table 6, we suspect that when people make an
All scales range from 1 to 9, in which 1 represents negative, sleepy, and
evaluation for the valence dimension, Americans may focus
being controlled. Min = minimal value; Max = maximal value; SD = on both the physical intensity of sounds and the context in
standard deviation which the sound is expressed; Japanese participants, on the
Behav Res (2018) 50:1415–1429 1427
Table 8 List of sounds that differ by more than two standard deviation between this study and the original version in the valence and arousal
dimensions for the IADS-2
Z-Distance is the distance between our study and the original version’s emotional ratings on the valence and arousal dimensions, which is converted into
a z-score
other hand, may pay more attention to the content of the (220), a party (365), a roller coaster (360), and a belch (702),
sounds. This interpretation should be considered in other sub- showed no difference in the arousal dimension; however, for
sequent research. valence, the original raters rated them as more positive than in
this study. For the sound of a thunderstorm (602), our partic-
Comparison of affective ratings for the IADS-2 sounds ipants rated it as more negative and higher in arousal than did
those in the original version.
We compared the affective ratings of the IADS-2 sounds in the Overall, although we must note carefully that our partici-
original version (Bradley & Lang, 2007a) with the results in pants evaluated the IADS-2 sounds in a highly similar way to
our study. Table 7 presents the means, standard deviations, the the participants in the original version, it is also true that some
minimal and maximal values for the ratings of arousal, va- slight differences have been found. These differences should
lence, and dominance from our study and the original study not be overlooked, particularly for the sounds closely related
separately. Very high Pearson correlations were obtained for to human activities.
all three ratings, of valence, arousal, and dominance, between
the two groups (rVal = .81, p < .001; rAro = .76, p < .001; rDom
= .71, p < .001). The results indicate that the two groups of Conclusion
participants rated the IADS-2 sounds in similar manners.
To evaluate the differences in each affective dimension, a In this article, we collected 935 affective sounds based on the
MANOVA was conducted with culture as a between-subjects IADS-2 and tried to expand and improve the existing affective
factor and with the affective dimensions as dependent vari- auditory stimuli database. The results showed that we could
ables. The ratings of arousal, F(1, 332) = 0.002, p = .96, η2 = successfully provide a larger standardized database of
.00, did not reach the level of statistical significance here, but emotional-evocative stimuli, which includes contents across
the ratings of dominance, F(1, 332) = 7.44, p < .01, η2 = .02, a wider range of semantic categories. Through our research,
and valence, F(1, 332) = 4.58, p < .05, η2 = .01, showed the new affective sound stimulus standardized database
significant differences between the two groups. The (IADS-E) may enable the comparison of results across various
Japanese participants rated sounds as less positive and more kinds of auditory research.
dominant than did the original participants (Table 7). To clar- On the other hand, it is obvious that real-life emotional
ify in more detail the differences for individual sounds in the experiences mostly rely on the presence of combined stimuli
IADS-2, we calculated the distances in the dimensions of va- coming from multiple sensory modalities (Baumgartner,
lence and arousal for each of the 167 sounds between our Esslen, & Jäncke, 2006). Information received may have been
study and the original version, and then converted each dis- transmitted through different modalities and channels.
tance into a z-score. Table 8 lists ten sounds for which the Auditory stimuli can increase the power of affective visual
standardized distance was greater than two standard devia- stimuli, and vice versa. The visual and auditory modalities
tions. The sounds closely related to human activities, such as can integrate to provide a complete assessment of the emo-
erotic sounds (202, 216, 201, 215, and 210), a boy laughing tional qualities of a situation or an object (Gerdes et al., 2014).
1428 Behav Res (2018) 50:1415–1429
Through this article, we expected to be able to provide a reli- research on auditory affective processing. Behavior Research
Methods, 40, 531–539. https://doi.org/10.3758/BRM.40.2.531
able auditory stimulus database to examine the mechanism of
Bradley, M., & Lang, P. J. (1994). Measuring emotion: The Self-
emotional processing of different modalities and promote this Assessment Manikin and the semantic differential. Journal of
kind of emotional research. Behavior Therapy and Experimental Psychiatry, 25, 49–59.
As we mentioned earlier, there are still some limitations to https://doi.org/10.1016/0005-7916(94)90063-9
Bradley, M. M., & Lang, P. P. J. (1999). International Affective Digitized
our study. First, as compared with a visual stimulus database
Sounds (IADS-1): Stimuli, instruction manual, and affective ratings
such as the IAPS (Lang et al., 2008), although the IAPS did (Technical Report No. B-2). Gainesville, FL: University of Florida,
not make the number of stimuli in each arousal–valence quad- Center for Research in Psychophysiology.
rant equal either, it is obvious that the IAPS’s distribution is Bradley, M. M., & Lang, P. J. (2000). Affective reactions to acoustic
stimuli. Psychophysiology, 37, 204–215. https://doi.org/10.1111/
relatively balanced. In response to this study, we are commit-
1469-8986.3720204
ted to increasing the number of sounds for the quadrants that Bradley, M. M., & Lang, P. J. (2007a). The International Affective
have few sounds and trying to develop a more practical affec- Digitized Sounds: Affective ratings of sounds and instruction
tive sound stimulus database. Second, though the auditory manual (Technical Report No. B-3). Gainesville, FL: University of
Florida, NIMH Center for the Study of Emotion and Attention.
stimuli were less sensitive than visual stimuli are to gender
Bradley, M. M., & Lang, P. J. (2007b). Affective Norms for English Words
influences, there were still some individual sound stimuli with (ANEW): Instruction manual and affective ratings (Technical
large differences in gender. Therefore, when we select exper- Report No. C-1). Gainesville, FL: University of Florida, Center for
imental stimuli for psychological experiments, it is necessary Research in Psychophysiology.
to pay attention to the specific individual sounds. Third, one Buodo, G., Sarlo, M., Mento, G., Messerotti Benvenuti, S., & Palomba,
D. (2017). Unpleasant stimuli differentially modulate inhibitory pro-
might want to argue that the new database is just a local data- cesses in an emotional Go/NoGo task: An event-related potential
base for Japanese research, because our main participants study. Cognition and Emotion, 31, 127–138. https://doi.org/10.
were Japanese college students. Even though we have deter- 1080/02699931.2015.1089842
mined that the participants in our study and in the original Castonguay, A. L., Sabiston, C. M., Crocker, P. R. E., & Mack, D. E.
(2014). Development and validation of the Body and Appearance
version rated sound stimuli in highly similar manners, whether Self-Conscious Emotions Scale (BASES). Body Image, 11, 126–
or not the emotional ratings obtained in this study would be 136. https://doi.org/10.1016/j.bodyim.2013.12.006
applicable to the elderly, children, or those in other countries Castro, S. L., & Lima, C. F. (2010). Recognizing emotions in spoken
remains to be explored in further studies. Fourth, in this study language: A validated set of Portuguese sentences and
pseudosentences for research on emotional prosody. Behavior
we assessed only three basic emotions (happiness, sadness, Research Methods, 42, 74–81. https://doi.org/10.3758/BRM.42.1.
and fear); to offer a more complete characterization of the 74
basic emotions aroused by the IADS-E stimuli, further re- Choi, Y., Lee, S., Choi, I. M., Jung, S., Park, Y. K., & Kim, C. (2015).
search should measure other basic emotions, such as disgust, International Affective Digitized Sounds in Korea: A cross-cultural
adaptation and validation study. Acta Acustica United with Acustica,
anger, and surprise. 101, 134–144. https://doi.org/10.3813/AAA.918811
Choi, Y., Lee, S., Jung, S., Choi, I.-M., Park, Y.-K., & Kim, C. (2016).
Author note This research was supported by the Center of Innovation Erratum to: Development of an auditory emotion recognition func-
Program of the Japan Science and Technology Agency (JST). In addition, tion using psychoacoustic parameters based on the International
we thank Syouichi Shiota, Madoca Miyagi, and Shiho Kashihara for their Affective Digitized Sounds. Behavior Research Methods, 48, 827.
kind advice on the manuscript and analysis. Special thanks to Yuko https://doi.org/10.3758/s13428-015-0596-x
Sonobe, Shogo Aida, Narumi Yamagata, Masashi Nishijima, Aiko Czigler, I., Cox, T. J., Gyimesi, K., & Horváth, J. (2007). Event-related
Nagao, and Noriko Miura for their assistance with the experiments and potential study to aversive auditory stimuli. Neuroscience Letters,
data collection. 420, 251–256. https://doi.org/10.1016/j.neulet.2007.05.007
da Silva, S. P., & Backs, R. W. (2015). Cardiac response during auditory
selective attention to tones and affective sounds. Psychophysiology,
52, 1099–1105. https://doi.org/10.1111/psyp.12432
References Drače, S., Efendić, E., Kusturica, M., & Landzo, L. (2013). Cross-cultural
validation of the Binternational affective picture system^(IAPS) on a
sample from bosnia and herzegovina. Psihologija, 46, 17–26. doi:
Armony, J. L., Aubé, W., Angulo-Perkins, A., Peretz, I., & Concha, L. https://doi.org/10.2298/PSI1301017D
(2015). The specificity of neural responses to music and their rela- Droit-Volet, S., Ramos, D., Bueno, J. L. O., & Bigand, E. (2013). Music,
tion to voice processing: An fMRI-adaptation study. Neuroscience emotion, and time perception: The influence of subjective emotional
Letters, 593, 35–39. https://doi.org/10.1016/j.neulet.2015.03.011 valence and arousal? Frontiers in Psychology, 4, 417:1–12. doi:
Banse, R., & Scherer, K. R. (1996). Acoustic profiles in vocal emotion https://doi.org/10.3389/fpsyg.2013.00417
expression. Journal of Personality and Social Psychology, 70, 614– Dufey, M., Fernández, A. M., & Mayol, R. (2011). Adding support to
636. https://doi.org/10.1037/0022-3514.70.3.614 cross-cultural emotional assessment: Validation of the International
Baumgartner, T., Esslen, M., & Jäncke, L. (2006). From emotion percep- Affective Picture System in a Chilean sample. Universitas
tion to emotion experience: Emotions evoked by pictures and clas- Psychologica, 10, 521–533.
sical music. International Journal of Psychophysiology, 60, 34–43. Ekman, P. (1992). Are there basic emotions? Psychological Review, 99,
https://doi.org/10.1016/j.ijpsycho.2005.04.007 550–553. https://doi.org/10.1037/0033-295X.99.3.550.
Belin, P., Fillion-Bilodeau, S., & Gosselin, F. (2008). The Montreal Esslen, M., Pascual-Marqui, R. D., Hell, D., Kochi, K., & Lehmann, D.
Affective Voices: A validated set of nonverbal affect bursts for (2004). Brain areas and time course of emotional processing.
Behav Res (2018) 50:1415–1429 1429
NeuroImage, 21, 1189–1203. https://doi.org/10.1016/j.neuroimage. Marchewka, A., Zurawski, L., Jednoróg, K., & Grabowska, A. (2014).
2003.10.001 The Nencki Affective Picture System (NAPS): Introduction to a
Fabiani, M., Kazmerski, V. A., Cycowicz, Y. M., & Friedman, D. (1996). novel, standardized, wide-range, high-quality, realistic picture data-
Naming norms for brief environmental sounds: Effects of age and base. Behavior Research Methods, 46, 596–610. https://doi.org/10.
dementia. Psychophysiology, 33, 462–475. 3758/s13428-013-0379-1
Fazio, R. H. (2001). On the automatic activation of associated evalua- Marin, M. M., Gingras, B., & Bhattacharya, J. (2012). Crossmodal trans-
tions: An overview. Cognition and Emotion, 15, 115–141. https:// fer of arousal, but not pleasantness, from the musical to the visual
doi.org/10.1080/0269993004200024 domain. Emotion, 12, 618–631. https://doi.org/10.1037/a0025020
Gerdes, A. B. M., Wieser, M. J., & Alpers, G. W. (2014). Emotional Mikels, J. A., Fredrickson, B. L., Larkin, G. R., Lindberg, C. M., Maglio,
pictures and sounds: A review of multimodal interactions of emotion S. J., & Reuter-Lorenz, P. A. (2005). Emotional category data on
cues in multiple domains. Frontiers in Psychology, 5, 1351:1–10. images from the International Affective Picture System. Behavior
https://doi.org/10.3389/fpsyg.2014.01351 Research Methods, 37, 626–630. https://doi.org/10.3758/
Iacobucci, D., & Duhachek, A. (2003). Advancing alpha: Measuring BF03192732
reliability with confidence. Journal of Consumer Psychology, 13, Oldfield, R. C. (1971). The assessment and analysis of handedness: The
478–487. https://doi.org/10.1207/S15327663JCP1304_14 Edinburgh inventory. Neuropsychologia, 9, 97–113. https://doi.org/
Iwata, N., & Higuchi, H. R. (2000). Responses of Japanese and American 10.1016/0028-3932(71)90067-4
university students to the STAI items that assess the presence or Pell, M. D., & Kotz, S. A. (2011). On the time course of vocal emotion
absence of anxiety. Journal of Personality Assessment, 74, 48–62. recognition. PLoS ONE, 6, e27256. https://doi.org/10.1371/journal.
https://doi.org/10.1207/S15327752JPA740104 pone.0027256
Jaquet, L., Danuser, B., & Gomez, P. (2012). Music and felt emotions: Redondo, J., Fraga, I., Padrón, I., & Piñeiro, A. (2008). Affective ratings
How systematic pitch level variations affect the experience of pleas- of sound stimuli. Behavior Research Methods, 40, 784–790. https://
antness and arousal. Psychology of Music, 42, 51–70. https://doi. doi.org/10.3758/BRM.40.3.784
org/10.1177/0305735612456583 Schindler, S., & Kissler, J. (2016). Selective visual attention to emotional
Juslin, P. N., & Laukka, P. (2003). Communication of emotions in vocal words: Early parallel frontal and visual activations followed by in-
expression and music performance: Different channels, same code? teractive effects in visual cortex. Human Brain Mapping, 37, 3575–
Psychological Bulletin, 129, 770–814. https://doi.org/10.1037/ 3587. https://doi.org/10.1002/hbm.23261
0033-2909.129.5.770 Schirmer, A., & Kotz, S. A. (2006). Beyond the right hemisphere: Brain
Krosnick, J. A., & Presser, S. (2010). Question and questionnaire design. mechanisms mediating vocal emotional processing. Trends in
In P. V. Marsden & J. D. Wright (Eds.), Handbook of survey re- Cognitive Sciences, 10, 24–30. https://doi.org/10.1016/j.tics.2005.
search (pp. 263–314). Bingley, UK: Emerald Group. 11.009
Lang, P. J. (1980). Behavioral treatment and bio-behavioral assessment: Shimizu, H., & Imae, K. (1981). Development of Japanese collegiate
Computer applications. In J. B. Sidowski, J. H. Johnson, & T. A. version of State–Trait Anxiety Inventory (in Japanese). Japanese
Williams (Eds.), Technology in mental health care delivery (pp. Journal of Educational Psychology, 29, 62–67.
119–137). Norwood, NJ: Ablex. Shuman, V., Clark-Polner, E., Meuleman, B., Sander, D., & Scherer, K.
Lang, P. J., Bradley, M. M., & Cuthbert, B. N. (1997). International R. (2015). Emotion perception from a componential perspective.
Affective Picture System (IAPS): Technical manual and affective Cognition and Emotion, 31, 47–56. https://doi.org/10.1080/
ratings (Technical Report No. A-1). Gainesville, FL: University of 02699931.2015.1075964
Florida, NIMH Center for the Study of Emotion and Attention. Soares, A. P., Pinheiro, A. P., Costa, A., Frade, C. S., Comesaña, M., &
Lang, P. J., Bradley, M. M., & Cuthbert, B. N. (2008). International Pureza, R. (2013). Affective auditory stimuli: Adaptation of the
affective picture system (IAPS): Affective ratings of pictures and International Affective Digitized Sounds (IADS-2) for European
instruction manual (Technical Report A-8). Gainesville, FL: Portuguese. Behavior Research Methods, 45, 1168–1181. https://
University of Florida, Center for Research in Psychophysiology. doi.org/10.3758/s13428-012-0310-1
Lang, P. J., Greenwald, M. K., Bradley, M. M., & Hamm, A. O. (1993). Spielberger, C.D., Gorsuch, R. L., & Lushene, R. E. (1970). STAI:
Looking at pictures: Affective, facial, visceral, and behavioral reac- Manual for the State–Trait Anxiety Inventory (BSelf-Evaluation
tions. Psychophysiology, 30, 261–273. https://doi.org/10.1111/j. Questionnaire^). Palo Alto, CA: Consulting Psychologists Press.
1469-8986.1993.tb03352.x Stevenson, R. A, & James, T. W. (2008). Affective auditory stimuli:
Lindquist, K. A. (2010). The brain basis of emotion: A meta-analytic Characterization of the International Affective Digitized Sounds
review. Dissertation Abstracts International, B: Sciences and (IADS) by discrete emotional categories. Behavior Research
Engineering, 71, 2744. https://doi.org/10.1017/ Methods, 40, 315–321. https://doi.org/10.3758/BRM.40.1.315
S0140525X11000446 Suk, H. J., & Irtel, H. (2010). Emotional response to color across media.
Liu, P., & Pell, M. D. (2012). Recognizing vocal emotions in Mandarin Color Research and Application, 35, 64–77. https://doi.org/10.
Chinese: A validated database of Chinese vocal emotional stimuli. 1002/col.20554
Behavior Research Methods, 44, 1042–1051. https://doi.org/10. Vieillard, S., Peretz, I., Gosselin, N., Khalfa, S., Gagnon, L., & Bouchard,
3758/s13428-012-0203-3 B. (2008). Happy, sad, scary and peaceful musical excerpts for re-
Lolli, S. L., Lewenstein, A. D., Basurto, J., Winnik, S., & Loui, P. (2015). search on emotions. Cognition and Emotion, 22, 720–752. https://
Sound frequency affects speech emotion perception: Results from doi.org/10.1080/02699930701503567
congenital amusia. Frontiers in Psychology, 6, 1340:1–10. https:// Zillmann, D. (1988). Mood management through communication
doi.org/10.3389/fpsyg.2015.01340 choices. American Behavioral Scientist, 31, 327–340.