TY - JOUR
T1 - The emotional component of Infant Directed-Speech
T2 - A cross-cultural study using machine learning
AU - Parlato-Oliveira, Erika
AU - Chetouani, Mohamed
AU - Cadic, Jean Maximilien
AU - Viaux, Sylvie
AU - Ghattassi, Zeineb
AU - Xavier, Jean
AU - Ouss, Lisa
AU - Feldman, Ruth
AU - Muratori, Filippo
AU - Cohen, David
AU - Saint-Georges, Catherine
N1 - Funding Information:
The study was supported by the Agence Nationale de la Recherche (ANR-12-SAMA-006) and the Groupement de Recherche en Psychiatrie (GDR-3557). It was partially performed in the Labex SMART (ANR- 11-LABX-65), which is supported by French state funds and managed by the ANR in the Investissements d'Avenir program under reference ANR-11-IDEX-0004-02. The sponsors had no involvement in the study design, data analysis, or interpretation of the results.
Publisher Copyright:
© 2019 Elsevier Masson SAS
PY - 2020/3
Y1 - 2020/3
N2 - Backgrounds: Infant-directed speech (IDS) is part of an interactive loop that plays an important role in infants’ cognitive and social development. The use of IDS is universal and is composed of linguistic and emotional components. However, whether the emotional component has similar acoustics characteristics has not been studied automatically. Methods: We performed a cross-cultural study using automatic social signal processing techniques (SSP) to compare IDS across languages. Our speech corpus consisted of audio-recorded vocalizations from parents during interactions with their infant between the ages of 4 and 18 months. It included 6 databases of five languages: English, French, Hebrew (two databases: mothers/fathers), Italian, and Brazilian Portuguese. We used an automatic classifier that exploits the acoustic characteristics of speech and machine learning methods (Support Vector Machines, SVM) to distinguish emotional IDS and non-emotional IDS. Results: Automated classification of emotional IDS was possible for all languages and speakers (father and mother). The uni-language condition (classifier trained and tested in the same language) produced moderate to excellent classification results, all of which were significantly different from chance (P < 1 × 10−10). More interestingly, the cross-over condition (IDS classifier trained in one language and tested in another language) produced classification results that were all significantly different from chance (P < 1 × 10−10). Conclusion: The automated classification of emotional and non-emotional components of IDS is possible based on the acoustic characteristics regardless of the language. The results found in the cross-over condition support the hypothesis that the emotional component shares similar acoustic characteristics across languages.
AB - Backgrounds: Infant-directed speech (IDS) is part of an interactive loop that plays an important role in infants’ cognitive and social development. The use of IDS is universal and is composed of linguistic and emotional components. However, whether the emotional component has similar acoustics characteristics has not been studied automatically. Methods: We performed a cross-cultural study using automatic social signal processing techniques (SSP) to compare IDS across languages. Our speech corpus consisted of audio-recorded vocalizations from parents during interactions with their infant between the ages of 4 and 18 months. It included 6 databases of five languages: English, French, Hebrew (two databases: mothers/fathers), Italian, and Brazilian Portuguese. We used an automatic classifier that exploits the acoustic characteristics of speech and machine learning methods (Support Vector Machines, SVM) to distinguish emotional IDS and non-emotional IDS. Results: Automated classification of emotional IDS was possible for all languages and speakers (father and mother). The uni-language condition (classifier trained and tested in the same language) produced moderate to excellent classification results, all of which were significantly different from chance (P < 1 × 10−10). More interestingly, the cross-over condition (IDS classifier trained in one language and tested in another language) produced classification results that were all significantly different from chance (P < 1 × 10−10). Conclusion: The automated classification of emotional and non-emotional components of IDS is possible based on the acoustic characteristics regardless of the language. The results found in the cross-over condition support the hypothesis that the emotional component shares similar acoustic characteristics across languages.
KW - Cross-cultural
KW - Machine learning
KW - Mother-child interaction
KW - Motherese
KW - Social signal processing
UR - http://www.scopus.com/inward/record.url?scp=85075434423&partnerID=8YFLogxK
U2 - 10.1016/j.neurenf.2019.10.004
DO - 10.1016/j.neurenf.2019.10.004
M3 - Article
AN - SCOPUS:85075434423
SN - 0222-9617
VL - 68
SP - 106
EP - 113
JO - Neuropsychiatrie de l'Enfance et de l'Adolescence
JF - Neuropsychiatrie de l'Enfance et de l'Adolescence
IS - 2
ER -