Crowdsourcing Affective Annotations via fNIRS-BCI

Research output: Contribution to journalJournal articleResearchpeer-review

Standard

Crowdsourcing Affective Annotations via fNIRS-BCI. / Ruotsalo, Tuukka; Makela, Kalle; Spape, Michiel.

In: IEEE Transactions on Affective Computing, 2024.

Research output: Contribution to journalJournal articleResearchpeer-review

Harvard

Ruotsalo, T, Makela, K & Spape, M 2024, 'Crowdsourcing Affective Annotations via fNIRS-BCI', IEEE Transactions on Affective Computing. https://doi.org/10.1109/TAFFC.2023.3273916

APA

Ruotsalo, T., Makela, K., & Spape, M. (2024). Crowdsourcing Affective Annotations via fNIRS-BCI. IEEE Transactions on Affective Computing. https://doi.org/10.1109/TAFFC.2023.3273916

Vancouver

Ruotsalo T, Makela K, Spape M. Crowdsourcing Affective Annotations via fNIRS-BCI. IEEE Transactions on Affective Computing. 2024. https://doi.org/10.1109/TAFFC.2023.3273916

Author

Ruotsalo, Tuukka ; Makela, Kalle ; Spape, Michiel. / Crowdsourcing Affective Annotations via fNIRS-BCI. In: IEEE Transactions on Affective Computing. 2024.

Bibtex

@article{2badc29ed87e40aab23aa65de8041c53,
title = "Crowdsourcing Affective Annotations via fNIRS-BCI",
abstract = "Affective annotation refers to the process of labeling media content based on the emotions they evoke. Since such experiences are inherently subjective and depend on individual differences, the central challenge is associating digital content with its affective, interindividual experience. Here, we present a first-of-its-kind methodology for affective annotation directly from brain signals by monitoring the affective experience of a crowd of individuals via functional near-infrared spectroscopy (fNIRS). An experiment is reported in which fNIRS was recorded from 31 participants to develop a brain-computer interface (BCI) for affective annotation. Brain signals evoked by images were used to draw predictions about the affective dimensions that characterize the stimuli. By combining annotations, the results show that monitoring crowd responses can draw accurate affective annotations, with performance improving significantly with increases in crowd size. Our methodology demonstrates a proof-of-concept to source affective annotations from a crowd of BCI users without requiring any auxiliary mental or physical interaction.",
keywords = "Affective computing, Annotations, Brain modeling, Crowdsourcing, Electroencephalography, Emotion classification, fNIRS, Functional near-infrared spectroscopy, Manuals, Media, Pattern classification",
author = "Tuukka Ruotsalo and Kalle Makela and Michiel Spape",
note = "Publisher Copyright: Author",
year = "2024",
doi = "10.1109/TAFFC.2023.3273916",
language = "English",
journal = "IEEE Transactions on Affective Computing",
issn = "1949-3045",
publisher = "IEEE Signal Processing Society",

}

RIS

TY - JOUR

T1 - Crowdsourcing Affective Annotations via fNIRS-BCI

AU - Ruotsalo, Tuukka

AU - Makela, Kalle

AU - Spape, Michiel

N1 - Publisher Copyright: Author

PY - 2024

Y1 - 2024

N2 - Affective annotation refers to the process of labeling media content based on the emotions they evoke. Since such experiences are inherently subjective and depend on individual differences, the central challenge is associating digital content with its affective, interindividual experience. Here, we present a first-of-its-kind methodology for affective annotation directly from brain signals by monitoring the affective experience of a crowd of individuals via functional near-infrared spectroscopy (fNIRS). An experiment is reported in which fNIRS was recorded from 31 participants to develop a brain-computer interface (BCI) for affective annotation. Brain signals evoked by images were used to draw predictions about the affective dimensions that characterize the stimuli. By combining annotations, the results show that monitoring crowd responses can draw accurate affective annotations, with performance improving significantly with increases in crowd size. Our methodology demonstrates a proof-of-concept to source affective annotations from a crowd of BCI users without requiring any auxiliary mental or physical interaction.

AB - Affective annotation refers to the process of labeling media content based on the emotions they evoke. Since such experiences are inherently subjective and depend on individual differences, the central challenge is associating digital content with its affective, interindividual experience. Here, we present a first-of-its-kind methodology for affective annotation directly from brain signals by monitoring the affective experience of a crowd of individuals via functional near-infrared spectroscopy (fNIRS). An experiment is reported in which fNIRS was recorded from 31 participants to develop a brain-computer interface (BCI) for affective annotation. Brain signals evoked by images were used to draw predictions about the affective dimensions that characterize the stimuli. By combining annotations, the results show that monitoring crowd responses can draw accurate affective annotations, with performance improving significantly with increases in crowd size. Our methodology demonstrates a proof-of-concept to source affective annotations from a crowd of BCI users without requiring any auxiliary mental or physical interaction.

KW - Affective computing

KW - Annotations

KW - Brain modeling

KW - Crowdsourcing

KW - Electroencephalography

KW - Emotion classification

KW - fNIRS

KW - Functional near-infrared spectroscopy

KW - Manuals

KW - Media

KW - Pattern classification

UR - http://www.scopus.com/inward/record.url?scp=85159807717&partnerID=8YFLogxK

U2 - 10.1109/TAFFC.2023.3273916

DO - 10.1109/TAFFC.2023.3273916

M3 - Journal article

AN - SCOPUS:85159807717

JO - IEEE Transactions on Affective Computing

JF - IEEE Transactions on Affective Computing

SN - 1949-3045

ER -

ID: 383788909