Skip to main content

Advertisement

Log in

Learning Low-Dimensional Semantics for Music and Language via Multi-Subject fMRI

  • Original Article
  • Published:
Neuroinformatics Aims and scope Submit manuscript

Abstract

Embodied Cognition (EC) states that semantics is encoded in the brain as firing patterns of neural circuits, which are learned according to the statistical structure of human multimodal experience. However, each human brain is idiosyncratically biased, according to its subjective experience, making this biological semantic machinery noisy with respect to semantics inherent to media, such as music and language. We propose to represent media semantics using low-dimensional vector embeddings by jointly modeling the functional Magnetic Resonance Imaging (fMRI) activity of several brains via Generalized Canonical Correlation Analysis (GCCA). We evaluate the semantic richness of the resulting latent space in appropriate semantic classification tasks: music genres and language topics. We show that the resulting unsupervised representations outperform the original high-dimensional fMRI voxel spaces in these downstream tasks while being more computationally efficient. Furthermore, we show that joint modeling of several subjects increases the semantic richness of the learned latent vector spaces as the number of subjects increases. Quantitative results and corresponding statistical significance testing demonstrate the instantiation of music and language semantics in the brain, thereby providing further evidence for multimodal embodied cognition as well as a method for extraction of media semantics from multi-subject brain dynamics.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

Data Availability

The data used for these experiments are based on the following repositories: https://openneuro.org/datasets/ds000113/versions/1.3.0 (MG) and https://osf.io/crwz7 (LT243 and LT384).

Code Availability

The source code is freely available for use from https://gitlab.hlt.inesc-id.pt/fraposo/fmri-gcca-pub.

References

Download references

Funding

Francisco Afonso Raposo is supported by a PhD scholarship granted by Fundação para a Ciância e a Tecnologia (FCT), with reference SFRH/BD/135659/2018. Additionally, this work was supported by Portuguese national funds through FCT, with reference UIDB/50021/2020.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Francisco Afonso Raposo.

Ethics declarations

Conflict of Interest

The authors declare that they have no conflict of interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Raposo, F.A., Martins de Matos, D. & Ribeiro, R. Learning Low-Dimensional Semantics for Music and Language via Multi-Subject fMRI. Neuroinform 20, 451–461 (2022). https://doi.org/10.1007/s12021-021-09560-5

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12021-021-09560-5

Keywords

Navigation