Juan Sebastián Gómez-Cañón
Postdoctoral Scholar, Psychiatry
Bio
Juan S. Gómez-Cañón is a researcher, engineer and musician from Colombia. He holds a Ph.D. in Information and Communication Technologies from the Universitat Pompeu Fabra (Barcelona, Spain). During his Ph.D., Juan researched human-centered and trustworthy machine learning methods to predict the emotions in music. His research focuses on deep learning, human-centered ML, personalization, dataset curation, and digital signal processing. Juan also holds a M.Sc. in Media Technology (Technische Universität Ilmenau, Germany), a B.Sc. in Electronics Engineering and a B.A. in Music (Universidad de los Andes, Colombia).
Stanford Advisors
-
David Hong, Postdoctoral Faculty Sponsor
-
Daniel Bowling, Postdoctoral Research Mentor
All Publications
-
Personalisation and profiling using algorithms and not-so-popular Colombian music: goal-directed mechanisms in music emotion recognition
EPJ DATA SCIENCE
2025; 14 (1): 80
Abstract
This work investigates how personalised Music Emotion Recognition (MER) systems may lead to sensitive profiling when applied to musically induced emotions in politically charged contexts. We focus on traditional Colombian music with explicit political content, including (1) vallenatos and social songs aligned with the left-wing guerrilla Fuerzas Armadas Revolucionarias de Colombia (FARC), and (2) corridos linked to sympathisers of the right-wing paramilitary group Autodefensas Unidas de Colombia (AUC). Using data from 49 participants with diverse political leanings, we train personalised machine learning models to predict induced emotional responses - particularly negative emotions. Our findings reveal that political identity plays a significant role in shaping emotional experiences of music with explicit political content, and that emotion recognition models can capture this variation to a certain extent. These results raise critical concerns about the potential misuse of emotion recognition technologies. What is often framed as a tool for wellbeing and emotional regulation could, in politically sensitive contexts, be repurposed for user profiling. This work highlights the ethical risks of deploying AI-driven emotion analysis without safeguards, particularly among populations that are politically or socially vulnerable. We argue that subjective emotional responses may constitute sensitive personal data, and that failing to account for their sociopolitical context could amplify harm and exclusion.The online version contains supplementary material available at 10.1140/epjds/s13688-025-00595-1.
View details for DOI 10.1140/epjds/s13688-025-00595-1
View details for Web of Science ID 001613812700001
View details for PubMedID 41244640
View details for PubMedCentralID PMC12615516
-
Beyond a Western Center of Music Information Retrieval: A Bibliometric Analysis of the First 25 Years of ISMIR Authorship
TRANSACTIONS OF THE INTERNATIONAL SOCIETY FOR MUSIC INFORMATION RETRIEVAL
2025; 8 (1): 372-387
View details for DOI 10.5334/tismir.265
View details for Web of Science ID 001620706400001
-
Participant and Musical Diversity in Music Psychology Research
Music & Science
2025; 8
View details for DOI 10.1177/20592043251317180
-
Globally, songs and instrumental melodies are slower and higher and use more stable pitches than speech: A Registered Report.
Science advances
2024; 10 (20): eadm9797
Abstract
Both music and language are found in all known human societies, yet no studies have compared similarities and differences between song, speech, and instrumental music on a global scale. In this Registered Report, we analyzed two global datasets: (i) 300 annotated audio recordings representing matched sets of traditional songs, recited lyrics, conversational speech, and instrumental melodies from our 75 coauthors speaking 55 languages; and (ii) 418 previously published adult-directed song and speech recordings from 209 individuals speaking 16 languages. Of our six preregistered predictions, five were strongly supported: Relative to speech, songs use (i) higher pitch, (ii) slower temporal rate, and (iii) more stable pitches, while both songs and speech used similar (iv) pitch interval size and (v) timbral brightness. Exploratory analyses suggest that features vary along a "musi-linguistic" continuum when including instrumental melodies and recited lyrics. Our study provides strong empirical evidence of cross-cultural regularities in music and speech.
View details for DOI 10.1126/sciadv.adm9797
View details for PubMedID 38748798
View details for PubMedCentralID PMC11095461
-
Bridging cognitive neuroscience and education: Insights from EEG recording during mathematical proof evaluation
TRENDS IN NEUROSCIENCE AND EDUCATION
2024; 35: 100226
Abstract
Much of modern mathematics education prioritizes symbolic formalism even at the expense of non-symbolic intuition, we contextualize our study in the ongoing debates on the balance between symbolic and non-symbolic reasoning. We explore the dissociation of oscillatory dynamics between algebraic (symbolic) and geometric (non-symbolic) processing in advanced mathematical reasoning during a naturalistic design.Employing mobile EEG technology, we investigated students' beta and gamma wave patterns over frontal and parietal regions while they engaged with mathematical demonstrations in symbolic and non-symbolic formats within a tutor-student framework. We used extended, naturalistic stimuli to approximate an authentic educational setting.Our findings reveal nuanced distinctions in neural processing, particularly in terms of gamma waves and activity in parietal regions. Furthermore, no clear overall format preference emerged from the neuroscientific perspective despite students rating symbolic demonstrations higher for understanding and familiarity.
View details for DOI 10.1016/j.tine.2024.100226
View details for Web of Science ID 001241428400001
View details for PubMedID 38879197
-
Audio-based Music Retrieval
Information Retrieval: Advanced Topics and Techniques
Association for Computing Machinery. 2024; 1: 539-575
View details for DOI 10.1145/3674127.3702959
-
TROMPA-MER: an open dataset for personalized music emotion recognition
JOURNAL OF INTELLIGENT INFORMATION SYSTEMS
2023; 60 (2): 549-570
View details for DOI 10.1007/s10844-022-00746-0
View details for Web of Science ID 000855794800001
-
Music Emotion Recognition: Toward new, robust standards in personalized and context-sensitive applications
IEEE SIGNAL PROCESSING MAGAZINE
2021; 38 (6): 106-114
View details for DOI 10.1109/MSP.2021.3106232
View details for Web of Science ID 000711718500019
-
Let's agree to disagree: Consensus Entropy Active Learning for Personalized Music Emotion Recognition
Proceedings of the 22nd International Society for Music Information Retrieval Conference (ISMIR)
2021: 237-245
View details for DOI 10.5281/ZENODO.5624399
-
Emotion Annotation of Music: A Citizen Science Approach
edited by HernandezLeo, D., Hishiyama, R., Zurita, G., Weyers, B., Nolte, A., Ogata, H.
SPRINGER INTERNATIONAL PUBLISHING AG. 2021: 51-66
View details for DOI 10.1007/978-3-030-85071-5_4
View details for Web of Science ID 000775767900004
-
LANGUAGE-SENSITIVE MUSIC EMOTION RECOGNITION MODELS: ARE WE REALLY THERE YET?
IEEE. 2021: 576-580
View details for DOI 10.1109/ICASSP39728.2021.9413721
View details for Web of Science ID 000704288400116
-
Transfer learning from speech to music: towards language-sensitive emotion recognition models
IEEE. 2021: 136-140
View details for Web of Science ID 000632622300028
-
Cross-Dataset Music Emotion Recognition: an End-to-End Approach
Late breaking/Demo of the 21st International Society for Music Information Retrieval Conference (ISMIR)
2020
View details for DOI 10.5281/zenodo.4076771
-
Joyful for you and tender for us: the influence of individual characteristics and language on emotion labeling and classification
Proceedings of the 21th International Society for Music Information Retrieval Conference (ISMIR)
2020: 853-860
View details for DOI 10.5281/zenodo.4245567
-
Jazz Solo Instrument Classification with Convolutional Neural Networks, Source Separation, and Transfer Learning
Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR)
2018: 577-584
View details for DOI 10.5281/zenodo.1492480
https://orcid.org/0000-0002-2544-6311