SMC 2019 Scientific Program

SMC 2019 Scientific Program is already available. You can download a summary in PDF format HERE.


There is also available a small book that includes the SMC 2019 Program and Abstracts.


WEDNESDAY, MAY 29

ORAL SESSION S1. Sonic interactions
Wednesday, May 29 (11:30- 13.00)
Session Chair: Stefania Serafin
S1.1. Towards a High-Performance Platform for Sonic Interaction Interfaces
Stefano Fasciani and Manohar Vohra
S1.2. Digital Manufacturing for Musical Applications: A Survey of Current Status and Future Outlook
Doga Cavdir
S1.3. Real Time Audio Digital Signal Processing with Faust and the Teensy
Romain Michon, Yann Orlarey, Stéphane Letz and Dominique Fober
S1.4. Sound Design through Large Audience Interaction
Kjetil Falkenberg Hansen, Martin Ljungdahl-Eriksson and Ricardo Atienza
S1.5. Evaluating a Continuous Sonic Interaction: Comparing a Performable Acoustic and Digital Everyday Sound
Fiona Keenan and Sandra Pauletto
Oral Session S2. Nordic SMC
Wednesday, May 29 (14:30- 16:30)
Session Chair: Vesa Välimäki
S2.1. Adaptive Loudness Compensation in Music Listening
Leonardo Fierro, Jussi Rämö and Vesa Välimäki
S2.2. Toward Automatic Tuning of the Piano
Joonas Tuovinen, Jamin Hu and Vesa Välimäki.
S2.3. Real-time Control of Large-scale Modular Physical Models using the Sensel Morph
Silvin Willemsen, Nikolaj Andersson, Stefania Serafin and Stefan Bilbao
S2.4. An Interactive Music Synthesizer for Gait Training in Neurorehabilitation
Prithvi Kantan and Sofia Dahl
S2.5. From Vocal Sketching to Sound Models by Means of a Sound-Based Musical Transcription System
Claudio Panariello, Mattias Sköld, Emma Frid and Roberto Bresin
S2.6. Tempo and Metrical Analysis by Tracking Multiple Metrical Levels Using Autocorrelation
Olivier Lartillot and Didier Grandjean
Poster Session 1. Wednesday, May 29
Session Chair: Hanna Järveläinen
P1.1. DAW-Integrated Beat Tracking for Music Production
Brett Dalton, David Johnson and George Tzanetakis
P1.2. Interaction-based Analysis of Freely Improvised Music
Stefano Kalonaris
P1.3. Mechanical Entanglement: A Collaborative Haptic-Music Performance
Alexandros Kontogeorgakopoulos, George Sioros and Odysseas Klissouras
P1.4. State Dependency - Audiovisual Interaction through Brain States
Patrick Neff, Jan Schacher and Daniel Bisig
P1.5. Perceptual Evaluation of Modal Synthesis for Impact-Based Sounds
Adrián Barahona and Sandra Pauletto
P1.6. VIBRA - Technical and Artistic Issues in an Interactive Dance Project
Andreas Bergsland, Sigurd Saue and Pekka Stokke
P1.7. Musical Tempo and Key Estimation with Directional Convolutional Neural Networks
Hendrik Schreiber and Meinard Müller
P1.8. The Viking HRTF Dataset
Simone Spagnol, Kristján Bjarki Purkhús, Sverrir Karl Björnsson and Runar Unnthórsson
P1.9. Performing with Sound Sample-Controlled Gloves and Light-Controlled Arms
Frank Pecquet, Fotis Moschos, David Fierro and Justin Pecquet
P1.10. Melody Identification in Standard MIDI Files
Zheng Jiang and Roger Dannenberg
P1.11. Automatic Chord-Scale Recognition using Harmonic Pitch Class Profiles
Emir Demirel, Baris Bozkurt and Xavier Serra
Demo Session D1. Wednesday, May 29
Session Chair: Alberto Peinado
D1.1. Exciting Digital Resonators with Analogue Sound
Max Neupert and Clemens Wegener
D1.2. Melody Slot Machine
Masatoshi Hamanaka
D1.3. OM-AI: A Toolkit to Support AI-Based Computer-Assisted Composition Workflows in OpenMusic
Anders Vinjar and Jean Bresson
D1.4. URALi: a proposal of approach to real-time audio synthesis in Unity
Enrico Dorigatti
D1.5. A Sequencer with Decoupled Track Timing
Silvan David Peter and Gerhard Widmer
D1.6. Musicypher: Music for Message Encryption
Víctor Jaime Marín and Alberto Peinado
D1.7. A Platform for Processing Sheet Music and Developing Multimedia Application
Fu-Hai Frank Wu
D1.8. Capturing the Reaction Time to Distinguish between Voice and Music
Alejandro Villena-Rodríguez, Lorenzo J. Tardón, Isabel Barbancho, Ana M. Barbancho, Irene Gómez-Plazas and María-José Varela-Salinas
D1.9. Physical Models and Real-Time Control with the Sensel Morph
Silvin Willemsen , Stefan Bilbao, Nikolaj Andersson and Stefania Serafin


THURSDAY, MAY 30

Oral Session S3: Augmented and virtual realities
Thursday, May 30 (09:00- 10:30)
Session Chair: Marcella Mandanici
S3.1. Comparison and Implementation of Data Transmission Techniques through Analog Audio Signals in the Context of Augmented Mobile Instruments
Romain Michon, Yann Orlarey, Stéphane Letz and Dominique Fober
S3.2. Mass-Interaction Physical Models for Sound and Multi-Sensory Creation: Starting Anew
Jerome Villeneuve and James Leonard
S3.3. Exploring the Effects of Diegetic and Non-diegetic Audiovisual Cues on Decision-making in Virtual Reality
JAnil Çamci
S3.4. OSC-XR: A Toolkit for Extended Reality Immersive Music Interfaces
David Johnson, Daniela Damian and George Tzanetakis
S3.5. No Strings Attached: Force and Vibrotactile Feedback in a Guitar Simulation
Andrea Passalenti, Razvan Paisa, Niels Christian Nilsson, Nikolaj S. Andersson, Federico Fontana, Rolf Nordahl and Stefania Serafin
Oral Session S4. SMC tools and methodologies
Thursday, May 30 (12:00- 13:00)
Session Chair: Emma Frid
S4.1. A Framework for the Evaluation of Interpolated Synthesizer Parameter Mapping
Darrell Gibson and Richard Polfreman
S4.2. Composing with Sounds: Designing an Object Oriented Daw for the Teaching of Sound-Based Composition
Stephen Pearse, Leigh Landy, Duncan Chapman, David Holland and Mihai Eni
S4.3. Insights in Habits and Attitudes Regarding Programming Sound Synthesizers: A Quantitative Study
Gordan Krekovic
Oral Session S5. Sound synthesis & analysis
Thursday, May 30 (14:30- 16:30)
Session Chair: Federico Avanzini
S5.1. Experimental Verification of Dispersive Wave Propagation on Guitar Strings
Dmitri Kartofelev, Joann Arro and Vesa Välimäki
S5.2. Real-Time Modeling of Audio Distortion Circuits with Deep Learning
Eero-Pekka Damskägg, Lauri Juvela and Vesa Välimäki
S5.3. MI-GEN~: An Efficient and Accessible Mass-Interaction Sound Synthesis Toolbox
James Leonard and Jerome Villeneuve
S5.4. Combining Texture-Derived Vibrotactile Feedback, Concatenative Synthesis and Photogrammetry for Virtual Reality Rendering
Eduardo Magalhães, Emil Rosenlund Høeg, Gilberto Bernardes, Jon Ram Bruun-Pedersen, Stefania Serafin and Rolf Nordhal
S5.5. Percussion synthesis using loopback frequency modulation oscillators
Jennifer Hsu and Tamara Smyth
S5.6. Deep Linear Autoregressive Model for Interpretable Prediction of Expressive Tempo
Akira Maezawa
S5.7. Metrics for the Automatic Assessment of Music Harmony Awareness in Children
Federico Avanzini, Adriano Baratè, Luca Andrea Ludovico and Marcella Mandanici
Poster Session P2, Thursday, May 30
Session Chair: Anja Volk
P2.1. RaveForce: A Deep Reinforcement Learning Environment for Music Generation
Qichao Lan, Jim Tørresen and Alexander Refsum Jensenius
P2.2. Music Temperaments Evaluation Based on Triads
Meihui Tong and Satoshi Tojo
P2.3. Composing space in the space: An Augmented and Virtual Reality sound spatialization system
Giovanni Santini
P2.4. Graph Based Physical Models for Sound Synthesis
Pelle Juul Christensen and Stefania Serafin
P2.5. ADPET: Exploring the Design, Pedagogy, and Analysis of a Mixed Reality Application for Piano Training
Lynda Gerry, Sofia Dahl and Stefania Serafin
P2.6. Chord Prediction with The Annotated Beethoven Corpus
Kristoffer Landsnes, Liana Mehrabyan, Victor Wiklund, Fabian C. Moss, Robert Lieck and Martin Rohrmeier
P2.7. Sonic Characteristics of Robots in Films
Adrian B. Latupeirissa, Emma Frid and Roberto Bresin
P2.8. Virtual Reality Music Intervention to Reduce Social Anxiety in Adolescents Diagnosed with Autism Spectrum Disorder
Ali Adjorlu, Nathaly Belen Betancourt Barriga and Stefania Serafin
P2.9. Teach Me Drums: Learning Rhythms through the Embodiment of a Drumming Teacher in Virtual Reality
Mie Moth-Poulsen, Tomasz Bednarz, Volker Kuchelmeister and Stefania Serafin
P2.10. Real-time Mapping of Periodic Dance Movements to Control Tempo in Electronic Dance Music
Lilian Jap and Andre Holzapfel
P2.11. Increasing Access to Music in SEN Settings
Tom Davis, Daniel Pierson and Ann Bevan
Demo Session D2. Thursday, May 30
Session Chair: Hendrik Schreiber
D2.1. Interacting with Musebots (that don't really listen)
Arne Eigenfeldt
D2.2. Extending Jamsketch: An Improvisation Support System
Akane Yasuhara, Junko Fujii and Tetsuro Kitahara
D2.3. Visualizing Music Genres using a Topic Model
Swaroop Panda, Vinay P. Namboodiri and Shatarupa Thakurta Roy
D2.4. CompoVOX: Real-Time Sonification of Voice
Daniel Hernán Molina Villota, Antonio Jurado-Navas and Isabel Barbancho
D2.5. Facial Activity Detection to Monitor Attention and Fatigue
Oscar Cobos, Jorge Munilla, Ana M. Barbancho, Isabel Barbancho and Lorenzo J. Tardón
D2.6. The Chordinator: An Interactive Music Learning Device
Eamon McCoy, John Greene, Jared Henson, James Pinder, Jonathon Brown and Claire Arthur
D2.7. Automatic Chord Recognition in Music Education Applications
Sascha Grollmisch and Estefania Cano
D2.8. Sonic Sweetener Mug
Signe Lund Mathiesen, Derek Victor Byrne and Qian Janice Wang


FRIDAY, MAY 30

Oral Session S6. Music information processing
Friday, May 31 (09:00- 10:30)
Session Chair: Roger Dannenberg
S6.1. Learning to Generate Music with BachProp
Florian Colombo, Johanni Brea and Wulfram Gerstner
S6.2. Offline Score Alignment for Realistic Music Practice
Yucong Jiang, Fiona Ryan, David Cartledge and Christopher Raphael
S6.3. Piano Score-Following by Tracking Note Evolution
Yucong Jiang and Christopher Raphael
S6.4. Adaptive Score-Following System by Integrating Gaze Information
Kaede Noto, Yoshinari Takegawa and Keiji Hirata
S6.5. Alternative Measures: A Musicologist Workbench for Popular Music
Beach Clark and Claire Arthur
Oral Session S7. Multimodality and (e)motions
Friday, May 31 (12:00- 13:00)
Session Chair: Sofia Dahl
S7.1. VocalistMirror: A Singer Support Interface for Avoiding Undesirable Facial Expressions
Kin Wah Edward Lin, Tomoyasu Nakano and Masataka Goto
S7.2. Audiovisual perception of arousal, valence, and effort in contemporary cello performance
Hanna Järveläinen
S7.3. Dancing dots - Investigating the Link Between Dancer and Musician in Swedish Folk Dance
Olof Misgeld, Andre Holzapfel and Sven Ahlbäck
Oral Session S8. Machine learning
Friday, May 31 (14:30- 16:30)
Session Chair: Olivier Lartillot
S8.1. Conditioning a Recurrent Neural Network to synthesize musical instrument transients
Lonce Wyse and Muhammad Huzaifah
S8.2. Predicting Perceived Dissonance of Piano Chords Using a Chord-Class Invariant CNN and Deep Layered Learning
Juliette Dubois, Anders Elowsson and Anders Friberg
S8.3. Belief Propagation algorithm for Automatic Chord Estimation
Vincent P. Martin, Sylvain Reynal, Dogac Basaran and Hélène-Camille Crayencour
S8.4. HMM-Based Glissando Detection for Recordings of Chinese Bamboo Flute
Changhong Wang, Emmanouil Benetos, Xiaojie Meng and Elaine Chew
S8.5. Towards CNN-based Acoustic Modeling of Seventh Chords for Automatic Chord Recognition
Christon-Ragavan Nadar, Jakob Abeßer and Sascha Grollmisch
S8.6. From Jigs and Reels to Schottisar och Polskor: Generating Scandinavian-like Folk Music with Deep Recurrent Networks
Simon Mossmyr, Eric Hallström, Bob L. Sturm, Victor Hansjons Vegeborn and Jonas Wedin
S8.7. Modeling and Learning Rhythm Structure
Francesco Foscarin, Florent Jacquemard and Philippe Rigaux
Poster Session P3. Friday, May 31
Session Chair: Jean Bresson
P3.1. Autoencoders for music sound modeling: a comparison of linear, shallow, deep, recurrent and variational models
Fanny Roche, Thomas Hueber, Samuel Limier and Laurent Girin
P3.2. Polytopic reconfiguration: a graph-based scheme for the multiscale transformation of music segments and its perceptual assessment
Valentin Gillot and Frédéric Bimbot
P3.3. Non-Linear Contact Sound Synthesis for Real-Time Audio-Visual Applications using Modal Textures
Martin Maunsbach and Stefania Serafin
P3.4. Analysis of Vocal Ornamentation in Iranian Classical Music
Sepideh Shafiei
P3.5. VUSAA: An Augmented Reality Mobile App for Urban Soundwalks
Josue Moreno and Vesa Norilo
P3.6. A Framework for Multi-f0 Modeling in SATB Choirs
Helena Cuesta, Emilia Gómez and Pritish Chandna
P3.7. Representations of Self-Coupled Modal Oscillators with Time-Varying Sounding Frequency
Tamara Smyth and Jennifer Hsu
P3.8. SonaGraph. A cartoonified spectral model for music composition
Andrea Valle
P3.9. Sound in Multiples: Synchrony and Interaction Design of Coupled-Oscillator Networks
Nolan Lem
P3.10. Jazz Mapping an Analytical and Computational Approach to Jazz Improvisation
Dimitrios Vassilakis, Anastasia Georgaki and Christina Anagnostopoulou
P3.11. Visual Pitch Estimation
A. Sophia Koepke, Olivia Wiles and Andrew Zisserman
Demo Session D3. Friday, May 31
Session Chair: Ana M. Barbancho
D3.1. Miningsuite: A Comprehensive Matlab Framework for Signal, Audio and Music Analysis, Articulating Audio and Symbolic Approaches
Olivier Lartillot
D3.2. Drawing Geometric Figures with Braille Description through a Speech Recognition System
África Chamorro, Ana M. Barbancho, Isabel Barbancho and Lorenzo J. Tardón
D3.3. Interactive Music Training System
Daniel Moreno, Isabel Barbancho, Ana M. Barbancho and Lorenzo J. Tardón
D3.4. Copying Clave - A Turing Test
Simon Blackmore
D3.5. Resonance Improviser: A System for Transmitting the Embodied Sensations of Vocalization Between Two People During Improvisation
Tejaswinee Kelkar and Lynda Gerry
D3.6. Finding New Practice Material through Chord-Based Exploration of a Large Music Catalogue
Johan Pauwels and Mark B. Sandler
D3.7. Internal Complexity for Exploratory Interaction
Mads Hobye
D3.8. Adaptive Body Movement Sonification in Music and Therapy
Christian Baumann, Johanna Friederike Baarlink and Jan-Torsten Milde