Advertisement

Automatic Visualization and Graphical Editing of Virtual Modeling Networks for the Open-Source Synth-A-Modeler Compiler

  • Edgar BerdahlEmail author
  • Peter Vasil
  • Andrew Pfalz
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9775)

Abstract

Synth-A-Modeler (SaM) is an open-source environment for developing audio-haptic interactions. It is based on the paradigm that a user develops a virtual model by interconnecting virtual objects and adjusting their parameters. The SaM Designer has been created to help assist users in automatically visualizing virtual models and in graphically editing them. Certain tasks, such as modifying the parameters for a group of objects, can be completed much faster in the Designer than by using a text editor. The SaM Designer was able to automatically visualize 72 test models using an example set of visualization parameters. With these tools, new virtual models such as a “waveguide drum” can be discovered, potentially facilitating the creation of novel audio-haptic interactions.

Keywords

Haptic Virtual environments Virtual modeling Sound synthesis Digital waveguides 

References

  1. 1.
    Adrien, J.M.: Representations of musical signals. In: The Missing Link: Modal Synthesis, pp. 269–297. MIT Press, Cambridge (1991)Google Scholar
  2. 2.
    Bak, J., Verplank, W., Gauthier, D.: Motors, music and motion. In: Proceedings of the Ninth International Conference on Tangible, Embedded, and Embodied Interaction, TEI 2015, pp. 367–374. ACM, New York (2015)Google Scholar
  3. 3.
    Berdahl, E., Kontogeorgakopoulos, A., Overholt, D.: HSP v2: haptic signal processing with extensions for physical modeling. In: Proceedings of the Haptic Audio Interaction Design Conference, Copenhagen, Denmark, pp. 61–62, September 2010Google Scholar
  4. 4.
    Berdahl, E., Smith III., J.: An introduction to the Synth-A-Modeler compiler: modular and open-source sound synthesis using physical models. In: Proceedings of the Linux Audio Conference, Stanford, CA, USA, April 2012Google Scholar
  5. 5.
    Cadoz, C.: The physical model as metaphor for musical creation. pico..TERA, a piece entirely generated by a physical model. In: Proceedings of the International Computer Music Conference, Göteborg, Sweden (2002)Google Scholar
  6. 6.
    Cadoz, C., Luciani, A., Florens, J.L.: Responsive input devices and sound synthesis by simulation of instrumental mechanisms: the cordis system. Comput. Music J. 8(3), 60–73 (1984)CrossRefGoogle Scholar
  7. 7.
    Castagne, N., Cadoz, C.: Creating music by means of ‘physical thinking’: the musician oriented Genesis environment. In: Proceedings of 5th International Conference on Digital Audio Effects, Hamburg, Germany, pp. 169–174, September 2002Google Scholar
  8. 8.
    Colgate, J.E., Hogan, N.: Robust control of dynamically interacting systems. Int. J. Control 48(1), 65–88 (1988)MathSciNetCrossRefzbMATHGoogle Scholar
  9. 9.
    Conti, F., Barbagli, F., Balaniuk, R., Halg, M., Lu, C., Morris, D., Sentis, L., Warren, J., Khatib, O., Salisbury, K.: The chai libraries. In: Proceedings of EuroHaptics, Dublin, Ireland, pp. 496–500, 6–9 July 2003Google Scholar
  10. 10.
    Cook, P., Scavone, G.: The synthesis toolkit (STK), version 2.1. In: Proceedings of International Computer Music Conference, Beijing, China (1999)Google Scholar
  11. 11.
    Ellis, N., Bensoam, J., Caussé, R.: Modalys demonstration. In: Proceedings of International Computer Music Conference (ICMC 2005), Barcelona, Spain, pp. 101–102 (2005)Google Scholar
  12. 12.
    Hu, Y.: Efficient and high quality force-directed graph drawing. Mathematica J. 10, 37–71 (2005)Google Scholar
  13. 13.
    Karjalainen, M.: Blockcompiler: efficient simulation of acoustic and audio systems. In: Proceedings of the 114th Convention of the Audio Engineering Society, Preprint #5756, Amsterdam, The Netherlands, 22–25 March 2003Google Scholar
  14. 14.
    Kontogeorgakopoulos, A., Cadoz, C.: Cordis anima physical modeling and simulation system analysis. In: Proceedings of 4th Sound and Music Computing Conference, Lefkada, Greece, pp. 275–282, July 2007Google Scholar
  15. 15.
    Lee, N., Berdahl, E., Niemeyer, G., Smith III., J.: Practical implementation of low-latency DSP for feedback control of sound. In: Proceedings of Acoustics 2008, Paris, France, July 2008Google Scholar
  16. 16.
    Leonard, J., Cadoz, C.: Physical modelling concepts for a collection of multisensory virtual musical instruments. In: Proceedings of the International Conference on New Instruments for Musical Expression, Baton Rouge, LA, USA, 1–3 June 2015Google Scholar
  17. 17.
    Orlarey, Y., Fober, D., Letz, S.: FAUST: an efficient functional approach to DSP programming. In: New Computational Paradigms for Computer Music. Edition Delatour, Sampzon, France (2009)Google Scholar
  18. 18.
    Pai, D., van den Doel, K., James, D., Lang, J., Lloyd, J., Richmond, J., Yau, S.: Scanning physical interaction behavior of 3d objects. In: Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques (SIGGRAPH), Los Angeles, CA, USA, pp. 87–96, 12–17 August 2001Google Scholar
  19. 19.
    SenseGraphics AB: H3D API Manual. Kista, Sweden, June 2014Google Scholar
  20. 20.
    Sinclair, S., Wanderley, M.: Extending DIMPLE: a rigid body haptic simulator for interactive control of sound. In: Proceedings of 4th International Conference on Enactive Interfaces, November 2007Google Scholar
  21. 21.
    Smith, J.O.: Physical Audio Signal Processing: For Virtual Musical Instruments and Audio Effects. W3K Publishing. http://ccrma.stanford.edu/jos/pasp/

Copyright information

© Springer International Publishing Switzerland 2016

Authors and Affiliations

  1. 1.School of MusicLouisiana State UniversityBaton RougeUSA
  2. 2.Center for Computation and TechnologyLouisiana State UniversityBaton RougeUSA
  3. 3.Audio Communication GroupTechnical University of BerlinBerlinGermany

Personalised recommendations