Skip to content

Latest commit

 

History

History
81 lines (58 loc) · 5.93 KB

README.md

File metadata and controls

81 lines (58 loc) · 5.93 KB

GIMLeT – Gestural Interaction Machine Learning Toolkit

A set of Max patches for gesture analysis, interactive machine learning, and gesture-sound interaction design. GIMLeT features a modular design that allows to easily share meaningfully structured data between several gesture tracking devices, machine learning, and sound synthesis modules.

NOTE: the PoseNet implementation used in this package is now deprecated and therefore is likely not going to work. I don't think I will have time to fix that in the foreseeable future and therefore anyone's contrubution is welcome, see this issue.

Installation

Install the required packages

  1. Download the odot package .zip file: https://github.com/CNMAT/CNMAT-odot/releases/download/1.3.0/odot-Max-1.3.0.zip
  2. Open the .zip file and copy the odot folder in your /Max 8/Packages folder.
  3. Download the modosc package .zip file: https://github.com/motiondescriptors/modosc/archive/main.zip
  4. Open the .zip file and copy the modosc folder in your /Max 8/Packages folder.
  5. Download the GIMLeT package .zip file: https://github.com/federicoVisi/GIMLeT/archive/main.zip
  6. Open the .zip file and copy the GIMLeT folder in your /Max 8/Packages folder.

Install the TouchOSC layout

  1. Install TouchOSC on your smarphone (iOS or Android, you'll find it on the app store)
  2. Sync the /TouchOSC/GIMLeT_TouchOSC_remote.touchosc using this guide: https://hexler.net/docs/touchosc-editor-sync
  3. Connect TouchOSC to your computer followng this guide: https://hexler.net/docs/touchosc-configuration-connections-osc
  4. Make sure the outgoing OSC port in the TouchOSC settings (see link above) is the same as the RmtCtrl Port shown in the gimlet.ml.ann module.

Launch the example patches

Launch Max, click on Extras->"GIMLeT examples" on the menu bar, choose an example.

Video Tutorials

  1. Installation and linear regression with artifical neural networks: https://youtu.be/Dace1sHy1IM
  2. Gesture following with PoseNet and GVF: https://youtu.be/GoNqiCvVgoY

Dependencies

Included in the package

Installed separately

Literature

Book chapter with an overview of interactive machine learning of musical gesture (please cite this if you use the package in a research project)

Visi, F. G., & Tanaka, A. (2021). Interactive Machine Learning of Musical Gesture. In E. R. Miranda (Ed.), Handbook of Artificial Intelligence for Music: Foundations, Advanced Approaches, and Developments for Creativity. Springer, 2021.

Paper on the Gesture Variation Follower algorithm

Caramiaux, B., Montecchio, N., Tanaka, A., & Bevilacqua, F. (2014). Adaptive Gesture Recognition with Variation Estimation for Interactive Systems. ACM Transactions on Interactive Intelligent Systems, 4(4), 1–34. https://doi.org/10.1145/2643204

Acknowledgements and history

The project was initiated as a collaboration between Federico Visi and Hochschule für Musik und Theater Hamburg, Germany, within the framework of the KiSS: Kinetics in Sound and Space project.

gimlet.mangle is based on a synth design by Atau Tanaka. The data recorder in gimlet.ml.ann is based on a design by Michael Zbyszyński.

Further development was carried out by FV as part of a postdoctoral research position at GEMM))) Gesture Embodiment and Machines in Music – Piteå School of Music – Luleå University of Technology, Sweden.

The package is being used and developed further in several projects including:

Contact

mail[at]federicovisi[dot]com

www.federicovisi.com