Ryan McGee has released a new sound design app for iOS, VOSIS, that synthesizes sound based on the greyscale image pixel data from photos or live video input.
OSIS is an interactive image sonification interface that creates complex wavetables by raster scanning greyscale image pixel data.
Using a multi-touch screen to play image regions of unique frequency content rather than a linear scale of frequencies, it becomes a unique performance tool for experimental and visual music. A number of image filters controlled by multi-touch gestures add variation to the sound palette. On a mobile device, parameters controlled by the accelerometer add another layer expressivity to the resulting audio-visual montages.
Animated/Directed by Ambar Navarro
Music by Hyperbubble
Additional Animation by
Quique Rivera Rivera
Isabela Dos Santos
Post-Prod done by Julian Petschek
Shot at BE∆RD H∆US
In this experiment with the Jasuto Modular synth (http://www.jasuto.com/main/) , Mark has got Noise node modulating FM of a Triangle wave and a sample. The noise node is moving slowly via motion modulation. He added an Accelerometer device which is modulating LFO FM of Square node and delay time. His iPad is held to a mic stand with The Gig Easy Mount (https://thegigeasy.com/) which allows him to move the iPad around pretty aggressively without worrying about a drop. The video also illustrates that motion is relative and even though he manually move the Noise node, it keeps moving with the recorded motion relative to where he drags it.
For more info:
Heineken are using scent with music to improve the experience of dance. With the Scenthesizer, the two claim to be ‘pushing boundaries in music and scent’.
In a nutshell, the system allows on-the-fly mixing of scents, which are then delivered on cue into the audience. The idea is to take DJing, which already incorporates sound and light, and add controlled scents to the experience.
The accelerating oscillations of the washing machine spin cycle are mimicked with the Korg Monotribe. The dual speed setting on the LFO allows the performer to push the speed up into the audio range allowing for FM synthesis like textures. This piece was shown, along with similar work, at my MFA thesis exhibition in May 2013.
NOTE: This is a binaural recording mixed with a monophonic, analog, synthesizer performance. Please use headphones to experience the binaural effect. For more info please visit http://audiocookbook.org/duet-for-syn…
This is a binaural recording mixed with a monophonic, analog, synthesizer performance. Please use headphones to experience the binaural effect.
This video created by Caleb Coppock (vimeo.com/calebcoppock) illustrates the time scope (from dusk until dawn) of the “In Habit: Living Patterns” performance at Northern Spark, June 2012. The music for the video was composed by John Keston for the sixteenth and final vignette in the sequence titled, “Energy.”
In the video I have focused on illustrating how one might use two iPad synthesizer apps and a hardware synthesizer together including Cassini, Sunrizer, and the MKS-80. The BS3X is used as both the iPad interface and MKS-80 controller. No computer is required, but a simple change of cable allows for a computer to be integrated into this setup because the MOTU UltraLite interface and standalone mixer has MIDI I/O. In other words two MIDI interfaces are still necessary with a computer, but prior to this experiment I was only using the BS3X as a controller for the MKS-80 and bypassing the class compliant USB MIDI interface functionality. Since the USB hub was required I also added the QuNexus to the setup. This was dedicated to feeding notes into the arpeggiator in Cassini. The keyboard controller was split so that in the low end I could play the MKS-80 effect then tweak it with the BS3X knobs and sliders as it decayed. In the upper end of the same keyboard I played a lead sound programmed in Sunrizer.
A sound installation that uses the Touch Board to create a ‘scientific skin’ that responds to touch with sound.
“In early 2013 Sabine Seymour of Moondial approached us and asked whether we’d be interested in developing a piece of work which could live for three months in Vienna’s MAK Museum. We asked artist Fabio Antinori, designer Alicja Pytlewska to work with us to explore the relationship between interface, interaction, printing and sound. What we created was an interactive installation which invited visitors to approach and even touch three interactive screenprints. The output of the interaction was a generative soundscape which will continue to evolve as the piece is interacted with. The project was full of technical challenges whether it was pushing the Touch Board to work with 2M high sensors, large scale printing on Tyvek or most importantly, developing a graphical language that could tie the project together. Make sure to take a look at more photos of the installation here:
Check out the Touch Board on Kickstarter!
A new project, AUUG Motion Synth, explores the possibilities the iPhone and iPod touch, converting the devices into wearable motion-controlled instruments. Using it, you can control iOS synths, desktop music apps and even hardware.
The platform is made up of three components:
The AUUG grip positions the screen of an iPhone or iPod touch to be played by the fingers and secures the device to the user’s hand during motion: The AUUG app converts your iOS device’s motion sensor data into signals for shaping sound, and transfers these signals to other iOS sound apps or external devices. The AUUG app does not produce its own sounds, but instead acts to control other iOS audio apps running on the same device (as well as external devices). The AUUG app can be installed on iPhones (4S and up) and iPod touch devices (5th generation and up). The AUUG cloud is online service at that allows AUUG app presets to be shared with other users, via your iPhone or iPod touch (no laptop or desktop needed).
Background video information:
VISIT OUR KICKSTARTER CAMPAIGN: kck.st/1efP3Rf (ending Dec 19, 2013)
VIDEO MUSICIANS AND CONTENT:
Introduction soundtrack: The second half of the video (2.12 to 4.17 mins) includes looped sections of ‘Silver Wheel’ (from the ‘Solar Driftwood’ album) by Australian band Cracked Actor crackedactor.com.au.
Samuel David Graeme [holyeucharistline.com]: (0.16 to 0.21 mins) ‘Animoog’ synth app running on same device.
Sonia Vaikyl: (0.22 to 0.33 mins) ‘Magellan Jnr’ synth app running on same device.
Video sample: The ‘video scrub’ demonstration (0.34 to 1.01 mins) by Joshua Young uses a 2 second clip of Evalena Marie [evalenamarie.com] from the independent feature film ‘Visionary’ [facebook.com/visionarymovie] directed by Ben Proulx.
Eesha Hunjon [youtube.com/channel/UCzatgvFHc7_coVeH2Tq1Klw]: (1.02 to 1.15 mins) ‘Audiobus’ app (hosting Harmony Voice and Garageband apps) running on same device, with mic input via iRig Pre.
Becki Whitton [facebook.com/aphirism]: (1.16 to 1.31 mins) ‘Audiobus’ app (hosting Magellan Jnr app) running on same device, with mic input via iRig Pre.
Adam Cook [adamjamescook.com]: (1.46 to 2.03 mins) Nord Stage 2 keyboard, with MIDI input via iRig MIDI.
Joshua Young [auug.com]: (0.00 to 0.15 mins) ‘Animoog’ synth app running on same device; (0.34 to 1.01 mins) AUUG ‘video scrub’ software running on laptop; (1.32 to 1.45 mins) Voice Live Touch 2, with MIDI input via iRig MIDI. The backing track (0.00 to 2.03 mins) each musician composed and performed their parts to and the motion-reactive visuals were also created by Joshua Young.
Process: The video musicians used the backing track as a basis to compose their own parts and performed them live during filming, while simultaneously controlling the motion-reactive visuals behind them (via Wi-Fi MIDI messages sent from the AUUG Motion Synth to laptop software and back-projected onto the screen behind them). Their estimated practice times can be found at auug.com/musician-composition-and-practice-time.
Gravitation in Time (2013)
Composition for Orchestra, Robotic Instruments, Leap motion and Electronics
Composed by GayBird
Gravity Glock (robotic instrument) concept and design: GayBird, Thomas Ip
Engineering and programming: Thomas Ip
Gravity, as an ultimate natural force on Earth, people living there may easily neglect its presence, and not to mention to its influence in our seemingly ordinary everyday experience.
We are used to take such magical force for granted, however, the potential energy that it can generate is indeed powerful. My work – Gravitation in Time, is intended to draw our attention to such force again.
To perform this piece, I have custom-made two pieces of robotic instruments – Gravity Glock. They are activated by the gravitational force, and incorporate with mechanical engineering and electronic sound processing, to provide a breath-taking audio-visual experience to audience.
Musically, the constantly progressing serpentine melodies symbolised the state of unawareness presence of gravity. The piece is structured from scattered to polyphonic texture, and developed into an integrity ensemble of repetitive melodies, that demonstrated the weight of gravity in an abstract sense.
Great Violin Concertos – Iskandar Widjaja Plays Wieniawski
Presented by Hong Kong Sinfonietta
Date: 12 October, 2013
Venue: Concert Hall, Hong Kong City Hall
Experimental play with KORG Volca beats, bass, keys and Amazon Kindle ( using kindle’s “Text to speech” function ) The text is from Volca bass’s manual
The Hot Hand USB Wireless MIDI Controller bypasses traditional knobs and faders, allowing users to manipulate software parameters with wireless motion-control. Hot Hand USB delivers an entirely new method of dynamic control to DAW (digital audio workstations) or live-performance software like Ableton Live, Traktor, Reason, Serato DJ, and Logic/MainStage.
- Wireless 3 axis motion sensor ring & USB receiver
- Driverless USB MIDI operation for Windows/Mac
- Rechargeable Lithium-ion battery for 6 hour usage time.
- Capable of transmitting over 100 feet.
- Works with any MIDI compatible audio software (Ableton, Reason, Logic, Traktor, etc).