The project uses the gest,js library to detect up, down, left and right hand gestures and that triggers sound files to start, stop or add an effect. This one three different computers was playing percussion, a string instrument names sarang and a flute but ideally we could add more instruments ergo musicians. These musicians are instructed by a conductor.
We like web pages and web applications. The experience of the modern web browsing is within a browser window or a window tab, which indeed is an user-centered and practical service design.
Web pages are like paintings, the browser window is the canvases and pigments of the paintings. We focus more on the contents made out of canvas and pigments, but we might rarely notice the canvas and pigments themselves also have great stories to tell.
Browser lab is an experiment project that treats browser windows as the contents of expression, the material of design and the storyteller of emotion, not only the canvas. In browser lab, instead of interacting with HTML elements within only one window, users build a connection with the browser windows themselves directly. It's a playground to gain more knowledge about browser window through playing with its virtual physical attributes, like size, position, color, layout and sound.
In browser lab, 3 scenes were built separately to show the novel experience under the essence of chrome browser.
The first scene is a color grid layout to play with multi-window interfaces, trying to experiment the color and sound of windows, and how they might be organized as a group.
The second scene is a touch interaction with one window, trying to experiment the emotion and physical connections between the browser and human beings.
The third scene is a browser window overlay sculpture, trying to experiment capacity for storytelling of browser windows. More scenes and experiment are being added.
This project offers the viewer a series of different musical interfaces (software) and allows them to interact with both the audio and the visual elements which are being projected onto a screen. The idea is to provide the audience with a multi-sensory AV experience by tightly weaving sound, graphics, and physical gestures (via mouse/wacom/drum pads/faders).
Data Art, Interactive Music, Nature of Code (Animation and Physics), Nature of Code (Intelligence and Learning), Shader Studio
My partner Or Fleisher and I have populated a desert landscape with variously shaped bounceable objects where each object is tied to an array of ambient music sounds. Related sonic clusters of sound which span the eq spectrum are grouped and tied to objects, and the sounds are triggered as their corresponding object bounces on the ground. By bouncing the object harder it travels higher, effectively stretching the cycle length of each array’s loop. And as each object pulls randomly from an array of note values, unpredictable melodies and combinations of harmonious sounds unfold and play out as the user travels around the desert tending to their dynamic soundscape in motion.
The Max Patch opens the webcam feed and uses a low resolution. The jit.change object filters out repeating frames and the jit.rgb2luma converts ARGB color to Black and White. I used the jit.scissors object to divide the frame into six equal regions. The jit.3m object reports minimum, mean and maximum values from the webcam feed and the < 100 object compares the value. If the value is less than 100 it sends a bang to trigger a sound. I have mapped several sounds to some regions to give more intuitive control to the gesture in sculpting the music. To make the sounds visually expressive, I integrated a generative aspect that emits particles each time the region of the frame is triggered. I layered the particles with some visuals that further enhance the compelling interaction of the gesture with sound. The visuals are composed of FFTs and glowing violet hues. When you play the touch-less instrument, the visual feedback is effectively responding to the gesture in front of the webcam.