Iraqi born Wafaa Bilal has become known for provocative interactive video installations. Many of Bilal’s projects over the past few years have addressed the dichotomy of the virtual vs. the real. He attempts to keep in mind the relationship of the viewer to the artwork, one of his main objectives being to transform the normally passive experience of viewing art into an active participation. In this, his latest effort, Domestic Tension, viewers can log onto the internet to contact or “shoot” Bilal with paintball guns. Bilal’s objective is to raise awareness of virtual war and privacy, or lack thereof, in the digital age. During the course of the exhibition, Bilal will confine himself to the gallery space. Over the duration, people will have 24-hour virtual access to the space via the Internet. They will have the ability to watch Bilal and interact with him through a live web-cam and chat room. Should they choose to do so, viewers will also have the option to shoot Bilal with a paintball gun, transforming the virtual experience into a very physical one. Bilal’s self imposed confinement is designed to raise awareness about the life of the Iraqi people and the home confinement they face due to the both the violent and the virtual war they face on a daily basis. This sensational approach to the war is meant to engage people who may not be willing to engage in political dialogue through conventional means. Domestic Tension will depict the suffering of war not through human displays of dramatic emotion, but rather through engaging people in the sort of playful interactive video game with which they are familiar.
The post Domestic Tension | Wafaa Bilal appeared first on generactive :: generative + interactive + art + design.
This installation done by Sam Van Aken in 2006 was essentially realized through the stacking and attaching of home stereo speakers. The speakers combined into a wall that played a random sampling of the quote “oh my God” which spanned an emotional spectrum from horror to sarcasm to sheer ecstasy. Through the use of a computer program designed by the artist and a seemingly infinite amount of speaker wire and extension cord, the quote was looped to individual speakers increasing one at a time until the entire wall screamed “oh my God” in a variety of tones and contexts at the same time. Remarkably, the range of human emotions the artist captured using only this singular quote was reflected through the viewers’ laughter, disgust, tears, etc.
Sam Van Aken is represented in New York by Ronald Feldman Fine Arts and Michael Klein Arts.
Sam Van Aken’s site
The post oh my god | Sam Van Aken appeared first on generactive :: generative + interactive + art + design.
The Cabinet of Curiousness is an antique wooden card catalogue with 20 drawers. Functioning as an interactive piece, the opening of each drawer activates a voice or piece of music from within the cabinet. The audience, assuming the role of a DJ, may experience the clarity of sound from one drawer or a cacophony of sounds from numerous drawers opened simultaneously as the cabinet is played like an instrument. A contrast emerges between the obsolete system of cataloguing single pieces of data and our current tendency to inundate ourselves with excessive information. An investigation of knowledge, time, and our relationship to objects and music.
Work by Janet Cardiff & George Miller
Materials: Unique oak card catalogue with speakers and audio
Dimensions: 52 X 17 1/2 X 27 inches (132.08 X 44.45 X 68.58 cm)
The post The Cabinet of Curiousness appeared first on generactive :: generative + interactive + art + design.
How do we search for alien life if it’s nothing like the life that we know? At TEDxUIUC Christoph Adami shows how he uses his research into artificial life — self-replicating computer programs — to find a signature, a ‘biomarker,’ that is free of our preconceptions of what life is.
The post Christoph Adami | Finding life we can’t imagine appeared first on generactive :: generative + interactive + art + design.
Installation and Performance
Fabricating the interfaced machine.
Interview created and produced by Sue Costabile for Cycling ’74.
The post Barney Haynes Installation & Performance appeared first on generactive :: generative + interactive + art + design.
Matthew Davidson aka Stretta is a talented guy. He’s an accomplished graphic artist and video producer/editor but we talked to him about his music. Stretta’s music is lush, modest and dreamy in the tradition of Brian Eno but it definitely has character of its own. Stretta comes from a tradition of modular synthesis that led him to discover Max/MSP.
The following is reposted from an interview by Marsha Vdovin here.
Tell me a little about your background.
I grew up in Iowa, and in 1988 I was trying to figure out what college to go to. I applied to one school, and found myself out on the East Coast, at Berklee College of Music. While I was at Berklee, one of my professors introduced me to Max. That was 1992. I’ve been using Max ever since.
I was interested in music and technology, so growing up in Iowa — pre-internet—all the information I had access to was books and magazines. It’s not at all like the hot and cold running information that we have on tap these days, where you can be anywhere in the world and learn about any subject very quickly.
As an example, I recently developed an interest in photography. This is a subject I knew nothing about. With the internet, and the instant turnaround of digital photography, being able to see other people’s work, inspect the meta data, I was able to learn a great deal in a short amount of time. Today, taking up any new interest like electronic music is far easier than when I started, I can tell you that much.
Digital photography opened up so much for me. I was able to do it without spending money, which was incredible. People were able to see my work all over the world, without me spending money.
Right. The spending money thing is analogous to what life was like before digital recording. A reel of tape costs money, so when you’re rolling, when you’ve hit the record button, there’s money at stake. That was the same thing with photography, you’re burning film. Now it doesn’t cost anything to drop the shutter, and now it doesn’t cost anything to play with digital audio. This accelerates the learning process.
I loved your Way-Geeky Time Line.
[Laughs.] You’ve done your research. Looking back, I realized that computers helped me express myself, so it was the correlation of operating systems or computers and what was happening in my life was significant. My first computer at home that I had access to was an Apple Lisa. The first time I used it, it was like touching the future. It was like someone got in a time machine, kidnapped a computer, then brought it back to the current day. I’d never experienced anything like that before.
I guess you could apply the oft-used term “paradigm shift.” I hate to use that word, but I can’t really think of anything better to describe what it was like going from computers with a green phosphorous screen to a black-and-white bitmap display where you click on objects and open them up. It’s not hyperbole to say that that changed my life.
I only had that machine for three months, then it was replaced by a 128K Macintosh. I was definitely one of the very early Mac users, and I’ve been fortunate in my choice of careers and work, as I’ve never had to use a Windows machine. Even in the dark days of the ’90s.
Did you take to Max right away?
I remember the night I was exposed to Max. Afterwards, I stood outside Berklee and put my head back and looked up at the sky, imagining how far this thing went. I recognized it and I knew it was one of those things I could spend years playing with, and never really see the end of the potential.
I’m very fortunate to watch Max evolve, sprouting audio, making all these technological leaps, and then the leap to OS X. It continues to be more capable, while retaining its essential core.
My favorite toy growing up was Legos, and I see a commonality. People I talk to who are into modular synthesizers, or into Max, there’s this commonality of “Did you play with Legos when you grew up?” “Yeah.” So it’s like that. It’s like Legos for music.
I like that granularity of control. It sits in this weird space, between commercial music applications and programming languages. Max is somewhere in between these two things. It allows you to create and customize your environment without programming and compiling.
I’m not a programmer—there is something about procedural languages, text-based, linear thinking that I don’t get along with. Max is non-linear, it moves in all directions, it’s real time. If you’re a guitarist, you understand how guitar pedals and patch cords work. You plug this into this and this other thing. I think this is a metaphor that is compatible with musicians.
If you understand these things, then understanding Max comes intuitively. When your creations evolve, and they tend to get more complex, you look back at it and you think, “How did I even understand this to begin with?” Because it looks really complicated. But then you break it down into smaller parts, and you can see how everything works.
So, is Max your primary music-making tool?
No. I would be surprised to hear anyone say that it is, simply because we live in this age where we have so many amazing tools available to musicians. There has never been a better time, from a technological standpoint, to be a musician. So while there are people who can dedicate themselves monk-like to a particular tool—Charles Cohen comes to mind. He’s been using a Buchla Music Easel for forty years. That’s his thing, and he knows it inside out. I admire that. We need people like that to be able to dedicate themselves to an instrument, but I don’t have that kind of dedication.
Have you gone the Jitter road? Have you combined your photography with Max?
[Laughs.] No. Like I said, Max is one of those things that you could spend the rest of your life dedicating yourself to the possibilities, and not exhaust them all. Based on my interest in video, and photography, I am definitely interested in in Jitter, but I haven’t come close to exhausting all the ideas I have for audio and MIDI within Max yet.
I think if someone came to me and said, “You know, we want you to do a live performance, and we want there to be video,” yeah, [laughs] I would fast-track my Jitter education.
Also, I don’t think there’re enough video-y applications for the Monome. The Monome is very audio-centric right now and there isn’t any good reason for that. The Monome, in conjunction with Jitter, would be very powerful.
What is it about the Monome that draws you to it?
Probably it was all the years of Max prior to it. You spend all this time with Max, and then you think to yourself, “Gosh, I really wish I had a controller to go along with this, to provide input and feedback.” People would come out with controllers, and they would be overly specific, or they wouldn’t do the thing that you wanted to do.
Then you started seeing people building their own controllers. Do-it-yourself kits became available, like the iCube, where you could hook up sensors and other analog sources and it would provide a MIDI output. That was a good move forward.
But when I saw the Monome, I just thought, “Oh, of course. I know exactly what I would do with that.” I think that’s partly why Max has been the default language of choice for Monome developers. They’re very well suited for each other. There are no labels of any sort, there’s no pre-determined, prescribed usage to the Monome. It is exactly what you’re looking for if you’d done anything in Max at all in the past.
How did the Max 5 change affect you?
I was using Max 4 up until about two or three months ago. I knew about Max 5, I knew what was going on with the environment, and I thought it was a very necessary, gutsy move for the company. And from what I could tell, at least two solid years of engineering, while adding no new features or capabilities to the software, redoing the user interface from the ground up, with a completely new framework. That’s the right way to do things.
If they were a larger company, they would find a way to screw it up. “You want to do what? For how long? That’s ridiculous.” But the change from Max 4 to Max 5 is as significant as the change OS 9 to OS X.
I was talking to Nick Rothwell as recently as September, telling him that I think it’s time for me to move into Max 5. He said, “Well, once you start using Max 5, you’re never going to go back.” Intellectually I believed him, but deep down, I was like, “Yeah, well we’ll just see about that”— because it is a big change. And oh, he was right. [laughs] I have a Max 5 license on one computer and a Max 4 license on another computer. I can’t bring myself to use Max 4 anymore.
I took to it in a fairly short amount of time. I think the main change, in terms of capabilities for Max 5, is being able to think in metric units. You can think in terms of 16th notes and 8th notes, and you don’t have to worry about milliseconds, or converting this to samples. That makes everything a lot easier. The idea of a global transport, and having access to metrical units is a really big deal for me. That was huge.
What’s your favorite object?
The Coll object.
And why is that?
I use it in every single patch. It’s familiar, like an old friend. I know it, and I know how to use it. I’m constantly learning new things about it. I think if you’re doing anything that manipulates or stores little bits of data, you have to get comfortable with the Coll object.
It seems to be pretty fast. I don’t have any problem extracting data in a timely fashion from it. If you have a Coll object and a Metro, you have the entire basis of a whole variety of step sequencers with a timed beat. You can do all sorts of magic with just those two things.
I don’t think it’s very sexy if you look at it. The object that I really liked before the Coll object was Table. It was more limited and approachable than Coll, but it had a graphical interface. You had two-dimensional data that you could manipulate directly with the mouse. But the Coll object is a lot more flexible. With the Monome, the face of the Coll object is now tangible.
Often, I’ll peek inside patches of other developers to see how they do things. Sometimes, they’re doing some sort of complex mathematical abstraction, which is satisfying from an intellectual point of view, but I’m more likely to simply dump the values I want into a data object like Coll. It kind of feels like cheating, but it gets the job done.
Have you been working in Max for Live?
Yeah. Most of the work I’ve been doing recently has been in Max for Live.
I think if Max by itself had a weak point, it would be that it doesn’t have a decent time line. A time line is one of those features that represents infinite mission creep. Ultimately, what you want is a full-featured DAW. So, putting Max inside a mature DAW is the best solution here.
Prior to Max for Live, most of the things I made were only of interest only to me, due to the dependencies involved. In the beginning, the dependencies were racks of hardware. At Berklee, I had codified what I learned about harmony into software, but to make it do anything you had to use external synthesizers and sound generators. Nothing ever made it out of the lab.
Later, you could use soft synths, but that still involved a lot of setup. You had to load the virtual instruments and effects, perform complicated routings, and deal with sync issues. It wasn’t really plug-and-play. I couldn’t take this, and then give it to someone else, and have it be as useful for them.
Now with Max for Live, suddenly the things that I make are portable to other people. I can make these little tools, these little performance things that take real time input, and then outputs something that’s musically interesting.
That also has ramifications for live performance. I did a recent video using Max 5. It involved a software harmonizer, effects, recording multiple tracks into a DAW, and complex MIDI routings, and that’s like, four different applications, all combined. It took a good hour or so to set this one performance piece up. So, it’s not easy for me to reproduce that performance again, let alone string together a set of pieces to perform. Now, with Max for Live, you can put all these combinations of elements together, all of your soft synths, all your routing, all your effects in this one environment, and save it. Then you can recall it. I can’t tell you what a huge thing that is.
Max for Live also addresses the issue of a DAW trying to be all things to all people…
But they try to be.
Well, they try to be, and then that’s where the user interface breaks down. The application sprouts these weird appendages, and after two years of that and you end up with something that becomes incomprehensible and un-maintainable. Especially if you’re not willing to take the time to go in and refine the user interface, or piss off your existing user base by throwing out old, crusty features that a small percentage of your user base relies on. But if you jettisoned that code, then you could bring your DAW forward, develop faster and make your code more reliable.
So what Max for Live does for Live users is it allows people to create this customized environment to do the things that they need to do, without bringing the entire DAW down.
So you can see how Max and Live need each other. Max gets a fully featured timeline, and Live gets a mature environment for user customization.
So, you’re giving away your Max for Live ‘Monome suite’?
It’s free for anyone to download. I’m beta testing a new release right now that adds support for multiple Monomes. So if you have multiple Monomes, you can have one that’s switching between these applications, and another one that’s switching between another set of applications.
I’m replacing all the user interface objects with Live objects, and that enables parameters to be stored and automated. I just sent out a beta of that yesterday, and I’ll hopefully be getting some bug reports and actually making that an official release in the near future.
So, what is Stretta?
Stretta began as a vanity record label. I bought the domain back around 1996. It became clear to me that people weren’t buying music, so the idea of a record label really didn’t make sense anymore. Simultaneously, I was noticing the importance of personal branding on the internet because there are so many forces competing for attention. If you release something, you’ll see a huge spike of interest that falls off rapidly. It doesn’t matter if you spent two years working on something or two days, you’ll see the same spike, then everyone moves on to the next thing. From that I concluded that the better strategy is to release smaller things on a more consistent basis, and this is where having a memorable brand becomes useful.
‘Matthew Davidson’ is not very memorable, and it is kind of long. So, since I already had the domain—and short, pronounceable domain names are a rare commodity these days—I use Stretta. It is short and memorable and consistent across all these social media platforms.
The post Matthew Davidson aka Stretta appeared first on generactive :: generative + interactive + art + design.
An interactive installation by Quiet Ensemble that have mice running wheels playing music boxes. I love the low tech sophistication of this piece. While they run around they can play a lullaby by Brahms, Schubert or Mozart.
The post Orchestra Da Camera appeared first on generactive :: generative + interactive + art + design.
Korinsky Studio consists of Abel, Carlo and Max Korinsky. They mainly focus on their shared passion: exploring the possibilities of using sound in vertical surfaces. 3845 m/s is their newest installation using their own software, in a former coal power plant in Berlin. See the Korinsky Studio website for more information about their work.
Documentary about the work of Berlin-based art collective “Korinsky – Atelier für vertikale Flächen” and their sound installation 3845 m/s
The post Korinsky Studios 3845 m/s Sound Installation appeared first on generactive :: generative + interactive + art + design.
Building Sound is a project instigated by Ella Finer and Fabrizio Manco, PhD candidates at Roehampton University, London.
Building Sound is an on-going research experiment in ways to describe and articulate experiences of sound making and reception within theatre, theory and practice.
As both doctoral studies are concerned with an interrogation of sound within theatrical space, the aim of building sound is to provide thinkers and practitioners an opportunity to offer their own ideas from their respective practices about sound within an actual and a virtual space.
These sites in which voices will interweave with their own particular social and cultural definitions of aurality will it is hoped create an open forum in which ideas can complement, collide and construct.
As well as hosting information about the project and storing archival sound from the symposium, the form of this website itself is the result of an attempt to question how to build a website beyond text.
Both the symposium and website are investigations into interdisciplinary dialogues about working with sound.
The post Building Sound Research Experiment appeared first on generactive :: generative + interactive + art + design.
Omnivisu was a temporary interactive installation which took place at the S/U station Warschauer Str. from july 7th through 17th — Tower of Light in Oberbaum city
The S/U station Warschauer street is one of Berlin’s most important interchange stations, especially at night. From the Warschauer bridge, a wide panorama over the center of Berlin presents itself and near the bank of the river Spree, where the Berlin wall used to run. Here the characteristic tower of the former light–bulb industry Narva rises. This is not only a symbol of the desolated GDR–industries, but also a relict of the new–economy boom in berlin and its ending as well as the arrival of media corporations like MTV and Universal.
A gazing tower
The unmistakable landmark of the area is transformed into a building with human character, equipped with the eyes of the people who interact with the installation. They can participate directly and in real-time through a showcase which is placed on the busy site of the bridge. Once somebody looks into it, their eyes are filmed. The video signal is transmitted and projected on the facade of the building. A big brother who sees the world with your eyes.
The post Omnivisu by Richard The appeared first on generactive :: generative + interactive + art + design.