Image Image Image Image Image

generative

01

May

[radical] signs of life

Through responsive dance, [radical] signs of life externalizes the mind’s non-hierarchical distribution of thought. Music is generated from the dancers‘ muscles and blood flow via biophysical sensors that capture sound waves from the performers’ bodies. This data triggers complex neural patterns to be projected onto multiple screens as 3D imagery. As the audience interacts with the images produced, they enter into a dialogue with the dancers.

The post [radical] signs of life appeared first on generactive :: generative + interactive + art + design.

17

Apr

Acusma | Chelpa Ferro

Chelpa Ferro is well known for squeezing a rhythmical sound from seemingly non-musical devices such as electric toothbrushes, drills, sewing machines, or juice makers, and using them in their installations and performances. At The Aldrich, the Acusma installation will fill the gallery with a sound resembling a group of people coming together to sing. However, the sound does not visually match the source, which turns out to be a series of beautiful Brazilian ceramic vases spread out on the gallery floor, with loudspeakers playing up to five different recorded voices inside each vessel.

Curator Mónica Ramírez-Montagut says, “In Chelpa Ferro’s work, the blend of high-tech equipment (speakers, cables, computers, and sophisticated computer programming) is integrated with traditional Brazilian crafts and domestic objects, providing a new and surprising visual representation of sound and conferring an aura of mystery upon these mundane objects.”

3898251745 a4e558c760 z Acusma | Chelpa Ferro

5937599563 4bd78b7f33 b Acusma | Chelpa Ferro

3898250773 4c023cb25a z Acusma | Chelpa Ferro

acusma chelpa ferro Acusma | Chelpa Ferro

The post Acusma | Chelpa Ferro appeared first on generactive :: generative + interactive + art + design.

17

Apr

Waveforms | The Art of Sound

Waveforms is an exhibition of interdisciplinary art works showcasing sound as the principle component. This exhibition explores, through an examination of creative and artistic practices, the interface of sound and new media technologies. The works in the exhibit include a number of trans-disciplinary interactions and collaborations that include sound in the context of visual and spatial artistic practices, including sound sculpture, installations, and performance works.

Featured artists include Carrie Bodle, Jacob C. Hammes, Norbert Herber/Rowland Ricketts, Tesia Kosmalski, Lou Mallozzi, Shannon McMullen/Fabian Winkler, Stephanie Rowden, and Jesse Seay. The list includes both emerging and established artists, primarily from the Midwest. Norbert Herber, artist and faculty member at Indiana University, will moderate the Symposium/discussion.

Because this exhibit is about the exchange of thought among a variety of disciplines, it will highlight the role that artistic practice and creative thinking have in our appropriation and understanding of technology, including practice and dialogue related to changes within both the natural and man-made environment. By comparing material practices in a variety of creative fields, new insights are gained into the way that artists and collaborators recognize the “new” and how the “new” becomes a building block for future discoveries.

The post Waveforms | The Art of Sound appeared first on generactive :: generative + interactive + art + design.

17

Apr

Tim Hawkinson

Tim Hawkinson is known for creating complex sculptural systems through surprisingly simple means. Inspiration for many of Hawkinson’s pieces has been the re-imagining of his own body and what it means to make a self-portrait of this new or fictionalized body. Sculptures are often re-purposed out of materials which then artist then mechanizes through hand-crafted electrical circuitry.

Tim Hawkinson is featured in the Season 2 episode “Time” of the Art21 series “Art:21 — Art in the Twenty-First Century”.

TimHawkinson drip Tim Hawkinson

large Tim Hawkinson

uberorgan1 Tim Hawkinson

uberorgan2 Tim Hawkinson

emoter detail 2 Tim Hawkinson

The post Tim Hawkinson appeared first on generactive :: generative + interactive + art + design.

17

Apr

oh my god | Sam Van Aken

This installation done by Sam Van Aken in 2006 was essentially realized through the stacking and attaching of home stereo speakers. The speakers combined into a wall that played a random sampling of the quote “oh my God” which spanned an emotional spectrum from horror to sarcasm to sheer ecstasy. Through the use of a computer program designed by the artist and a seemingly infinite amount of speaker wire and extension cord, the quote was looped to individual speakers increasing one at a time until the entire wall screamed “oh my God” in a variety of tones and contexts at the same time. Remarkably, the range of human emotions the artist captured using only this singular quote was reflected through the viewers’ laughter, disgust, tears, etc.

Sam Van Aken is represented in New York by Ronald Feldman Fine Arts and Michael Klein Arts.

Sam Van Aken’s site

OMG Monitor Detail 2 oh my god | Sam Van Aken

OMG Install Back View 1 oh my god | Sam Van Aken

OMG Install Side Front View 1 oh my god | Sam Van Aken

oh my god by sam van aken oh my god | Sam Van Aken

The post oh my god | Sam Van Aken appeared first on generactive :: generative + interactive + art + design.

16

Apr

Christoph Adami | Finding life we can’t imagine

How do we search for alien life if it’s nothing like the life that we know? At TEDxUIUC Christoph Adami shows how he uses his research into artificial life — self-replicating computer programs — to find a signature, a ‘biomarker,’ that is free of our preconceptions of what life is.

The post Christoph Adami | Finding life we can’t imagine appeared first on generactive :: generative + interactive + art + design.

16

Apr

Matthew Davidson aka Stretta

Matthew Davidson aka Stretta is a talented guy. He’s an accomplished graphic artist and video producer/editor but we talked to him about his music. Stretta’s music is lush, modest and dreamy in the tradition of Brian Eno but it definitely has character of its own. Stretta comes from a tradition of modular synthesis that led him to discover Max/MSP.

The following is reposted from an interview by Marsha Vdovin here.

Tell me a little about your background.

I grew up in Iowa, and in 1988 I was trying to figure out what college to go to. I applied to one school, and found myself out on the East Coast, at Berklee College of Music. While I was at Berklee, one of my professors introduced me to Max. That was 1992. I’ve been using Max ever since.

I was interested in music and technology, so growing up in Iowa — pre-internet—all the information I had access to was books and magazines. It’s not at all like the hot and cold running information that we have on tap these days, where you can be anywhere in the world and learn about any subject very quickly.

As an example, I recently developed an interest in photography. This is a subject I knew nothing about. With the internet, and the instant turnaround of digital photography, being able to see other people’s work, inspect the meta data, I was able to learn a great deal in a short amount of time. Today, taking up any new interest like electronic music is far easier than when I started, I can tell you that much.

Digital photography opened up so much for me. I was able to do it without spending money, which was incredible. People were able to see my work all over the world, without me spending money.

Right. The spending money thing is analogous to what life was like before digital recording. A reel of tape costs money, so when you’re rolling, when you’ve hit the record button, there’s money at stake. That was the same thing with photography, you’re burning film. Now it doesn’t cost anything to drop the shutter, and now it doesn’t cost anything to play with digital audio. This accelerates the learning process.

I loved your Way-Geeky Time Line.

[Laughs.] You’ve done your research. Looking back, I realized that computers helped me express myself, so it was the correlation of operating systems or computers and what was happening in my life was significant. My first computer at home that I had access to was an Apple Lisa. The first time I used it, it was like touching the future. It was like someone got in a time machine, kidnapped a computer, then brought it back to the current day. I’d never experienced anything like that before.

I guess you could apply the oft-used term “paradigm shift.” I hate to use that word, but I can’t really think of anything better to describe what it was like going from computers with a green phosphorous screen to a black-and-white bitmap display where you click on objects and open them up. It’s not hyperbole to say that that changed my life.

I only had that machine for three months, then it was replaced by a 128K Macintosh. I was definitely one of the very early Mac users, and I’ve been fortunate in my choice of careers and work, as I’ve never had to use a Windows machine. Even in the dark days of the ’90s.

Did you take to Max right away?

I remember the night I was exposed to Max. Afterwards, I stood outside Berklee and put my head back and looked up at the sky, imagining how far this thing went. I recognized it and I knew it was one of those things I could spend years playing with, and never really see the end of the potential.

I’m very fortunate to watch Max evolve, sprouting audio, making all these technological leaps, and then the leap to OS X. It continues to be more capable, while retaining its essential core.

My favorite toy growing up was Legos, and I see a commonality. People I talk to who are into modular synthesizers, or into Max, there’s this commonality of “Did you play with Legos when you grew up?” “Yeah.” So it’s like that. It’s like Legos for music.

I like that granularity of control. It sits in this weird space, between commercial music applications and programming languages. Max is somewhere in between these two things. It allows you to create and customize your environment without programming and compiling.

I’m not a programmer—there is something about procedural languages, text-based, linear thinking that I don’t get along with. Max is non-linear, it moves in all directions, it’s real time. If you’re a guitarist, you understand how guitar pedals and patch cords work. You plug this into this and this other thing. I think this is a metaphor that is compatible with musicians.

If you understand these things, then understanding Max comes intuitively. When your creations evolve, and they tend to get more complex, you look back at it and you think, “How did I even understand this to begin with?” Because it looks really complicated. But then you break it down into smaller parts, and you can see how everything works.

So, is Max your primary music-making tool?

No. I would be surprised to hear anyone say that it is, simply because we live in this age where we have so many amazing tools available to musicians. There has never been a better time, from a technological standpoint, to be a musician. So while there are people who can dedicate themselves monk-like to a particular tool—Charles Cohen comes to mind. He’s been using a Buchla Music Easel for forty years. That’s his thing, and he knows it inside out. I admire that. We need people like that to be able to dedicate themselves to an instrument, but I don’t have that kind of dedication.

Have you gone the Jitter road? Have you combined your photography with Max?

[Laughs.] No. Like I said, Max is one of those things that you could spend the rest of your life dedicating yourself to the possibilities, and not exhaust them all. Based on my interest in video, and photography, I am definitely interested in in Jitter, but I haven’t come close to exhausting all the ideas I have for audio and MIDI within Max yet.

I think if someone came to me and said, “You know, we want you to do a live performance, and we want there to be video,” yeah, [laughs] I would fast-track my Jitter education.

Also, I don’t think there’re enough video-y applications for the Monome. The Monome is very audio-centric right now and there isn’t any good reason for that. The Monome, in conjunction with Jitter, would be very powerful.

What is it about the Monome that draws you to it?

Probably it was all the years of Max prior to it. You spend all this time with Max, and then you think to yourself, “Gosh, I really wish I had a controller to go along with this, to provide input and feedback.” People would come out with controllers, and they would be overly specific, or they wouldn’t do the thing that you wanted to do.

Then you started seeing people building their own controllers. Do-it-yourself kits became available, like the iCube, where you could hook up sensors and other analog sources and it would provide a MIDI output. That was a good move forward.

But when I saw the Monome, I just thought, “Oh, of course. I know exactly what I would do with that.” I think that’s partly why Max has been the default language of choice for Monome developers. They’re very well suited for each other. There are no labels of any sort, there’s no pre-determined, prescribed usage to the Monome. It is exactly what you’re looking for if you’d done anything in Max at all in the past.

How did the Max 5 change affect you?

I was using Max 4 up until about two or three months ago. I knew about Max 5, I knew what was going on with the environment, and I thought it was a very necessary, gutsy move for the company. And from what I could tell, at least two solid years of engineering, while adding no new features or capabilities to the software, redoing the user interface from the ground up, with a completely new framework. That’s the right way to do things.

If they were a larger company, they would find a way to screw it up. “You want to do what? For how long? That’s ridiculous.” But the change from Max 4 to Max 5 is as significant as the change OS 9 to OS X.

I was talking to Nick Rothwell as recently as September, telling him that I think it’s time for me to move into Max 5. He said, “Well, once you start using Max 5, you’re never going to go back.” Intellectually I believed him, but deep down, I was like, “Yeah, well we’ll just see about that”— because it is a big change. And oh, he was right. [laughs] I have a Max 5 license on one computer and a Max 4 license on another computer. I can’t bring myself to use Max 4 anymore.

I took to it in a fairly short amount of time. I think the main change, in terms of capabilities for Max 5, is being able to think in metric units. You can think in terms of 16th notes and 8th notes, and you don’t have to worry about milliseconds, or converting this to samples. That makes everything a lot easier. The idea of a global transport, and having access to metrical units is a really big deal for me. That was huge.

What’s your favorite object?

The Coll object.

And why is that?

I use it in every single patch. It’s familiar, like an old friend. I know it, and I know how to use it. I’m constantly learning new things about it. I think if you’re doing anything that manipulates or stores little bits of data, you have to get comfortable with the Coll object.

It seems to be pretty fast. I don’t have any problem extracting data in a timely fashion from it. If you have a Coll object and a Metro, you have the entire basis of a whole variety of step sequencers with a timed beat. You can do all sorts of magic with just those two things.

I don’t think it’s very sexy if you look at it. The object that I really liked before the Coll object was Table. It was more limited and approachable than Coll, but it had a graphical interface. You had two-dimensional data that you could manipulate directly with the mouse. But the Coll object is a lot more flexible. With the Monome, the face of the Coll object is now tangible.

Often, I’ll peek inside patches of other developers to see how they do things. Sometimes, they’re doing some sort of complex mathematical abstraction, which is satisfying from an intellectual point of view, but I’m more likely to simply dump the values I want into a data object like Coll. It kind of feels like cheating, but it gets the job done.

Have you been working in Max for Live?

Yeah. Most of the work I’ve been doing recently has been in Max for Live.

I think if Max by itself had a weak point, it would be that it doesn’t have a decent time line. A time line is one of those features that represents infinite mission creep. Ultimately, what you want is a full-featured DAW. So, putting Max inside a mature DAW is the best solution here.

Prior to Max for Live, most of the things I made were only of interest only to me, due to the dependencies involved. In the beginning, the dependencies were racks of hardware. At Berklee, I had codified what I learned about harmony into software, but to make it do anything you had to use external synthesizers and sound generators. Nothing ever made it out of the lab.

Later, you could use soft synths, but that still involved a lot of setup. You had to load the virtual instruments and effects, perform complicated routings, and deal with sync issues. It wasn’t really plug-and-play. I couldn’t take this, and then give it to someone else, and have it be as useful for them.

Now with Max for Live, suddenly the things that I make are portable to other people. I can make these little tools, these little performance things that take real time input, and then outputs something that’s musically interesting.

That also has ramifications for live performance. I did a recent video using Max 5. It involved a software harmonizer, effects, recording multiple tracks into a DAW, and complex MIDI routings, and that’s like, four different applications, all combined. It took a good hour or so to set this one performance piece up. So, it’s not easy for me to reproduce that performance again, let alone string together a set of pieces to perform. Now, with Max for Live, you can put all these combinations of elements together, all of your soft synths, all your routing, all your effects in this one environment, and save it. Then you can recall it. I can’t tell you what a huge thing that is.

Max for Live also addresses the issue of a DAW trying to be all things to all people…

But they try to be.

Well, they try to be, and then that’s where the user interface breaks down. The application sprouts these weird appendages, and after two years of that and you end up with something that becomes incomprehensible and un-maintainable. Especially if you’re not willing to take the time to go in and refine the user interface, or piss off your existing user base by throwing out old, crusty features that a small percentage of your user base relies on. But if you jettisoned that code, then you could bring your DAW forward, develop faster and make your code more reliable.

So what Max for Live does for Live users is it allows people to create this customized environment to do the things that they need to do, without bringing the entire DAW down.

So you can see how Max and Live need each other. Max gets a fully featured timeline, and Live gets a mature environment for user customization.

So, you’re giving away your Max for Live ‘Monome suite’?

It’s free for anyone to download. I’m beta testing a new release right now that adds support for multiple Monomes. So if you have multiple Monomes, you can have one that’s switching between these applications, and another one that’s switching between another set of applications.

I’m replacing all the user interface objects with Live objects, and that enables parameters to be stored and automated. I just sent out a beta of that yesterday, and I’ll hopefully be getting some bug reports and actually making that an official release in the near future.

So, what is Stretta?

Stretta began as a vanity record label. I bought the domain back around 1996. It became clear to me that people weren’t buying music, so the idea of a record label really didn’t make sense anymore. Simultaneously, I was noticing the importance of personal branding on the internet because there are so many forces competing for attention. If you release something, you’ll see a huge spike of interest that falls off rapidly. It doesn’t matter if you spent two years working on something or two days, you’ll see the same spike, then everyone moves on to the next thing. From that I concluded that the better strategy is to release smaller things on a more consistent basis, and this is where having a memorable brand becomes useful.

‘Matthew Davidson’ is not very memorable, and it is kind of long. So, since I already had the domain—and short, pronounceable domain names are a rare commodity these days—I use Stretta. It is short and memorable and consistent across all these social media platforms.

The post Matthew Davidson aka Stretta appeared first on generactive :: generative + interactive + art + design.

16

Apr

20Hz by Semiconductor

Semiconductor is artist duo Ruth Jarman and Joe Gerhardt. Through moving image works they explore the material nature of our world and how we experience it, questioning our place in the physical universe. Their unique approach has won them many awards and prestigious fellowships such as the Gulbenkian Galapagos, Smithsonian Artists Research and the NASA Space Sciences. Their work is part of several international public collections and has been exhibited globally including Venice Bienniale, The Royal Academy, Hirshhorn Museum, BBC, ICA and the Exploratorium.

20Hz observes a geo-magnetic storm occurring in the Earth’s upper atmosphere. Working with data collected from the CARISMA radio array and interpreted as audio, we hear tweeting and rumbles caused by incoming solar wind, captured at the frequency of 20 Hertz. Generated directly by the sound, tangible and sculptural forms emerge suggestive of scientific visualisations. As different frequencies interact both visually and aurally, complex patterns emerge to create interference phenomena that probe the limits of our perception


05.00 minutes / HD / 2011
HD single channel and HD 3D single channel.
A Semiconductor work by Ruth Jarman and Joe Gerhardt.
Audio Data courtesy of CARISMA, operated by the University of Alberta, funded by the Canadian Space Agency. Special thanks to Andy Kale
20Hz is co-commissioned by Arts Santa Monica + Lighthouse . Supported by the British Council.
Commissioned for the Invisible Fields Exhibition at Arts Santa Monica, Barcelona. 2011-2012.

Semiconductor 20Hz 9 20Hz by Semiconductor

Semiconductor 20Hz 2 20Hz by Semiconductor

Semiconductor 20Hz 10 20Hz by Semiconductor

Semiconductor 20Hz 4 20Hz by Semiconductor

The post 20Hz by Semiconductor appeared first on generactive :: generative + interactive + art + design.

15

Apr

Nervous System Generative Design Studio

Nervous System is a design studio founded in 2007 by Jessica Rosenkrantz and Jesse Louis-Rosenberg. They work at the intersection of science, art, and technology, and create using a novel process that employs computer simulation to generate designs and digital fabrication to realize products. Drawing inspiration from natural phenomena, Jesse & Jessica write computer programs mimicking processes and patterns found in nature and use those programs to create unique and affordable art, jewelry, and housewares. In this talk at Eyeo they discuss their obsession with the way patterns form in nature and their attempts to adapt those methods for design.

http://www.shapeways.com/shops/nervous

Hyphae Nervous Inside 01 150x150 Nervous System Generative Design Studio 0721nervous2 150x150 Nervous System Generative Design Studio work nervous portrait Ken Richardson 150x150 Nervous System Generative Design Studio steelBracelets blackRadiolaria large 150x150 Nervous System Generative Design Studio h11b vesselPendant black model large 150x150 Nervous System Generative Design Studio 0408nervous1 150x150 Nervous System Generative Design Studio 5709019163 0ab3c4ef4e b 150x150 Nervous System Generative Design Studio nervous system deconstruction cell cycle 1 150x150 Nervous System Generative Design Studio 20071120 IMGP0971 large 150x150 Nervous System Generative Design Studio

The post Nervous System Generative Design Studio appeared first on generactive :: generative + interactive + art + design.