There is a lot of great pseudoscience around what one’s body resistance or frequency can reveal/do, or how it can be changed. Some examples include:
Like my automaton, I wanted to work with moiré patterns again. I simply would like to make a visualization of the participant’s body resistance, or thats all I have right now, itll get more flesh to it as I work on it. Heres a prototype of what the sketch will generate:
The color is generated by taking the average color from the participant using the kinect’s camera. Im also using the kinect so it can stand by itself as an installation, and will only be activated when someone is standing riiight in front of it. As the person is standing in front of it, the kinect will be taking in information about that person. After the person walks away (as detected from the kinect) it will generate the picture using all the information it gathered and post it to a tumblr blog, with a number so the participant can see which picture represents them.
Here are some links that gave me the idea for this thing:
An article about the history of privacy and surveillance and the NSA
Angels with a bunch of faces
Art about modern surveillance
Art about being watched by bots and algorithms
The Panopticon is an installation that simulates a digital creature with a hundred-sided polygon of faces as a body. It steals the faces of those who view it (using the depth and color data from a kinect) and uses them as tools of surveillance, acting as the center of a giant spherical panopticon. The database of faces is never purged, and you cannot remove yourself from the sphere once you are a part of it.
With this project, I want to try to take the unseen and secret processes of modern surveillance and have them manifested as a mythical digital object/creature.
I also want to explore the idea of “surveillance by identity” with this project. Today, surveillance is not just conducted by physical cameras/microphones, but also by the mass collection of data and the use of algorithms to parse the identity of people in the database and analyze them for threats.
I’ve been recently interested in the text-based gaming revival started by Porpentine. Porpentine created this platform called Twine where anyone is able to create their own text-based video game (some examples http://aliendovecote.com/intfic.html). The game borders an immersion of story/plot and contemporary poetic elements.
After thinking about this, I was inspired to create a sort of facial simplification via prose by creating a software that would recognize elements of your face (level of eyebrows, face structure, expression, facial direction, colour of shirt below face). The data would then give the participant a small snippet of prose based upon these elements.
I know that I would like to use passive audio recordings in my final project but I’m not exactly sure what I will do with them. My current best idea, seeks to primitively figure out when “interesting audio” is being made in the environment my object is placed in. When it believes interesting things are happening it sends the audio to be evaluated by a mechanical turk where they will decide what the “tone” of the audio is. For example the Turkers will respond with an “emotion” which corresponds to the audio. That emotion will be parsed into some kind of visual or audio response which can be recorded throughout the day.
Here’s an odd sketch.
Using a fog machine, projectors, and lenses, I hope to make an animated light sculpture in which I will narrate a series of events from my past using spirit-like figures.
Window into another world
In this I will create a user interactive scene display. I will use a kinect or other similar device to track user movement to make it appear that the user is looking through a window rather than at a rear projection screen, monitor, or TV. The scene I would project would change over time in order to narrate a small story.
Both ideas utilize the same mechanism set up diagramed above.
Sound and Body
I got this idea when we were first learning how to use the sound coding software of Maxus. I wanted to control the pitch and tone of a sound with the movement of my body. I want the sound to change when I bend my fingers and tilt my hand. Therefore, I would require the a flex sensor and a tilt sensor from Adafruit. For the sound one, I would only use one flex sensor to indicate volume and the tilt sensor to indicate pitch. As one would turn their palm upward, the pitch would lower and as the palm was turn downward, the pitch would become higher. And, as the pointer finger flexed, the the volume would lower and as the finger relaxed the sound would become louder.
Color and Body
This idea is similar to the sound one except it would use Processing to create interesting visuals. I would require three flexors one to control the red value, one the green value, and one the blue value. The tilt sensor would then be used to control the tone. As the fingers flexed, the values of RGB would decrease and as they relaxed the values would increase. Therefore, a clenched fist would create black and a relaxed hand could create white. The reason I came up with these ideas is because I am interested in creating an intimacy with a viewer and forcing them to move thus making the viewer more aware of their own control over their own physical actions.
My original idea was inspired by my looking outwards post, RUAH by Giulia Tomasello. Instead of emphasising the importance of diaphragmatic breathing, I wanted to explore the interaction between two people and flowers blooming. Like free hug that was once popular, I wanted to explore the interaction especially act of hugging between two people, since hug could potentially not only cheer up a person but also cure their sadness in their hearts. I would used motor and pressure sensor to make this project happen. I think I would mostly focus on making the motor open up the back of the clothing so that the inside could be revealed and work one making the flowers look as if they are alive and they are blooming more as the person was hugged more firmly and tightly.
My biggest step in this would be figuring out how to make the motor and the fabric work together to form something that would make it look naturally opening and look as if it was really something alive rather than something human made.
My idea came from the waves and how it would reflect the sky on to its surface. I wanted to create a simulated nature where there would be small reflective piece on each motor and as a group it would interact with human or show the simulated nature. I originally came up with the idea for the reflective piece as aluminium piece, glass, and acrylic but Golan also gave me an idea about using the dichroic glass and lighting.
My biggest step in this project would be figuring out how to show this with the lighting and the projection that might be projected on to the pieces. How it would be shattered and reflected both the lights and the projection and what effect it would cause the viewers is something to also think about.
For this final project, I intend on constructing a 2.6-foot humanoid robot named Halley that exists for the purpose of emulating a single student in a classroom. The robot would be remote controlled from another computer and be able to perform tasks such as raising a hand and asking questions. Halley would also be equipped with an Android phone for a face, allowing for a wide range of emotions to be displayed.
Here is a concept sketch:
And here is a concept rendering:
This would probably not be possible to build in the course of two weeks, however, fortunately for me, I have been planning for it since the beginning of last year and assembling it since the beginning of this semester, thanks to the support of an FRFAF Microgrant (am about 80% done with raw assembly), meaning I can get it done just in time for the purposes of a final project.
While I do not expect that I will be able to complete the full functionality as a robot student until after Winter Break, I do think I will be able to get some interesting behaviors working. Primarily, I expect to have actions mapped to certain keys such that when pressed, the robot will proceed to play that action. Some actions I will probably set are shaking hands, raising hands, looking up/down/left/right, standing up and sitting down.
Something funny to take out of this, the way I see it, is that students are so detached while in some classrooms and lecture halls that they may as well be robots.