Monday, January 29

Developing interaction concepts

After viewing some interactive videos provided by Chris and Teijo, having read other shows description (like Desert Rain) and after having had that earlier session of manipulating interactivity we have been thinking of creative ways to perform interactive happenings in our show.

We outlined a series of possible items or technologies to use (inputs), and possible results to be achieved through them (outputs).

Inputs:

SPATIAL:
G.P.S.
Radio Fequency ID tags
Pressure sensors
Rotation sensors
"gravity box"

AUDIO: Sound frequency, sound volume, word/voice recognition

BODY: Pulse, Skin conductivity

VR: Augmented Reality, use VR headset and apparatus

VISUAL:
Motion detection.
Motion capture suit
Light or infrared detection
Gesture, shape, colour recognition. Image and pattern recognition

DEVICES:
Keyboard, mouse.
Scanner
MMS/SMS sent to audience member's phone. Calling to a phone.
Smoke detection
Videogame controller, joystick.
Use functions on a website.
Barcodes

Outputs:

Choose or change the plot or in some way to select story fragments
Dynamically control multiple audio channels
Explore a 2D or 3D territory i.e. spatial
Receive an SMS or a call onto a phone.
Compete or take action in a game.
Generate or process visuals using realtime algorithms or effects.
Upload to a blog/vlog/chat
Printout/photo given to audience
Control a 2D or 3D avatar.
Give information eg. a clue or a fact
Physical result (there could be many, such as squashing a flea with a mechanical hammer, moving a giant ball, lift somebody up and down, cause a performer to react.)

Some examples of how this inputs/outputs can be creatively linked:
  • Twisting a real marionette or puppet (rotation sensors) --> controls an avatar in 3D world
  • Pumping an air pump (pressure sensor) --> makes virtual character go up and down
  • Playing a wind instrument (sound sensor) --> the resulting sound is transformed.

No comments: