We outlined a series of possible items or technologies to use (inputs), and possible results to be achieved through them (outputs).
Inputs:
SPATIAL:
G.P.S.
Radio Fequency ID tags
Pressure sensors
Rotation sensors
"gravity box"
AUDIO: Sound frequency, sound volume, word/voice recognition
BODY: Pulse, Skin conductivity
VR: Augmented Reality, use VR headset and apparatus
VISUAL:
Motion detection.
Motion capture suit
Light or infrared detection
Gesture, shape, colour recognition. Image and pattern recognition
DEVICES:
Keyboard, mouse.
Scanner
MMS/SMS sent to audience member's phone. Calling to a phone.
Smoke detection
Videogame controller, joystick.
Use functions on a website.
Barcodes
Outputs:
Choose or change the plot or in some way to select story fragments
Dynamically control multiple audio channels
Explore a 2D or 3D territory i.e. spatial
Receive an SMS or a call onto a phone.
Compete or take action in a game.
Generate or process visuals using realtime algorithms or effects.
Upload to a blog/vlog/chat
Printout/photo given to audience
Control a 2D or 3D avatar.
Give information eg. a clue or a fact
Physical result (there could be many, such as squashing a flea with a mechanical hammer, moving a giant ball, lift somebody up and down, cause a performer to react.)
Some examples of how this inputs/outputs can be creatively linked:
- Twisting a real marionette or puppet (rotation sensors) --> controls an avatar in 3D world
- Pumping an air pump (pressure sensor) --> makes virtual character go up and down
- Playing a wind instrument (sound sensor) --> the resulting sound is transformed.
No comments:
Post a Comment