Finally! Android Open Accessory Development Kit.
Kinesthetic Interfacing with Architecture
Choreography within architecture can be understood in two ways: ergonomically and performatively. Ergonomic choreography is the technically determined kinetics of the human body within space. It aims to control movement in predictable and repeatable ways within the built environment. Performance choreography focuses on the kinesthetic experience (the feelings associated with movement) of the human body in space. In contrast it aims to explore how the moving body “interacts” dynamically with architecture. Throughout the twentieth century, kinesthetics have undergone a conceptual evolution. Performance art in the 1960’s challenged our understanding of how the body moves around space and advancements in media technology have addressed new possibilities for communicating and controlling the environment through gesture recognition. As computing becomes less an immersive experience, associated with gaming and virtual reality environments in which reality is simulated or re-created, and more a material layer implemented into the built environment, interfacing between the two contexts takes on spatial constructs. The resulting architectures are composites, mixed realities of physical/digital space. This thesis explores the possibilities of performance for developing kinesthetic interfaces with architecture.
The project is constructed as an interactive field condition; a landscape constructed from a sequence of unforeseen tactile experiences, acting as a mediator between the physical and digital environment that define its programmatic performance. It seeks to situate the participant in an environment which reveals its and identity and function through performing; a communication resulting from the participants input: a series of hits, bumps, knocks, kicks or other forms of contact. Cataloged within the system as a series of locations and interactions, the field locates you architecturally within the stoa. The project also seeks to investigate the moving body’s ability to spatially orient itself in a terrain which must be activated in order to be perceived. As one progressively engages with the space they begin to create a more full embodied dialogue with the system, understanding the physical inputs, locations, and communications across the field.
This video shows the updated interaction with the GPS location coordinates being entered into the collective database of all users. The map shown on the contribute tab can be found at http://jodipfister.com/testsmap.php.
The remaining documentation can be found at http://jodipfister.wordpress.com/2011/05/07/smellscape-updated/.
I have made a few blog posts on my website which documents SoundClusters both as a process and a final product.
SoundClusters description > http://soundclusters.adamlaskowitz.com/about.php
Video documentation > http://soundclusters.adamlaskowitz.com/blog.php?id=26
Write-up of the process> http://soundclusters.adamlaskowitz.com/blog.php?id=28
Source Code > http://soundclusters.adamlaskowitz.com/blog.php?id=27
Download the app > http://soundclusters.adamlaskowitz.com/blog.php?id=25
Video Documentation of the project is currently a little rough…. more revisions coming soon!
The above is a link to the Smellscape description and demonstration.
Quick way to bypass screen orientation force close error:
For the most part, this is all you need to do:
switch to Application tab and
select desired Activity you wish to override orientation change behavior.