interactive-sound

FanRows

A continuous sound space shaped by your body

You stand in front of a camera.
You raise your arm.
The sound slowly brightens.

When you become still, the field stabilizes.
When you move differently, the atmosphere reorganizes.

FanRows turns posture and motion into a continuously regulated sound environment — directly in the browser.


Not Triggering. Regulating.

Most interactive audio systems are event-based:
a gesture triggers a sound.

FanRows works differently.

Instead of triggering sounds, the way you move
— how fast, how stable, how sustained —
gradually reshapes multiple layers of sound in real time.

Sound is not executed.
It behaves as a dynamic field.


A Closed Feedback Loop

FanRows operates as a continuous feedback system.

Body influences sound.
Sound influences perception.
Perception influences movement.

The interaction is cyclical and state-based rather than command-driven.

This enables:

  • self-stabilizing dynamics
  • perceptual drift
  • emergent regulation patterns
  • temporally bounded coherent states

Why This Matters

For artists:
FanRows offers a continuous sound environment shaped by movement rather than discrete control.

For researchers:
The system provides a structured experimental environment for investigating embodied regulation, state transitions, and dynamic coupling in real time.

Research Prototype

FanRows is an independent research-driven prototype exploring:

  • embodied interaction
  • applied cybernetic feedback systems
  • continuous state regulation
  • non-symbolic interface design

FanRows can be understood as an applied second-order cybernetic system investigating embodied state dynamics through continuous auditory feedback. Initial experimental data analysis focuses on identifying coherent state windows and dynamic stability patterns within the feedback loop.

The technical structure and system architecture are documented here:

View System Architecture →


FanRows — Continuous Embodied Audio Interaction

Next page About FanRows