
You stand in front of a camera.
You raise your arm.
The sound slowly brightens.
When you become still, the field stabilizes.
When you move differently, the atmosphere reorganizes.
FanRows turns posture and motion into a continuously regulated sound environment — directly in the browser.
Most interactive audio systems are event-based:
a gesture triggers a sound.
FanRows works differently.
Instead of triggering sounds,
the way you move
— how fast, how stable, how sustained —
gradually reshapes multiple layers of sound in real time.
Sound is not executed.
It behaves as a dynamic field.
FanRows operates as a continuous feedback system.
Body influences sound.
Sound influences perception.
Perception influences movement.
The interaction is cyclical and state-based rather than command-driven.
This enables:
For artists:
FanRows offers a continuous sound environment shaped by movement rather than discrete control.
For researchers:
The system provides a structured experimental environment for investigating embodied regulation, state transitions, and dynamic coupling in real time.
FanRows is an independent research-driven prototype exploring:
FanRows can be understood as an applied second-order cybernetic system investigating embodied state dynamics through continuous auditory feedback. Initial experimental data analysis focuses on identifying coherent state windows and dynamic stability patterns within the feedback loop.
The technical structure and system architecture are documented here:
FanRows — Continuous Embodied Audio Interaction
Next page About FanRows