
You stand in front of a camera.
You raise your arm.
The sound slowly brightens.
When you become still, the field stabilizes.
When you move differently, the atmosphere reorganizes.
FanRows turns posture and motion into a continuously regulated sound environment — directly in the browser.
Most interactive audio systems are event-based:
a gesture triggers a sound.
FanRows works differently.
Instead of triggering sounds,
the way you move —
how fast, how stable, how sustained —
gradually reshapes multiple layers of sound in real time.
Sound is not executed.
It behaves as a dynamic field.
FanRows operates as a continuous feedback system.
Body influences sound.
Sound influences perception.
Perception influences movement.
The interaction is cyclical, not command-based.
This enables:
For artists:
FanRows offers a new way of shaping sound without discrete control or symbolic interfaces.
For researchers:
The system provides a structured environment to investigate embodied regulation, state transitions, and non-symbolic interaction models in real time.
FanRows is an independent research-driven prototype exploring:
The technical structure and system architecture are documented here:
FanRows — Continuous Embodied Audio Interaction
Next page About FanRows