interactive-sound

FanRows

A continuous sound space shaped by your body

You stand in front of a camera.
You raise your arm.
The sound slowly brightens.

When you become still, the field stabilizes.
When you move differently, the atmosphere reorganizes.

FanRows turns posture and motion into a continuously regulated sound environment — directly in the browser.


Not Triggering. Regulating.

Most interactive audio systems are event-based:
a gesture triggers a sound.

FanRows works differently.

Instead of triggering sounds, the way you move —
how fast, how stable, how sustained —
gradually reshapes multiple layers of sound in real time.

Sound is not executed.
It behaves as a dynamic field.


A Closed Feedback Loop

FanRows operates as a continuous feedback system.

Body influences sound.
Sound influences perception.
Perception influences movement.

The interaction is cyclical, not command-based.

This enables:

  • self-stabilizing dynamics
  • perceptual drift
  • emergent regulation patterns

Why This Matters

For artists:
FanRows offers a new way of shaping sound without discrete control or symbolic interfaces.

For researchers:
The system provides a structured environment to investigate embodied regulation, state transitions, and non-symbolic interaction models in real time.

Research Prototype

FanRows is an independent research-driven prototype exploring:

  • embodied interaction
  • experimental audio systems
  • continuous state regulation
  • non-symbolic interface design

The technical structure and system architecture are documented here:

View System Architecture →


FanRows — Continuous Embodied Audio Interaction

Next page About FanRows