ToneForge Haptics is the tactile counterpart to everything you’ve already built. It completes the sensory loop by turning sound and visual behavior into physical feedback, without breaking determinism or production discipline.
Below is a complete, standalone PRD for ToneForge Haptics, written to integrate cleanly with Sequencer, State, Context, Mixer, Visualizer, Runtime, and Intelligence.
ToneForge Haptics
ToneForge Haptics is the procedural tactile feedback generation module of the ToneForge ecosystem. It converts sound and visual behavior into deterministic, expressive haptic patterns for controllers, mobile devices, wearables, and other feedback‑capable hardware.
ToneForge Haptics does not generate sound or visuals.
It defines how the system feels physically.
ToneForge Haptics operates alongside Visualizer and Mixer as a sensory output layer:
State / Context
↓
Sequencer
↓
Runtime
↓
Mixer / Visualizer / Haptics
Its purpose is to:
- reinforce sound and visual intent
- provide tactile clarity under load
- enhance immersion without overload
- unify multi‑sensory feedback
- support accessibility and comfort
- Deterministic haptic behavior
- Audio‑ and behavior‑driven patterns
- Cross‑device abstraction
- Low‑latency runtime execution
- Explicit, inspectable configuration
- Device‑specific driver management
- Manual waveform editing
- Cinematic vibration authoring
- AI‑driven improvisation
A haptic event represents a tactile response to a sound or behavior.
Examples:
- footstep pulse
- impact thump
- UI confirmation tick
- sustained rumble
Haptic events are behavior‑aligned, not raw waveforms.
A haptic pattern defines:
- intensity curve
- duration
- rhythm
- decay
{
"pattern": "impact_heavy",
"intensity": 0.8,
"duration": 120,
"decay": "fast"
}Patterns are deterministic and reusable.
ToneForge Haptics targets capability classes, not specific hardware:
- simple vibration
- dual‑motor controllers
- linear actuators
- wearable feedback
Device‑specific translation happens at the integration layer.
Sequencer drives:
- haptic timing
- rhythmic repetition
- pattern switching
Example:
- footsteps → rhythmic pulses
- sprinting → faster, sharper feedback
State influences:
- haptic intensity
- pattern selection
- feedback suppression
Example:
- stealth state → reduced haptics
- combat state → stronger feedback
Context influences:
- surface feel (soft vs hard)
- environmental dampening
- accessibility overrides
Example:
- water context → softer, longer pulses
Mixer coordinates:
- haptic intensity with audio loudness
- overload prevention
- priority‑based suppression
This prevents sensory fatigue.
Visualizer and Haptics remain perceptually aligned:
- visual bursts ↔ tactile pulses
- sustained visuals ↔ low‑frequency rumble
Intelligence may:
- detect overuse of haptics
- suggest intensity reductions
- recommend pattern consolidation
Haptics never auto‑adjust without approval.
{
"name": "footstep_gravel",
"pattern": "pulse",
"intensity": 0.4,
"duration": 60
}haptics.trigger("impact_heavy", {
intensity: 0.9
});Haptic triggers are:
- non‑blocking
- frame‑safe
- deterministic
ToneForge Haptics supports:
- global intensity scaling
- pattern simplification
- complete disablement
- accessibility‑aware presets
Comfort is a first‑class concern.
ToneForge Haptics guarantees:
- no hidden randomness
- bounded intensity
- explicit timing
- reproducible behavior
Haptic output is always inspectable.
- Minimal CPU overhead
- Low‑latency triggering
- Scales to many concurrent events
- Safe for mobile and low‑power devices
- Footstep feedback
- Weapon recoil
- UI confirmation
- Environmental effects
- Accessibility reinforcement
- XR and immersive experiences
Optional tooling includes:
- live haptic visualization
- intensity meters
- pattern previews
- overload detection
This makes tactile behavior observable.
- Advanced actuator support
- Network‑synchronized haptics
- User‑customizable haptic profiles
- Cross‑modal fatigue analysis
- Marketplace‑ready haptic patterns
Without a haptics layer:
- feedback relies solely on sound and visuals
- accessibility suffers
- immersion is incomplete
ToneForge Haptics ensures:
- clarity
- physical presence
- sensory balance
It turns sound behavior into felt experience.
ToneForge Haptics is the tactile expression layer of the ToneForge ecosystem.
It transforms sound and visual behavior into deterministic, expressive physical feedback—completing the multi‑sensory loop while preserving control, comfort, and production safety.
If you want next, the strongest follow‑ups are:
- a Haptics + Accessibility deep dive
- example haptic pattern libraries
- or a comparison to traditional controller vibration systems