Live VJ show for real-time projection mapping, where visuals don’t come “pre-rendered” from the past: they’re generated on the fly with AI and projected mapped onto a surface/stage to turn it into a dynamic canvas.
Tools:
TouchDesigner as the real-time engine: compositing, control, sync, warping/keystone, and mapping management.
StreamDiffusion for live AI generation: creating visuals and continuous variations during the show.
TouchDesigner as the real-time engine: compositing, control, sync, warping/keystone, and mapping management.
StreamDiffusion for live AI generation: creating visuals and continuous variations during the show.
What it does: reactive, evolving visuals, aligned to the mapping geometry, with instant changes driven by music/rhythm/inputs.
Outcome: an immersive, near “plug & play” experience (well… almost—reality always demands calibration), ready for performance with stable projection, a strong look, and real-time visual storytelling.
Workflow with StreamDiffusion in TouchDesigner, showing the nodes used for this work in progress