If you aren’t familiar, Syphon is a technology for applications to share video frames through the GPU for high-performance and low-latency. Within Unity, I setup Syphon textures using the KlaxSyphon Unity plugin. The four camera limit was based on hardware limitations - I was running all this on a Mac Mini with an external GPU - between simultaneously using QLab, Unity, multiple Syphon feeds, NDI and NDI to Blackmagic conversion, the hardware was maxed out at four Unity cameras. I then imported and reassembled the mobile inside Unity and setup four cameras - three for three camera angles, and one for an overhead view to send to lighting desk (more on this later) I modeled UV mapped the mobile in Blender within a model of the theatre to get the scale right. Using NDI Outlet, I was able to get key and fill video streaming from Unity to my UltraStudio Mini HD without too much hassle. Note, this app is Mac only and only supports Blackmagic UltraStudio devices. While this only gets 1.5 stars on the Mac App Store, it worked flawlessly for me. I found NDI Outlet from Sienna - which was the only software cost in the project (€150). The next challenge was to get this NDI + alpha feed to physical key / fill outputs over SDI. They also make a great Syphon plugin for Mac that I also use in this project. The solution here was to use the NDI plugin for Unity by Github user Keijiro - KlakNDI. I got this working first so I knew I could proceed with the greater idea. This was actually the trickiest part - finding a way to get key / fill feeds out of Unity. That meant sending them key / fill video feeds - a broadcast technique for live compositing whereby one feed is the content and the other is a greyscale feed generated from the transparency in the source, that enables a production vision mixer to overlay graphics on live camera source. The streaming was being handled by a professional streaming production team using digital broadcast gear, therefore I would need to fit in with their workflow and technology. It would slowly turn as mobiles do.Ĭatch 2: The live stream would be multi-camera so my visuals would need to look correct for different camera angles, and cut live as the camera cut. The elements of the mobile would be a mix of what looked like LED video panels and light orbs. Idea: my idea was to make a 3D Calder-esque mobile that would appear to be hanging above the players on the stage. The catch: the show was to be live-streamed only, no live audience so no nice on-stage visuals.Ĭonstraint: I’m generally not a fan of inter-cutting film / video art with live-performance, I like to see the musicians all the time, so I wanted to try to add some visual element that looked like it was in the space with the musicians. The project: Crash Ensemble were premiering a new piece - Wingform - by composer and musician Barry O'Halpin as part of New Music Dublin festival, and asked if I’d do some live visuals for it. The following is a brief write-up on the creative-tech side of a recent project I produced with Crash Ensemble, Ireland’s leading new music ensemble. This project benefited from research and development I conducted as part of a bursary from The Arts Council of Ireland. JOverlay visuals for live-streamed events using Unity Engine
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |