0
$\begingroup$

Recently I have been figuring out ways to do virtual production with blender, and I've already got live camera tracking set up. I have the virtual world in one scene and a plane in the other. Both cameras in both scenes have linked data, so they have the same location and rotation. To make the parallax correct on my projector, I need to UV project the camera view from the active camera onto the plane in the second scene. From there, I go into the side view and put that onto the screen. Everything works perfectly, but I just can't figure out how to get the live camera view as an image texture. Because of that, I can't move the location of the camera during a shot because I could only figure out how to use the render result as a texture. Does anyone know if it is possible to use a viewport camera feed like this? If so, how can I do it? (The purpose of this question is not just as a render, it is for real-time virtual production)

$\endgroup$
3

0

You must log in to answer this question.

Browse other questions tagged .