Need 4k, want 8k when those projectors become affordable.
Got rid of the mirror and use the upper half of a video camera fisheye lens; big aperture, cheaper, better image quality, more compact, and easier to maintain/calibrate.
I plan to dig into Dalai Felinto's code in the blender game engine, to create an image node in the rendering. Just need something for a simple dome quick look to set scenegraph/render parameters. For production, render in fisheye, and move that to vuo for further enhancements (eg, sound, midi & dmx).
I use Blender for animation, particularly for fisheye scenegraphs. For a complete movie I use vuo's warp to do the full video.
I have a 6m dome with a modified version of Paul Bourke's warped fisheye offset projection.
Getting the models correct, lighting correct, and camera optimal is always a problem of rendering. To actually view it I have to warp the fisheye. Dalai Felinto did a warped fisheye in the gaming engine 15y ago, but it did not work in practical animation rendering.
I thought I could tighten the QC loop using vuo to do the warping thru syphon. But the syphon code is barely there, so getting it into vuo to do the warp is problematic.
I think the solution is to write a warp node in blender and call it a day.
Screen capture is an option that Paul Bourke suggested on a private email. The problem here is managing displays and juggling different different sw packages simultaneously.
My current inclination is to do uv mapping within blender, as a node after rendering. Currently trying to figure out how to take the meshmap used in vuo, and convert it into a UV blender node.
Curious if anyone in the vuo community, besides Paul and me, are doing dome viz.