Instead of regenerating 20 times for the right angle, we can now move inside the scene

LLMs 1.7K points 236 comments 1 month ago

For the longest time, getting the right camera angle in AI images meant regenerating. Too high? Regenerate. Framing slightly off? Regenerate. Perspective not dramatic enough? Regenerate again. I’ve probably wasted more credits fixing angles than anything else. This time I tried something different instead of rerolling, I entered the generated image as a 3D scene and adjusted the camera from inside. Being able to physically move forward, lower the camera, shift perspective, and reframe without rewriting the prompt felt like a completely different workflow. It turns angle selection from guessing into choosing. The interesting part is that it changes how you think about prompting. You don’t need to over-describe camera positioning anymore if you can explore the space afterward. I used ChatGPT to define the base scene and then explored it in 3D inside Cinema Studio 2.0. Has anyone else here tried navigating inside generated scenes instead of regenerating? Curious if this changes how you approach composition.

More from r/ChatGPT