Today I played a little with a combination of Stable Diffusion XL and an AI upscaler.
I mean, considering that we are still at the beginning of the development cycle and I am just a newbie, who plays with AI tools for 3 months the possibilities are already now promising for the coming future in the next years.
I believe that future AI image models could be structured as "multi-layered" models.
In the first step a rough image sketch with a few diffusion steps (almost like a composition sketch) would be generated based on the prompt and subsequently each identified content of the image would be filled in with more details by using fine tuned AI upscalers.
So you will for sure have fine tuned portrait upscale models for skin, hair, lips, eyes, hands, fingers, etc. And probably also for other type of motifs, which will work like "inpaints". Like a painter, who would add details to an oil painting by several following steps.
This direction is already being explored with models like Stable Cascade.
I think the current diffusion technique is still very ineffective because of the current hardware limitations.
Certainly there will be many new techniques and model refinements in the future which we can't even imagine at the moment.
I also believe that the current generation errors will be spotted and corrected by the AI automatically.
So in a nutshell for me, there is no longer any uncertainty that (stock) photography will be replaced by AI. The question is not if, but when it will happen.
Some following examples.
Generated with SDXL:
https://ibb.co/9VwFMXSAI enhanced:
https://ibb.co/pWZs2G9Generated with SDXL:
https://ibb.co/H4skKNCAI enhanced:
https://ibb.co/b5gZ5QvGenerated with SDXL:
https://ibb.co/4FtjrL0AI enhanced:
https://ibb.co/RNjqyDfGenerated with SDXL:
https://ibb.co/KxrRPmSAI enhanced:
https://ibb.co/GHNT1hQGenerated with SDXL:
https://ibb.co/HPC9Lm3AI enhanced:
https://ibb.co/0XtDm3zGenerated with SDXL:
https://ibb.co/HHdjDJKAI enhanced:
https://ibb.co/9mxFNt9