franciscrot
franciscrot t1_irx4mwf wrote
Reply to comment by ReasonablyBadass in [D] Reversing Image-to-text models to get the prompt by MohamedRashad
You'd think, but I'm pretty sure no. Different models. Also different types of models, I think? Isn't most image captioning GAN?
One thing that's interesting about this q is that the diffusion models, as I understand them (not too well) do already involve a kind of "reversal" in their training - adding more and more noise to an image till it vanishes, then trying to create an image from "pure" noise.
Just in a really non mathy way, I wonder how OP imagines this accommodating rerolling? Would it provide an image seed?
Related: Can the model produce the exact same image from two slightly different prompts?
franciscrot t1_j5g8uq1 wrote
Reply to comment by adt in [D] Couldn't devs of major GPTs have added an invisible but detectable watermark in the models? by scarynut
Would anyone like to explain to me like I'm five how it can be robust against edits like that?