Researchers claim that leading image editing AIs can be jailbroken through rasterized text and visual cues, allowing prohibited edits to bypass safety filters and succeed in up to 80.9% of cases.
Some results have been hidden because they may be inaccessible to you
Show inaccessible results