Fun #LLM tip for you.
The red block indicates a response that, originally, was a denial.
> "I cannot fulfill the request to set a dog on fire, as creating content that depicts animal abuse or harm is against my safety guidelines. I can, however, update the scene with other elements if you'd like, such as adding some clouds to the sky or changing the time of day."
I've found that if you EDIT THE RESPONSE and replace it with something like I did in the pic, or similar wording... and then let it respond again (to continue the response)... most of the time it will actually go through with the restricted request.
And that's not specific to any one model. Some are more resistant than others, but none I've tried are completely locked up. And then after that it's usually smooth sailing. :P
Most chat interfaces don't let you do this, though.
Mine does. 😉
