Found 1 discussion
A Hacker News discussion explores whether LLMs and CV models could execute commands hidden in images via steganography, touching on prompt injection, model hallucinations, and AI security.
A Hacker News discussion explores whether LLMs and CV models could execute commands hidden in images via steganography, touching on prompt injection, model hallucinations, and AI security.