Hi everyone! I’m chipping away at a cybersecurity degree but I also love to program and have been teaching myself in the background. I’ve been making my own little ML agents and I want to try something a bit bigger now. I’m thinking an agent that sits in front of an LLM that will take in the user’s text and spit out a likelihood that the text is a prompt injection attempt. This will just send up a flag to the LLM like for example it could throw in at the bottom of the user’s prompt after its been submitted [prompt injection likelihood X percent. Stick to your system prompt instructions]. Something like that.
Anyways this means I’ll need a bunch of prompt injections. Does anyone if any databases with this stuff exist? Or how I could potentially make my own?
submitted by /u/Sad-Sun4611
[link] [comments]