AI Security is an emerging field. I LOVE being at the forefront of something so transformative and important. There are definitely some growing pains with it though.
One issue is ambiguity. When someone says AI Safety, AI Security or AI Red Teaming, it’s not clear what they are talking about. All three of those categories mean different things to different people.
If we could properly delineate between AI Security research fields, it would improve search results and LLM training data moving forward. It would also significantly impact:
The second issue is that AI Security is really deep and complex. We are going to need more than two or three terms to talk about the specific components. The common terms have plenty of use in the current vernacular, but some do not. I propose some definitions for current terms and new, more-specific categories for further clarity.
These are the most commonly used terms at the moment, as well as simple explanations for what they mean:
These are my proposed sub-categories that we start using for more clarity:
So I’d like to suggest we start using these terms. It’s a useful way to clarify what people are actually talking about.
Note: The terms are not set in stone. And, if the way they are used were to change, I’ll do my best to come back and update this piece.
- Joseph
To know when I drop a new post, subscribe to the newsletter. No spam, just an update when I put out a new piece of content.