For those of you who couldn't make it, I wanted to share a few key takeaways. This post aims to bring the spirit of [un]prompted directly to your desk, with my own unique observations.
![[un]prompted: Key Insights from the AI Security Practitioners Conference](https://cdn.prod.website-files.com/62b077774e7780ab60b8ff6d/69b9e4b6d0e4ffe083aa86b0_unprompted.png)
The biggest shift evident at the [un]prompted AI Security Practitioners Conference was the move from purely theoretical discussions about "what could go wrong" to concrete, battle-tested methodologies for "what is going wrong and how we fix it." It’s clear that AI security is rapidly evolving, from initial employee DLP use cases, to organization-wide focus around securing all THINGS A.I..
We published an episode of our This Week in AI Security Podcast right after the event, which you can watch here below.
In the episode, I shared some of my key thoughts around several major themes:
Overall, huge kudos to the team over at Knostic!
There were a number of other topics that I didn’t have enough time to cover in the 15-minute episode. Here are some of my thoughts below.
One theme was the urgent need for threat modeling tailored specifically to Large Language Models (LLMs) and generative AI systems. Traditional application security models often fall short, failing to account for the unique attack surface introduced by model weights, training data pipelines, and prompts themselves.
Key speaker sessions highlighted a new approach focusing on three main challenges:
While Prompt Injection remains a foundational concern, the conversation has matured to address more subtle and potentially damaging attack vectors.
Several talks detailed advanced adversarial examples designed not just to trick the model into an undesirable output, but to subtly shift its behavior over time or bypass safety filters without obvious jailbreaking language. This requires a defensive posture that looks beyond simple keyword blocking and into semantic understanding and anomaly detection on input and output data.
The focus is increasingly on how malicious actors can misuse the powerful capabilities of an AI system, even when it's technically operating "as intended." For example, using a coding assistant LLM to generate highly optimized malware code or leveraging an RAG system to exfiltrate proprietary data through cleverly crafted queries. This necessitates integrating "red teaming" early in the development lifecycle, simulating real-world abuse scenarios before deployment.
The conference provided a fantastic overview of the tools that are actually making a difference in AI security labs today. The consensus is that no single tool provides a complete solution, so a layered defense strategy is essential.
The core message is the need for an approach that includes:
Beyond the technical deep-dives, the most engaging discussions centered around the future. Below are some of my thoughts from conversations with AI security leaders that I had at the event: