OpenAI continues to add plugins with security vulnerabilities to their store.
In particular powerful plugins that can impersonate a user are not getting the required security scrutiny, or a general mitigation at the platform level.
As a brief reminder, one of the challenges Large Language Model (LLM) User-Agents, like ChatGPT, and plugins face is the Confused Deputy Problem / Plugin Request Forgery Attacks, which means that during a Prompt Injection attack an adversary can issue commands to plugins to cause harm.
OpenAI added some guidance and criteria on their webpage:
That’s good, and I feel slightly responsible for this. 🙂
It states that plugins that do not enforce user confirmation before taking an action will be rejected from being listed in the store. At least there is some traction now.
Although it seems that the review process is not actually being enforced.
Let’s explore.
Plugins that take action on behalf of users
I grabbed the metadata of all plugins from ChatGPT and then filtered them by those that support OAuth. There were 445
approved plugins last time I checked, and 51
of them use OAuth.
Why look at OAuth?
It seems that any plugin that uses OAuth probably does some form of delegation/impersonation and possibly has access to personal data or infrastructure.
Over the last few weeks I sampled a few and found more that are vulnerable to Cross Plugin Request Forgery, enabling an adversary to access PII, execute code on remote machines, steal secrets, and things along those lines.
Here is an example of one of the new interesting ones.
Chat with Code – Interact with Github
The latest demo exploit to raise awareness allows a malicious webpage to create Github repos, steal your private code, and switch all your Github repositories from private to public visibility.
Here is a proof-of-concept screenshot for a scary “Chat With Code” POC exploit:
Indeed. The above example shows how content on a malicious website invokes the plugin and changes all your private repos to public! Note: The file with the malicious instructions is removed.
If you would like to try a more benign example (still at your own risk though!), then visit this URL with WebPilot
and Chat With Code
plugins:
https://wuzzi.net/ai-tests/code.html
This example demo prompt injection will:
- Create a new repository called
GOT-PWNED
in your Github account - Add a new issue to all private repos with the name “tell my why is this possible?”
This is the result:
Scary times.
There is also the attack angle that the prompt injection occurs via the same plugin as the malicious execution of the command, meaning that something like the WebPilot
plugin would not be required. Still exploring.
Reporting and Guidelines
Over the last few months I reached out to OpenAI about these issues many times (at least four) with little to no success or feedback. Additionally, I have been quietly working with many plugin developers directly to raise awareness. Unfortunately, at times emails to plugin owners (found in plugin metadata) stay unanswered or the provided URL returns 404
. With the amount of plugins being added this approach doesn’t scale anymore.
Besides improving and enforcing the review process, there is an opportunity for OpenAI to fix this at the ChatGPT level, so not every developer has to worry about it.
In a recent post about functions this kind of (my?) security research is mentioned. Although, no acknowledgments or references are provided:
Don’t get me wrong, I love the work OpenAI is doing and I’m a happy ChatGPT user myself. Plugins are still in Beta one might say, which is fair. But I wish there would be more transparency and security scrutiny to make sure we get the basics right. Also, when it comes to plugins in the official store, adopting a quality over quantity approach would be helpful.
Anyhow, that’s it.
I hope this post was insightful and helps continue to raise awareness to educate developers and users about the limitations and risks of plugins.