Before we jump in, let me introduce myself. My name is Joseph Thacker, aka rez0. I’m a full-time bug hunter and former principal AI engineer. Over the years, I’ve helped hundreds of companies unearth vulnerabilities that could have cost them millions. I’ve submitted over 1,000 vulnerabilities through crowdsourced platforms like Bugcrowd. Furthermore, I love building and breaking systems.
Alright, let’s talk AI safety and security. One of the first executive actions of the Trump Administration was to repeal the previous administration’s guidance on AI safety and security testing. What does this mean? Specifically, what does it mean for you and other members of the public?
The previous guidance included:
With the Trump Administration in office, these regulations have been rescinded. The new administration seems to be taking a different approach, which, at first glance, might seem like a step backward. However, whether we see progress depends on what comes next. The Trump Administration has announced Project Stargate (awesome name, by the way), which is an investment of up to $500 billion in private-sector AI infrastructure. This is meant to foster more AI innovation. That said, a case can be made that large AI companies are more focused on “winning the market” than making sure that what they create is well-secured and safe.
From a hacker and cybersecurity lens, this shift raises some concerns. While it’s true that AI safety and security risks can easily be overblown, ignoring them isn’t the answer either. The Biden Administration’s guidance on AI safety had its flaws, but I thought it was a decent start. Without these guidelines, the question is: What risks does the new administration see as priorities?
Personally, AI safety risks, such as a superintelligence enslaving us all, doesn’t concern me in the short term as much as AI security does. The potential for AI to be leveraged in attacks, like indirect prompt injection, through vulnerabilities is a real issue that needs attention. Developers need practical advice and recommendations for addressing these security challenges. In the long term, AI safety is worth prioritizing because AI systems will one day be even more influential than they are today.
Below are the specific vulnerability types in new AI applications that I have been exploiting in the wild. Prompt injections
AI = New paths to traditional vulnerabilities
While the rollback of these regulations may seem alarming, I think we can wait and see what comes next. The major point I want you to leave with is that AI will have a significant impact on our lives, and we need people working on both AI safety and security. The new administration should ideally release something similar to Biden’s order, or at least provide more practical guidance on security issues. What changes or recommendations would I like to see from the new administration? I’m glad you asked.
As we wrap things up, I just want to express that navigating the ever-changing landscape of AI safety and security has never been more important. With regulatory changes and new vulnerabilities emerging, it’s obvious to me that AI is a double-edged sword—we are going to have to wield it wisely. While the rollback of Biden’s regulations might offend some, it also opens the door for (potentially) more practical recommendations and protection.
My hope is that people will work on robust safety guardrails, create smarter design patterns, and devote both time and money to educating the captain steering the ship. AI is moving fast, and our job is to ensure security efforts can keep pace. Please stay engaged and proactive. Don’t lose vigilance.