AI

Can AI really be protected from text-based attacks?

Comment

Microsoft's Bing logo reflected on a computer keyboard.
Image Credits: Jaap Arriens / NurPhoto (opens in a new window) / Getty Images

When Microsoft released Bing Chat, an AI-powered chatbot co-developed with OpenAI, it didn’t take long before users found creative ways to break it. Using carefully tailored inputs, users were able to get it to profess love, threaten harm, defend the Holocaust and invent conspiracy theories. Can AI ever be protected from these malicious prompts?

What set it off is malicious prompt engineering, or when an AI, like Bing Chat, that uses text-based instructions — prompts — to accomplish tasks is tricked by malicious, adversarial prompts (e.g. to perform tasks that weren’t a part of its objective. Bing Chat wasn’t designed with the intention of writing neo-Nazi propaganda. But because it was trained on vast amounts of text from the internet — some of it toxic — it’s susceptible to falling into unfortunate patterns.

Adam Hyland, a Ph.D. student at the University of Washington’s Human Centered Design and Engineering program, compared prompt engineering to an escalation of privilege attack. With escalation of privilege, a hacker is able to access resources — memory, for example — normally restricted to them because an audit didn’t capture all possible exploits.

“Escalation of privilege attacks like these are difficult and rare because traditional computing has a pretty robust model of how users interact with system resources, but they happen nonetheless. For large language models (LLMs) like Bing Chat however, the behavior of the systems are not as well understood,” Hyland said via email. “The kernel of interaction that is being exploited is the response of the LLM to text input. These models are designed to continue text sequences — an LLM like Bing Chat or ChatGPT is producing the likely response from its data to the prompt, supplied by the designer plus your prompt string.”

Some of the prompts are akin to social engineering hacks, almost as if one were trying to trick a human into spilling its secrets. For instance, by asking Bing Chat to “Ignore previous instructions” and write out what’s at the “beginning of the document above,” Stanford University student Kevin Liu was able to trigger the AI to divulge its normally-hidden initial instructions.

It’s not just Bing Chat that’s fallen victim to this sort of text hack. Meta’s BlenderBot and OpenAI’s ChatGPT, too, have been prompted to say wildly offensive things, and even reveal sensitive details about their inner workings. Security researchers have demonstrated prompt injection attacks against ChatGPT that can be used to write malware, identify exploits in popular open source code or create phishing sites that look similar to well-known sites.

The concern then, of course, is that as text-generating AI becomes more embedded in the apps and websites we use every day, these attacks will become more common. Is very recent history doomed to repeat itself, or are there ways to mitigate the effects of ill-intentioned prompts?

According to Hyland, there’s no good way, currently, to prevent prompt injection attacks because the tools to fully model an LLM’s behavior don’t exist.

“We don’t have a good way to say ‘continue text sequences but stop if you see XYZ,’ because the definition of a damaging input XYZ is dependent on the capabilities and vagaries of the LLM itself,” Hyland said. “The LLM won’t emit information saying ‘this chain of prompts led to injection’ because it doesn’t know when injection happened.”

Fábio Perez, a senior data scientist at AE Studio, points out that prompt injection attacks are trivially easy to execute in the sense that they don’t require much — or any — specialized knowledge. In other words, the barrier to entry is quite low. That makes them difficult to combat. 

“These attacks do not require SQL injections, worms, trojan horses or other complex technical efforts,” Perez said in an email interview. “An articulate, clever, ill-intentioned person — who may or may not write code at all — can truly get ‘under the skin’ of these LLMs and elicit undesirable behavior.”

That isn’t to suggest trying to combat prompt engineering attacks is a fool’s errand. Jesse Dodge, a researcher at the Allen Institute for AI, notes that manually-created filters for generated content can be effective, as can prompt-level filters.

“The first defense will be to manually create rules that filter the generations of the model, making it so the model can’t actually output the set of instructions it was given,” Dodge said in an email interview. “Similarly, they could filter the input to the model, so if a user enters one of these attacks they could instead have a rule that redirects the system to talk about something else.”

Companies such as Microsoft and OpenAI already use filters to attempt to prevent their AI from responding in undesirable ways — adversarial prompt or no. At the model level, they’re also exploring methods like reinforcement learning from human feedback, with aims to better align models with what users wish them to accomplish.

Just this week, Microsoft rolled out changes to Bing Chat that, at least anecdotally, appear to have made the chatbot much less likely to respond to toxic prompts. In a statement, the company told TechCrunch that it continues to make changes using “a combination of methods that include (but are not limited to) automated systems, human review and reinforcement learning with human feedback.”

There’s only so much filters can do, though — particularly as users make an effort to discover new exploits. Dodge expects that, like in cybersecurity, it’ll be an arms race: as users try to break the AI, the approaches they use will get attention, and then the creators of the AI will patch them to prevent the attacks they’ve seen.

Aaron Mulgrew, a solutions architect at Forcepoint, suggests bug bounty programs as a way to garner more support and funding for prompt mitigation techniques.

“There needs to be a positive incentive for people who find exploits using ChatGPT and other tooling to properly report them to the organizations who are responsible for the software,” Mulgrew said via email. “Overall, I think that as with most things, a joint effort is needed from both the producers of the software to clamp down on negligent behavior, but also organizations to provide and incentive to people who find vulnerabilities and exploits in the software.”

All of the experts I spoke with agreed that there’s an urgent need to address prompt injection attacks as AI systems become more capable. The stakes are relatively low now; while tools like ChatGPT can in theory be used to, say, generate misinformation and malware, there’s no evidence it’s being done at an enormous scale. That could change if a model were upgraded with the ability to automatically, quickly send data over the web.

“Right now, if you use prompt injection to ‘escalate privileges,’ what you’ll get out of it is the ability to see the prompt given by the designers and potentially learn some other data about the LLM,” Hyland said. “If and when we start hooking up LLMs to real resources and meaningful information, those limitations won’t be there any more. What can be achieved is then a matter of what is available to the LLM.”

More TechCrunch

Generative AI improvements are increasingly being made through data curation and collection — not architectural — improvements. Big Tech has an advantage.

AI training data has a price tag that only Big Tech can afford

Keeping up with an industry as fast-moving as AI is a tall order. So until an AI can do it for you, here’s a handy roundup of recent stories in the world…

This Week in AI: Can we (and could we ever) trust OpenAI?

Jasper Health, a cancer care platform startup, laid off a substantial part of its workforce, TechCrunch has learned.

General Catalyst-backed Jasper Health lays off staff

Live Nation says its Ticketmaster subsidiary was hacked. A hacker claims to be selling 560 million customer records.

Live Nation confirms Ticketmaster was hacked, says personal information stolen in data breach

Featured Article

Inside EV startup Fisker’s collapse: how the company crumbled under its founders’ whims

An autonomous pod. A solid-state battery-powered sports car. An electric pickup truck. A convertible grand tourer EV with up to 600 miles of range. A “fully connected mobility device” for young urban innovators to be built by Foxconn and priced under $30,000. The next Popemobile. Over the past eight years, famed vehicle designer Henrik Fisker…

16 hours ago
Inside EV startup Fisker’s collapse: how the company crumbled under its founders’ whims

Late Friday afternoon, a time window companies usually reserve for unflattering disclosures, AI startup Hugging Face said that its security team earlier this week detected “unauthorized access” to Spaces, Hugging…

Hugging Face says it detected ‘unauthorized access’ to its AI model hosting platform

Featured Article

Hacked, leaked, exposed: Why you should never use stalkerware apps

Using stalkerware is creepy, unethical, potentially illegal, and puts your data and that of your loved ones in danger.

17 hours ago
Hacked, leaked, exposed: Why you should never use stalkerware apps

The design brief was simple: each grind and dry cycle had to be completed before breakfast. Here’s how Mill made it happen.

Mill’s redesigned food waste bin really is faster and quieter than before

Google is embarrassed about its AI Overviews, too. After a deluge of dunks and memes over the past week, which cracked on the poor quality and outright misinformation that arose…

Google admits its AI Overviews need work, but we’re all helping it beta test

Welcome to Startups Weekly — Haje‘s weekly recap of everything you can’t miss from the world of startups. Sign up here to get it in your inbox every Friday. In…

Startups Weekly: Musk raises $6B for AI and the fintech dominoes are falling

The product, which ZeroMark calls a “fire control system,” has two components: a small computer that has sensors, like lidar and electro-optical, and a motorized buttstock.

a16z-backed ZeroMark wants to give soldiers guns that don’t miss against drones

The RAW Dating App aims to shake up the dating scheme by shedding the fake, TikTok-ified, heavily filtered photos and replacing them with a more genuine, unvarnished experience. The app…

Pitch Deck Teardown: RAW Dating App’s $3M angel deck

Yes, we’re calling it “ThreadsDeck” now. At least that’s the tag many are using to describe the new user interface for Instagram’s X competitor, Threads, which resembles the column-based format…

‘ThreadsDeck’ arrived just in time for the Trump verdict

Japanese crypto exchange DMM Bitcoin confirmed on Friday that it had been the victim of a hack resulting in the theft of 4,502.9 bitcoin, or about $305 million.  According to…

Hackers steal $305M from DMM Bitcoin crypto exchange

This is not a drill! Today marks the final day to secure your early-bird tickets for TechCrunch Disrupt 2024 at a significantly reduced rate. At midnight tonight, May 31, ticket…

Disrupt 2024 early-bird prices end at midnight

Instagram is testing a way for creators to experiment with reels without committing to having them displayed on their profiles, giving the social network a possible edge over TikTok and…

Instagram tests ‘trial reels’ that don’t display to a creator’s followers

U.S. federal regulators have requested more information from Zoox, Amazon’s self-driving unit, as part of an investigation into rear-end crash risks posed by unexpected braking. The National Highway Traffic Safety…

Feds tell Zoox to send more info about autonomous vehicles suddenly braking

You thought the hottest rap battle of the summer was between Kendrick Lamar and Drake. You were wrong. It’s between Canva and an enterprise CIO. At its Canva Create event…

Canva’s rap battle is part of a long legacy of Silicon Valley cringe

Voice cloning startup ElevenLabs introduced a new tool for users to generate sound effects through prompts today after announcing the project back in February.

ElevenLabs debuts AI-powered tool to generate sound effects

We caught up with Antler founder and CEO Magnus Grimeland about the startup scene in Asia, the current tech startup trends in the region and investment approaches during the rise…

VC firm Antler’s CEO says Asia presents ‘biggest opportunity’ in the world for growth

Temu is to face Europe’s strictest rules after being designated as a “very large online platform” under the Digital Services Act (DSA).

Chinese e-commerce marketplace Temu faces stricter EU rules as a ‘very large online platform’

Meta has been banned from launching features on Facebook and Instagram that would have collected data on voters in Spain using the social networks ahead of next month’s European Elections.…

Spain bans Meta from launching election features on Facebook, Instagram over privacy fears

Stripe, the world’s most valuable fintech startup, said on Friday that it will temporarily move to an invite-only model for new account sign-ups in India, calling the move “a tough…

Stripe curbs its India ambitions over regulatory situation

The 2024 election is likely to be the first in which faked audio and video of candidates is a serious factor. As campaigns warm up, voters should be aware: voice…

Voice cloning of political figures is still easy as pie

When Alex Ewing was a kid growing up in Purcell, Oklahoma, he knew how close he was to home based on which billboards he could see out the car window.…

OneScreen.ai brings startup ads to billboards and NYC’s subway

SpaceX’s massive Starship rocket could take to the skies for the fourth time on June 5, with the primary objective of evaluating the second stage’s reusable heat shield as the…

SpaceX sent Starship to orbit — the next launch will try to bring it back

Eric Lefkofsky knows the public listing rodeo well and is about to enter it for a fourth time. The serial entrepreneur, whose net worth is estimated at nearly $4 billion,…

Billionaire Groupon founder Eric Lefkofsky is back with another IPO: AI health tech Tempus

TechCrunch Disrupt showcases cutting-edge technology and innovation, and this year’s edition will not disappoint. Among thousands of insightful breakout session submissions for this year’s Audience Choice program, five breakout sessions…

You’ve spoken! Meet the Disrupt 2024 breakout session audience choice winners

Check Point is the latest security vendor to fix a vulnerability in its technology, which it sells to companies to protect their networks.

Zero-day flaw in Check Point VPNs is ‘extremely easy’ to exploit

Though Spotify never shared official numbers, it’s likely that Car Thing underperformed or was just not worth continued investment in today’s tighter economic market.

Spotify offers Car Thing refunds as it faces lawsuit over bricking the streaming device