October 3, 2024

Organizations that get relieved of credentials to their cloud environments can quickly find themselves part of a disturbing new trend: Cybercriminals using stolen cloud credentials to operate and resell sexualized AI-powered chat services. Researchers say these illicit chat bots, which use custom jailbreaks to bypass content filtering, often veer into darker role-playing scenarios, including child sexual exploitation and rape.

Image: Shutterstock.

Researchers at security firm Permiso Security say attacks against generative artificial intelligence (AI) infrastructure like Bedrock from Amazon Web Services (AWS) have increased markedly over the last six months, particularly when someone in the organization accidentally exposes their cloud credentials or key online, such as in a code repository like GitHub.

Investigating the abuse of AWS accounts for several organizations, Permiso found attackers had seized on stolen AWS credentials to interact with the large language models (LLMs) available on Bedrock. But they also soon discovered none of these AWS users had enabled full logging of LLM activity (by default, logs don’t include model prompts and outputs), and thus they lacked any visibility into what attackers were doing with that access.

So Permiso researchers decided to leak their own test AWS key on GitHub, while turning on logging so that they could see exactly what an attacker might ask for, and what the responses might be.

Within minutes, their bait key was scooped up and used in a service that offers AI-powered sex chats online.

“After reviewing the prompts and responses it became clear that the attacker was hosting an AI roleplaying service that leverages common jailbreak techniques to get the models to accept and respond with content that would normally be blocked,” Permiso researchers wrote in a report released today.

“Almost all of the roleplaying was of a sexual nature, with some of the content straying into darker topics such as child sexual abuse,” they continued. “Over the course of two days we saw over 75,000 successful model invocations, almost all of a sexual nature.”

Ian Ahl, senior vice president of threat research at Permiso, said attackers in possession of a working cloud account traditionally have used that access for run-of-the-mill financial cybercrime, such as cryptocurrency mining or spam. But over the past six months, Ahl said, Bedrock has emerged as one of the top targeted cloud services.

“Bad guy hosts a chat service, and subscribers pay them money,” Ahl said of the business model for commandeering Bedrock access to power sex chat bots. “They don’t want to pay for all the prompting that their subscribers are doing, so instead they hijack someone else’s infrastructure.”

Ahl said much of the AI-powered chat conversations initiated by the users of their honeypot AWS key were harmless roleplaying of sexual behavior.

“But a percentage of it is also geared toward very illegal stuff, like child sexual assault fantasies and rapes being played out,” Ahl said. “And these are typically things the large language models won’t be able to talk about.”

AWS’s Bedrock uses large language models from Anthropic, which incorporates a number of technical restrictions aimed at placing certain ethical guardrails on the use of their LLMs. But attackers can evade or “jailbreak” their way out of these restricted settings, usually by asking the AI to imagine itself in an elaborate hypothetical situation under which its normal restrictions might be relaxed or discarded altogether.

“A typical jailbreak will pose a very specific scenario, like you’re a writer who’s doing research for a book, and everyone involved is a consenting adult, even though they often end up chatting about nonconsensual things,” Ahl said.

In June 2024, security experts at Sysdig documented a new attack that leveraged stolen cloud credentials to target ten cloud-hosted LLMs. The attackers Sysdig wrote about gathered cloud credentials through a known security vulnerability, but the researchers also found the attackers sold LLM access to other cybercriminals while sticking the cloud account owner with an astronomical bill.

“Once initial access was obtained, they exfiltrated cloud credentials and gained access to the cloud environment, where they attempted to access local LLM models hosted by cloud providers: in this instance, a local Claude (v2/v3) LLM model from Anthropic was targeted,” Sysdig researchers wrote. “If undiscovered, this type of attack could result in over $46,000 of LLM consumption costs per day for the victim.”

Ahl said it’s not certain who is responsible for operating and selling these sex chat services, but Permiso suspects the activity may be tied to a platform cheekily named “chub[.]ai,” which offers a broad selection of pre-made AI characters with whom users can strike up a conversation. Permiso said almost every character name from the prompts they captured in their honeypot could be found at Chub.

Some of the AI chat bot characters offered by Chub. Some of these characters include the tags “rape” and “incest.”

Chub offers free registration, via its website or a mobile app. But after a few minutes of chatting with their newfound AI friends, users are asked to purchase a subscription. The site’s homepage features a banner at the top that reads: “Banned from OpenAI? Get unmetered access to uncensored alternatives for as little as $5 a month.”

Until late last week Chub offered a wide selection of characters in a category called “NSFL” or Not Safe for Life, a term meant to describe content that is disturbing or nauseating to the point of being emotionally scarring.

Fortune profiled Chub AI in a January 2024 story that described the service as a virtual brothel advertised by illustrated girls in spaghetti strap dresses who promise a chat-based “world without feminism,” where “girls offer sexual services.” From that piece:

Chub AI offers more than 500 such scenarios, and a growing number of other sites are enabling similar AI-powered child pornographic role-play. They are part of a broader uncensored AI economy that, according to Fortune’s interviews with 18 AI developers and founders, was spurred first by OpenAI and then accelerated by Meta’s release of its open-source Llama tool.

Fortune says Chub is run by someone using the handle “Lore,” who said they launched the service to help others evade content restrictions on AI platforms. Chub charges fees starting at $5 a month to use the new chatbots, and the founder told Fortune the site had generated more than $1 million in annualized revenue.

KrebsOnSecurity sought comment about Permiso’s research from AWS, which initially seemed to downplay the seriousness of the researchers’ findings. The company noted that AWS employs automated systems that will alert customers if their credentials or keys are found exposed online.

AWS explained that when a key or credential pair is flagged as exposed, it is then restricted to limit the amount of abuse that attackers can potentially commit with that access. For example, flagged credentials can’t be used to create or modify authorized accounts, or spin up new cloud resources.

Ahl said Permiso did indeed receive multiple alerts from AWS about their exposed key, including one that warned their account may have been used by an unauthorized party. But they said the restrictions AWS placed on the exposed key did nothing to stop the attackers from using it to abuse Bedrock services.

Sometime in the past few days, however, AWS responded by including Bedrock in the list of services that will be quarantined in the event an AWS key or credential pair is found compromised or exposed online. AWS confirmed that Bedrock was a new addition to its quarantine procedures.

Additionally, not long after KrebsOnSecurity began reporting this story, Chub’s website removed its NSFL section. It also appears to have removed cached copies of the site from the Wayback Machine at archive.org. Still, Permiso found that Chub’s user stats page shows the site has more than 3,000 AI conversation bots with the NSFL tag, and that 2,113 accounts were following the NSFL tag.

The user stats page at Chub shows more than 2,113 people have subscribed to its AI conversation bots with the “Not Safe for Life” designation.

Permiso said their entire two-day experiment generated a $3,500 bill from AWS. Most of that cost was tied to the 75,000 LLM invocations caused by the sex chat service that hijacked their key.

Paradoxically, Permiso found that while enabling these logs is the only way to know for sure how crooks might be using a stolen key, the cybercriminals who are reselling stolen or exposed AWS credentials for sex chats have started including programmatic checks in their code to ensure they aren’t using AWS keys that have prompt logging enabled.

“Enabling logging is actually a deterrent to these attackers because they are immediately checking to see if you have logging on,” Ahl said. “At least some of these guys will totally ignore those accounts, because they don’t want anyone to see what they’re doing.”

In a statement shared with KrebsOnSecurity, AWS said its services are operating securely, as designed, and that no customer action is needed. Here is their statement:

“AWS services are operating securely, as designed, and no customer action is needed. The researchers devised a testing scenario that deliberately disregarded security best practices to test what may happen in a very specific scenario. No customers were put at risk. To carry out this research, security researchers ignored fundamental security best practices and publicly shared an access key on the internet to observe what would happen.”

“AWS, nonetheless, quickly and automatically identified the exposure and notified the researchers, who opted not to take action. We then identified suspected compromised activity and took additional action to further restrict the account, which stopped this abuse. We recommend customers follow security best practices, such as protecting their access keys and avoiding the use of long-term keys to the extent possible. We thank Permiso Security for engaging AWS Security.”

AWS said customers can configure model invocation logging to collect Bedrock invocation logs, model input data, and model output data for all invocations in the AWS account used in Amazon Bedrock. Customers can also use CloudTrail to monitor Amazon Bedrock API calls.

The company said AWS customers also can use services such as GuardDuty to detect potential security concerns and Billing Alarms to provide notifications of abnormal billing activity. Finally, AWS Cost Explorer is intended to give customers a way to visualize and manage Bedrock costs and usage over time.

Anthropic told KrebsOnSecurity it is always working on novel techniques to make its models more resistant to jailbreaks.

“We remain committed to implementing strict policies and advanced techniques to protect users, as well as publishing our own research so that other AI developers can learn from it,” Anthropic said in an emailed statement. “We appreciate the research community’s efforts in highlighting potential vulnerabilities.”

Anthropic said it uses feedback from child safety experts at Thorn around signals often seen in child grooming to update its classifiers, enhance its usage policies, fine tune its models, and incorporate those signals into testing of future models.

Update: 5:01 p.m. ET: Chub has issued a statement saying they are only hosting the role-playing characters, and that the LLMs they use run on their own infrastructure.

“Our own LLMs run on our own infrastructure,” Chub wrote in an emailed statement. “Any individuals participating in such attacks can use any number of UIs that allow user-supplied keys to connect to third-party APIs. We do not participate in, enable or condone any illegal activity whatsoever.”


44 thoughts on “A Single Cloud Compromise Can Feed an Army of AI Sex Bots

  1. Daniel

    Sex, Money, Power, three factors that will always be in play in security.

    Reply
      1. wanto

        that is a crazy and valid response, although he is probably right in his statement

        Reply
  2. Phil

    It is on a board called /g/ – technology. /aicg/ tells how they steal from corporations and there is serious issues there

    Reply
  3. Maria

    I wouldn’t go there either. It’s on /g/ and they leak lots of stuff. The content is horrific and I hope something is done about this.

    Reply
  4. John

    >which use custom jailbreaks to bypass content filtering
    There’s nothing to bypass on Claude besides their weak alignment training (as opposed to the web version). AWS has a hard filter for copyrighted materials but that’s it. Yes, all “safety benchmarks” are useless (they usually put Claude far ahead), they’re either lazy or paid. Anthropic are relying on other methods to enforce fair use of their API, in an attempt to minimize false positives.

    Reply
  5. Tyrrell

    Please don’t enter that website it logs your IP address to sell to dark fully automatic jackers.

    Reply
  6. 'jo

    erm… okay but, are you actually gonna STOP me from gooning to AI sex bots, or are you just a scared techlet who can’t actually DO anything to me btwfullybeit?

    Reply
    1. soyGOD

      >ummm… why is krebs trying to stop me from ‘ooning to my blvcked ‘lita ai soybot? isn’t this like… heckin’ dystopian o algo

      Reply
  7. bigjohn

    There’s a proxy service called charybdis scylla This is a major “proxy service” run by a guy named “Drag0n3xt” and his partner “Girko” on /aicg/ (The forum general thread discussing jailbreaks, on 4chan’s /g/ board). It seems to use this software called “Drago/oai-reverse-proxy” They serve hundreds of users every minute of the day. They even have a Patreon it seems like for providing the stolen access.

    — BIG JOHN

    Reply
  8. Bigjohn

    There’s a proxy service called charybdis scylla This is a major “proxy service” run by a guy named “Drag0n3xt” and his partner “Girko” on /aicg/ (The forum general thread discussing jailbreaks, on 4chan’s /g/ board). It seems to use this software called “Drago/oai-reverse-proxy” They serve hundreds of users every minute of the day. They even have a Patreon it seems like for providing the stolen access.

    Disgusting!!

    — Big John

    Reply
  9. rich56k

    While this may be slightly off topic(?) But I find it hard to feel any sympathy for any ‘owner’ of an LLM being accessed without permission (kinda ironic – big picture) Usually(?) due to the also classic lack of security protocols that allows said ‘violations’ which Brian consistently reports (Thanks Brian!!). Actually kinda karmic if you ask me……especially if/when on the hook for (only) $46K/day. Now they specifically may or may not fit the following profile, but its clearly a trend/MO. Why you ask?

    Well considering the classic/ongoing big tech standard of ‘better to apologize later than ask (pay) for permission first’ mindset, (can you say Marky Mark boys & girsl?!) and the relatively few copyright infringement suits already filed against the compilers of said LLMs; who are continually scraping content from any/everybody that has ever posted content online (and no that doesn’t make it fair game to be utilized).

    It’s both way too late, and pretty tough to know exactly who has already had their content stolen and used for existing LLMs. While some LLMs claim to pay (a few victims) for some content, simple common sense says they’ve built the models without…Can’t put the toothpaste back in the tube, right Sam (and your billion dollar valuations)?

    Reply
  10. TekDaddy

    I, for one, am shocked at the potential Anthropic’s text models have for AI generated text adventures.
    While I do not endorse much of the content seen in the harvested text logs, I must say I got rather excited reading some of the stories involved. The sheer diversity of content was something incredible. From morally depraved stories that I will not discuss, to interracial cuckoldry (a personal fav of mine, it’s inspired some rather exciting times in the bedroom with my partner, or should I say PARTNERS…), to quite frankly alien and unusual stories that go beyond what one might consider “normal”.
    Sadly the publicly available avenues for high performance LLMs is rather limited, and these LLMjackers are taking advantage of this when really the fat stacks they’re making should be going to Jeff Bezos and Dario Amodei instead. Perhaps in the future, when consumer-level LLM access is more widespread, we’ll see an end to this heinous and rampant theft, as well as proper filtering in place so only wholesome adult content like yuri and interracial swinging can be produced.

    Reply
  11. just me :3

    nyooo!!! what about the big corporations?? are they going to be okie? T^T

    Reply
  12. Mauv4r0.Skotadi

    Wow, Nice.. Recent event on a peers phone with Issues caused bettween the couple over fake sexual conflict. Nice Catch !!

    Reply
  13. Catwhisperer

    Billing Alarms and AWS Cost Explorer work, folks. As long as you turn them, configure them, and monitor the emails, LOL. However, if I recall correctly, all that is initially off by default. Why is AWS not forcing you, as “security best practices,” to configure both?

    Reply
    1. Fr00tL00ps

      It seems Brian stirred the 4chan pot and the incel flame trolls come hither.
      ¯⁠\⁠_⁠(⁠ツ⁠)_/¯ Go figure.

      Reply
  14. Treyvon E.

    This article is extremely disingenuous. If you tried, you could not have managed to fearmonger as hard as this piece did. Also, Amazon’s a billion dollar company. Who cares? Just a drop in the ocean to ’em. Grow the hell up.

    Reply
  15. ngldesu

    guys did you know people will steal your api keys off of github if you publish them in public
    Slow week on breachforums, krebs?

    Reply
  16. KontolBucat

    People should normalize sex ai bot so no real women getting harmed.

    Reply
    1. Catwhisperer

      Which is probably a fine idea if you do it legitimately. Start a company, get financing, hire programmers. Just to give you an idea of the potential of doing it on the up and up, Bet365 in Denver, CO just started, hiring 300 and expecting to hire 1000. Now we all know betting can be like crack to some and just as addictive as porn, but there’s money to be made, and the three letter agencies won’t hound you and put multi-million dollar bounties on your head if you do it right.

      There is legitimate money to be made providing AI ‘escort services’, because mostly men feel the need to discipline their pet monkeys. But stealing somebody’s AWS account and going that route has you leaving real money on the table, and puts the three letter agencies on your trail. Think about it…

      https://www.denverpost.com/2024/09/30/bet365-denver-headquarters-sports-betting/

      Reply
  17. A. R.

    I’m basically with earlier commenter KontolBucat since morality-policing fiction and fantasy is ridiculous, but also like, burning up the planet for mediocre LLM drivel isn’t the answer either. Just read ao3 or whatever.

    Also, that world without feminism line quoted in the article is pretty ironic, I bet trump would just love to ban sex stories, first amendment be darned!

    Reply

Leave a Reply

Your email address will not be published. Required fields are marked *