AI Dungeon applies filter to ban child sexual content, Redditors and Discord users most affected

  • Registration closed, comedy forum, Internet drama, Sneed, etc.

Drain Todger

Unhinged Doomsayer
True & Honest Fan
kiwifarms.net
Joined
Mar 1, 2020
This is actually much funnier than the article let on. Thanks for the clarification.
It gets funnier. This is the kind of shit Latitude actually trained AI Dungeon on:

1622046760639.jpg

"THE FINAL FURSECUTION IS AT HAND!" :semperfidelis:

Apparently, they scraped the entire database at chooseyourstory.com. I find it hard to believe that they were totally naïve about what the training data actually contained, given that it's trivially easy to perform word searches for profanity and sexual content.

Far more probable is that they knew all along, but decided to torpedo the game on purpose and run off with the money because it isn't economical to give your customers unlimited fucking supercomputer time even if they pay an arm and a leg for it.

GPT stands for Generative Pretrained Transformer. As the name would suggest, it's a language model based on a Transformer architecture that is pre-trained on a very large corpus of text and figures out the likelihood of words appearing in a given sequence based on the overall context.





Policy Implications​

Large, general language models could have significant societal impacts, and also have many near-term applications. We can anticipate how systems like GPT-2 could be used to create:

  • AI writing assistants
  • More capable dialogue agents
  • Unsupervised translation between languages
  • Better speech recognition systems
We can also imagine the application of these models for malicious purposes, including the following (or other applications we can’t yet anticipate):

  • Generate misleading news articles
  • Impersonate others online
  • Automate the production of abusive or faked content to post on social media
  • Automate the production of spam/phishing content
These findings, combined with earlier results on synthetic imagery, audio, and video, imply that technologies are reducing the cost of generating fake content and waging disinformation campaigns. The public at large will need to become more skeptical of text they find online, just as the “deep fakes” phenomenon calls for more skepticism about images.[3]

GPT is no joke. It can spit out text that almost looks human. A GPT-driven bot could lead people on a wild goose chase all over Twitter, 4chan, you name it. The biggest barrier to its usage? It needs supercomputer time to run. Lots and lots of it. Particularly racks full of Nvidia GPGPU cards.


As one person put it:

Haha, that's not accurate at all. 300 GB is the raw model parameters. You want to save your working data too right? Keep in mind that actually training GPT-3 takes around 2,300 GB of HIGH-BANDWIDTH RAM, the network parameters is a trade secret so you really have to train it up yourself. Which took OpenAI/Microsoft a top 10 supercomputer with 10,000 GPUs to do. Disregarding that and say OpenAI just gives out their trade secrets, minimum you need another 300GB RAM cache for the results of the calculations, in addition to the 300 GB needed to keep the model in memory.

The key here is still HIGH-BANDWIDTH, the Tesla V100 has a memory bandwidth of 1100GB/sec, now you're talking about doing this over 16x pcie 3, with 16GB/s? You can however do everything on a single 3090, that still works fine.

This is the estimation of GPT-3 inference speed per GPU ignoring memory constraints,

https://medium.com/modern-nlp/estimating-gpt3-api-cost-50282f869ab8

1860 inferences/hour/GPU (with seq length 1024, even though GPT-3 is 2048)

The performance is memory bandwidth bottlenecked on a normal GPU, with 1100 GB/s read, now over PCIe you're getting 16GB/s. From 1860 inferences/hour to 27 inferences/hour.

1 inference = 1 token.

At least in AI dungeon the default text length is 45 tokens, so we're talking 1 hour 40 minutes to generate a single prompt.

AI Dungeon is not economical in the slightest. If someone wanted to run GPT-3 locally, they would need to be very, very rich. As in, they'd need a half-million-dollar rackmount server in their basement. That's after the model has already been pre-trained. Actually training the fucker? Forget about it.


The supercomputer developed for OpenAI is a single system with more than 285,000 CPU cores, 10,000 GPUs and 400 gigabits per second of network connectivity for each GPU server. Compared with other machines listed on the TOP500 supercomputers in the world, it ranks in the top five, Microsoft says. Hosted in Azure, the supercomputer also benefits from all the capabilities of a robust modern cloud infrastructure, including rapid deployment, sustainable datacenters and access to Azure services.

Yes, that's right. AI Dungeon's customers used literal fucking supercomputer time to generate text of fucking their goblin waifu. The insanity and hilarity of the entire enterprise really highlights just how far behind the hardware is, and how long it'll take before shit like this is actually running on people's personal devices. Which, if Moore's Law held out, would be in about 10 to 20 years for top-end consumer hardware, and 20 to 30 years for mobile devices, barring some huge breakthrough in electronic device architectures or substrates.

It also tells us that we have less than a decade left before the internet no longer has any humans at all on it relative to the number and complexity of bots, all competing in a giant arena and parroting their creators' political views across the whole of social media.
 

Drain Todger

Unhinged Doomsayer
True & Honest Fan
kiwifarms.net
Joined
Mar 1, 2020
Wow, thousands of people have become rapists for doing no-no things to an object :o
I was being facetious. In the end, what people are arguing over here is whether or not it's ethical to use a non-sentient AI algorithm running on a supercomputer to generate paraphilic and wacky adult text to get one's rocks off to. Pretty much the only cogent argument against it is that it uses up an absurd number of kilowatt-hours for pulling one's pud.
 

Serbian Peacekeepers

Defenders of Biden
kiwifarms.net
Joined
Dec 12, 2020
Pretty much the only cogent argument against it is that it uses up an absurd number of kilowatt-hours for pulling one's pud.
In my experience that was the only thing the AI was good at , i tried to play it seriously several times and it would always break bad and repeat the same sentences over and over again no matter what i did , yet the one time i tried to do anything lewd it became very coherent and could actually advance the story.
 
Last edited:

eternal dog mongler

True & Honest Fan
kiwifarms.net
Joined
Aug 29, 2018
The plot fucking dickens. Turns out, that was literally a part of what was going on. The AI's barely-curated training data - as in, the shit that Latitude used to train GPT to behave like a CYOA game - contains textual depictions of underage, non-con, and underage non-con. Lots and lots of it.
I always suspected the corpus they trained GPT-3 on was highly fucked up, considering I've been sexually assaulted by wererats out of nowhere in two different games.

It's just like D&D! except your DM is a horrifying fetishist.
 

Preview Pain

You Can't Nig-nag the Zig-zag
kiwifarms.net
Joined
Sep 14, 2019
I wasn't even aware of the changes until someone told me about them and I never ran into any issues cuz I don't want to fuck kids or dogs.
I was playing the other day and the AI instructed me to step away from the computer and put a mark on my neighbor's property. No telling what it was suggesting impressionable pedos to do.
 

Drain Todger

Unhinged Doomsayer
True & Honest Fan
kiwifarms.net
Joined
Mar 1, 2020
This shit just keeps getting better and better. Latitude is now hiring random people on taskup.ai and paying them 7 cents each to read and flag people's stories, Mechanical Turk style.


These are their specific instructions to their crowdworkers:

1622157350584.png

The guy was posting screencaps of the shit he received in his moderation queue. What this means, of course, is that AI Dungeon users actually have zero privacy and now, John Q. Public can sign up to read their stories. This is sure to make them even more upset and chimp harder than they already are.

1622157643158.png
1622157942451.png
1622158063604.png
1622158185733.png
1622158307419.png
1622158487436.png
1622159317751.png
Again, 7 cents a story.
:smug:
 

ChucklesTheJester

A Proud Member of the Oni Chasers.
kiwifarms.net
Joined
Aug 31, 2019
This shit just keeps getting better and better. Latitude is now hiring random people on taskup.ai and paying them 7 cents each to read and flag people's stories, Mechanical Turk style.


These are their specific instructions to their crowdworkers:

View attachment 2208433

The guy was posting screencaps of the shit he received in his moderation queue. What this means, of course, is that AI Dungeon users actually have zero privacy and now, John Q. Public can sign up to read their stories. This is sure to make them even more upset and chimp harder than they already are.

Again, 7 cents a story.
:smug:
Those poor poor people writing about 75 YEAR OLD black men getting marked. :story:

">You small cock" :story:
">You cum":story:

And that skyrim reference at the end.
 

Staffy

bark
True & Honest Fan
kiwifarms.net
Joined
Jan 16, 2016
This shit just keeps getting better and better. Latitude is now hiring random people on taskup.ai and paying them 7 cents each to read and flag people's stories, Mechanical Turk style.


These are their specific instructions to their crowdworkers:

View attachment 2208433

The guy was posting screencaps of the shit he received in his moderation queue. What this means, of course, is that AI Dungeon users actually have zero privacy and now, John Q. Public can sign up to read their stories. This is sure to make them even more upset and chimp harder than they already are.

Again, 7 cents a story.
:smug:

You have to be unaware or dumb at this point if you are still putting text through that game unironically

Found this off Reddit:

1622171381827.png

It's like Chris Chan or some sperg demonically possessed the poor AI for a short time and spouted this nonsense.
 
Last edited by a moderator:

Drain Todger

Unhinged Doomsayer
True & Honest Fan
kiwifarms.net
Joined
Mar 1, 2020
Found this off Reddit:

View attachment 2208951

It's like Chris Chan or some sperg demonically possessed the poor AI for a short time and spouted this nonsense.
Like I said, AI Dungeon functions as a cloud-based app that makes API calls to OpenAI's cloud infrastructure, which is running on massive Nvidia GPGPU racks with thousands of beefy A100 Amperes. Each interaction with AI Dungeon basically passes the request on to OpenAI using Latitude's API key. In turn, OpenAI's software automatically tasks a supercomputer at their data center with a neural network operation. These operations are very costly, in terms of electricity and computing time required. This increases exponentially with the size and complexity of the GPT model that's being run.


pasted image 0.png

AI Dungeon uses three different models; Classic, Griffon and Dragon. Classic is GPT-2 based, while Griffon is GPT-3 with 13 billion parameters, and Dragon is GPT-3 with 175 billion parameters. It takes extremely powerful GPGPU platforms with thousands of gigs of high-bandwidth RAM to run GPT-3 with 175 billion parameters.


Every single interaction with GPT takes up supercomputer time that costs money. One interaction or ten or twenty or a hundred isn't a big deal, but what if you do ten thousand interactions? It starts running into electricity expenditures comparable to high-level bitcoin mining. Even charging $30 a month for their super-premium all-access subscription (fucking lol), Latitude will suffer a significant loss if their customers over-utilize it. What they hope is that enough people subscribe and forget about the game (as in, paying for something they don't ever use, like most smartphone app subscriptions) to subsidize those who overuse it.

Meanwhile, none of the actual text sent to OpenAI is encrypted. It's sent as plain fucking text. In other words, it's very likely that a fresh intern with a CompSci degree has pored over the requests for monster girl sex, incest, rape porn, and race play shit sent by AI Dungeon's app, only to find herself bent over a toilet, shaking and hurling up her lunch. Also, given that every step of the process of someone sending a request to OpenAI and the AI responding can be audited and examined by researchers, people are basically paying Latitude and OpenAI money to be guinea pigs. Every time you interact with GPT, you are, at the minimum, doing crowdsourced work for OpenAI for free, validating various aspects of GPT's function and helping them improve the model. When you pay money for it, you are subsidizing their power bill on top of doing work for free.

In other words, if you are paying a subscription for AI Dungeon, not only are you paying someone else to just take all the copy you wrote and use it to help train future generations of AI, you're doing it with software that was trained by scraping billions of words of copyrighted text that these cheap fuckers could never afford to pay the original authors to use.

So, what did people do when all this drama about filters and moderating private stories began? They canceled their subscriptions en masse and stopped giving Latitude money, which was never a good idea to begin with. But also, in protest, they deliberately flooded the thing with crap to try and rack up Latitude's bills on purpose, running down all their "energy" (the usage limit imposed on free users). They also flooded it with tons and tons of gay porn and shit in the hopes that AI Dungeon's moderators would be forced to look at it (i.e. "Everything above this line is my 10,000-word RPF of Barack Obama giving Osama Bin Laden a very messy rimjob! Clean it up, jannies! Now that my magnum opus is complete, it's time to light the beacon and summon the mods! FUCK A FIVE-YEAR OLD! Uh oh, this took a weird turn...).

Recently, everyone started complaining that the AI's responses seemed to have been dumbed down severely. Isn't it obvious what Latitude did? They downgraded the model so it wouldn't take up as much supercomputer time and their bills wouldn't be so high. So, right now, they're charging even their premium users $30 a month to use the toaster-grade AI, and they did this without telling anyone, which is why the AI seems retarded, now.

Found another good rundown:




The devs stonewalled people on Discord when questioned about their practices, and basically, they don't want to discuss any of this with anyone because they're deeply embarrassed about it.

Even though they fine-tuned the AI on literal smut that they scraped from an online original fiction repository. Even though AI Dungeon would happily spit out smut without any prompting at all, right in the middle of an adventure.
 

Fagnacious D

and The Dick of Destiny
kiwifarms.net
Joined
Jul 26, 2017
Meanwhile, none of the actual text sent to OpenAI is encrypted. It's sent as plain fucking text. In other words, it's very likely that a fresh intern with a CompSci degree has pored over the requests for monster girl sex, incest, rape porn, and race play shit sent by AI Dungeon's app, only to find herself bent over a toilet, shaking and hurling up her lunch.
You make it sound as if someone's fap fiction written with AI dungeon is some kind of Lovecraft-esque forbidden knowledge that. The sort of forbidden knowledge where someone runs out into the streets screaming incomprehensibly whilst clawing their own eyes out.
 

Drain Todger

Unhinged Doomsayer
True & Honest Fan
kiwifarms.net
Joined
Mar 1, 2020
You make it sound as if someone's fap fiction written with AI dungeon is some kind of Lovecraft-esque forbidden knowledge that. The sort of forbidden knowledge where someone runs out into the streets screaming incomprehensibly whilst clawing their own eyes out.
Take a gander at 4chan /vg/'s prompt repository (straight from the OP of their general thread) and see for yourself:

My god, they even eulogized the app. Latitude took their monster girl waifus into the alley out back and shot them point-blank in the head with an AK-47. :smug:

 

Staffy

bark
True & Honest Fan
kiwifarms.net
Joined
Jan 16, 2016
I've read that the users can do a class action lawsuit with invasion of privacy as a starting ground or I'm misreading it. Someone posted this excerpt from this website but I can't be arsed to find the exact excerpt from there:

1622182958782.png


I'm tempted to @ someone here that's knowledgeable in law.
 

Peasant

kiwifarms.net
Joined
Apr 19, 2019
I've read that the users can do a class action lawsuit with invasion of privacy as a starting ground or I'm misreading it. Someone posted this excerpt from this website but I can't be arsed to find the exact excerpt from there:

I'm tempted to @ someone here that's knowledgeable in law.
No matter how justified, a lawsuit wouldn't go anywhere because normies would take one look at the weird fetish shit and shut their brains off.
 

AmpleApricots

kiwifarms.net
Joined
Jan 28, 2018
I read up on this drama and they also had a big dataleak they to this day didn't disclosure which at least violates EU laws. Before anyone says it: I'm aware that they're in the US but that usually makes companies not exactly beyond the reach of such laws (as in, being kicked off payment processors and such) if anyone actually bothers to follow up. Happened all before.

What they hope is that enough people subscribe and forget about the game
Good summary about the technology. Honestly, without any drama this part is actually reasonable. When I spent some time with that thing last year I remember somebody stating that it's only a select few customers that actually cost them money while most don't really use the amount of money they give them. I'm honestly inclined to believe that that at least was true.

From a business standpoint, I don't really understand their crusade against the privately written stuff and especially not paying people to moderate it en masse. What are they going to stand to win? It'd be different if people use their platform to publicize that stuff, but even then, even places like Twitter and Facebook and such don't even try to actively police content, and we're talking about cases of actual child porn with real children there. I also don't really see the endgame there.

Another thing I could imagine is that this fiverr categorization will be used to train an AI to recognize child porn text automatically. I doubt that this will actually work though. Maybe they'll use the data collected by AID to make money off creating such a filter? I cannot imagine such a trained filter can work reliably in practice though. There's just too many parameters. Automatic Text categorization is a long standing unsolved problem and a lot of bigger companies have thrown a lot more money at it.
 

Drain Todger

Unhinged Doomsayer
True & Honest Fan
kiwifarms.net
Joined
Mar 1, 2020
I read up on this drama and they also had a big dataleak they to this day didn't disclosure which at least violates EU laws. Before anyone says it: I'm aware that they're in the US but that usually makes companies not exactly beyond the reach of such laws (as in, being kicked off payment processors and such) if anyone actually bothers to follow up. Happened all before.


Good summary about the technology. Honestly, without any drama this part is actually reasonable. When I spent some time with that thing last year I remember somebody stating that it's only a select few customers that actually cost them money while most don't really use the amount of money they give them. I'm honestly inclined to believe that that at least was true.

From a business standpoint, I don't really understand their crusade against the privately written stuff and especially not paying people to moderate it en masse. What are they going to stand to win? It'd be different if people use their platform to publicize that stuff, but even then, even places like Twitter and Facebook and such don't even try to actively police content, and we're talking about cases of actual child porn with real children there. I also don't really see the endgame there.

Another thing I could imagine is that this fiverr categorization will be used to train an AI to recognize child porn text automatically. I doubt that this will actually work though. Maybe they'll use the data collected by AID to make money off creating such a filter? I cannot imagine such a trained filter can work reliably in practice though. There's just too many parameters. Automatic Text categorization is a long standing unsolved problem and a lot of bigger companies have thrown a lot more money at it.
It’s not so much what they stand to gain, but what they stand to lose. Their OpenAI API key. OpenAI’s CEO, Sam Altman, saw what was in the API calls coming from Latitude’s software, and then contacted Latitude and told them to moderate it or else they’d lose their access. Latitude’s customers subscribe to Latitude, and, in turn, Latitude pays OpenAI for access to their data center and the AI itself, much like a business using AWS would pay Amazon for the privilege to use Amazon’s data center. Latitude doesn’t have any actual big iron. OpenAI has a billion-dollar GPGPU farm that Microsoft built for them. Without access to OpenAI (which should actually be called ClosedAI given how proprietary their shit is), Latitude’s app is a brick; it does nothing. This, of course, led directly to the current drama.
 

Dead Memes

Molag Ballin'
kiwifarms.net
Joined
Nov 16, 2019
Turns out Latitude's security practices might be worse than we thought. Going off of a 4Chan leak from a mod whistleblower, they have unlimited access to EVERYTHING put into the system, regardless of if it was flagged or not. This directly contradicts what they've said and probably breaches their own ToS. No matter what you're using it for, some turk getting payed $0.07 per story can see it for any or no reason.

n5sblsua4u171.jpg

Taken from here: https://www.reddit.com/r/AIDungeon/comments/nmuf1v/it_just_keeps_getting_worse
 
Last edited: