Originality.AI looked at 8,885 long Facebook posts made over the past six years.
Key Findings
- 41.18% of current Facebook long-form posts are Likely AI, as of November 2024.
- Between 2023 and November 2024, the average percentage of monthly AI posts on Facebook was 24.05%.
- This reflects a 4.3x increase in monthly AI Facebook content since the launch of ChatGPT. In comparison, the monthly average was 5.34% from 2018 to 2022.
I wouldn’t be surprised, but I’d be interested to see what they used to make that determination. All of the AI detection I know of are prone to a lot of false-positives.
When I was looking for a job, I ran into a guide to make money using AI:
-
Choose a top selling book.
-
Ask Chat GPT to give a summary for each chapter.
-
Paste the summaries into Google docs.
-
Export as PDF.
-
Sell on Amazon as a digital “short version” or “study guide” for the original book.
-
Repeat with other books.
Blew my mind how much hot stinking garbage is out there.
-
The other 60% are old people re-sharing it.
6% old people re-sharing. The other 54% were bot accounts.
Thanks.
Now do Reddit comments.There’s an AI reply option now. Interested to know how far that is off just being part of the regular comments.
In the last month it has become a barrage. The algorithms also seem to be in overdrive. If I like something I get bombarded with more stuff like that within a day. I’d say 90% of my feed is shit that has nothing to do with anyone I know.
If it wasn’t a way to stay in touch with family and friends I’d bail.
A friend told me he saw 16 posts before he saw a post from a friend or page he’d liked.
I’m a big fan of a particularly virtual table-top tool called Foundry, which I use to host D&D games.
The Instagram algorithm picked this out of my cookies and fed it to Temu, which determined I must really like… lathing and spot-wielding and shit. So I keep getting ads for miniature industrial equipment. At-home tools for die casting and alloying and the like. From Temu! Absolutely crazy.
I made the mistake of clicking like on an Indian machine shop (I admired how they made do with crude conditions). Well now I get bombarded with not just those videos but Mexican welding shops, Pakistani auto repair places…
and, is the jury already in on which ai is most fuckable?
I’d tell you, but my area network appears to have already started blocking DeepSeek.
Deekseek that was not encrypting data
https://www.theregister.com/2025/01/30/deepseek_database_left_open/
Deleted my account a little while ago but for my feed I think it was higher. You couldn’t block them fast enough, and mostly obviously AI pictures that if the comments are to be believed as being actual humans…people believed were real. It was a total nightmare land. I’m sad that I have now lost contact with the few distant friends I had on there but otherwise NOTHING lost.
I was wondering who Facebook was for, good to know AI has low standards
Dead internet theory
FB has been junk for more than a decade now, AI or no.
I check mine every few weeks because I’m a sports announcer and it’s one way people get in contact with me, but it’s clear that FB designs its feed to piss me off and try to keep me doomscrolling, and I’m not a fan of having my day derailed.
I deleted facebook in like 2010 or so, because i hardly ever used it anyway, it wasn’t really bad back then, just not for me. 6 or so years later a friend of mine wanted to show me something on fb, but couldn’t find it, so he was just scrolling, i was blown away how bad it was, just ads and auto played videos and absolute garbage. And from what i understand, it just got worse and worse. Everyone i know now that uses facebook is for the market place.
My brother gave me his Facebook credentials so I could use marketplace without bothering him all the time. He’s been a liberal left-winger all his life but for the past few years he’s taken to ranting about how awful Democrats are (“Genocide Joe” etc.) while mocking people who believe that there’s a connection between Trump and Putin. Sure enough, his Facebook is filled with posts about how awful Democrats are and how there’s no connection between Trump and Putin - like, that’s literally all that’s on there. I’ve tried to get him to see that his worldview is entirely created by Facebook but he just won’t accept it. He thinks that FB is some sort of objective collator of news.
In my mind, this is really what sets social media apart from past mechanisms of social control. In the days of mass media, the propaganda was necessarily a one-size-fits-all sort of thing. Now, the pipeline of bullshit can be custom-tailored for each individual. So my brother, who would never support Trump and the Republicans, can nevertheless be fed a line of bullshit that he will accept and help Trump by not voting (he actually voted Green).
It’s such a cesspit.
I’m glad we have the Fediverse.
And 58.82% are likely generated by human junk then.
If you want to visit your old friends in the dying mall. Go to feeds then friends. Should filter everything else out.
That’s an extremely low sample size for this
8,855 long-form Facebook posts from various users using a 3rd party. The dataset spans from 2018 to November 2024, with a minimum of 100 posts per month, each containing at least 100 words.
seems like thats a good baseline rule and that was about the total number that matched it
With apparently 3 billion active users
Only summing up 9k posts over a 6 year stretch with over 100 words feels like an outreach problem. Conclusion could be drawn that bots have better reach
each post has to be 100 words with at least 100 posts a month
how many actual users do that?
I have no idea because I don’t use the site
But to say less than 0.0001% just seems hard to believe
I don’t use the site either but 100 words is a lot for a facebook post
deleted by creator
Not my Annie! No! Not my Annie!
This kind of just looks like an add for that companies AI detection software NGL.
this whole concept relies on the idea that we can reliably detect AI, which is just not true. None of these “ai detector” apps or services actually works reliably. They have terribly low success rates. the whole point of LLMs is to be indistinguishable from human text, so if they’re working as intended then you can’t really “detect” them.
So all of these claims, especially the precision to which they write the claims (24.05% etc), are almost meaningless unless the “detector” can be proven to work reliably.
Not enough attention is given to the literal arms race we find ourselves in. Most big tech buzz is all “yay innovation!” Or “oh no, jobs!”
Don’t get me wrong, the impact AI will have on pretty much every industry shouldn’t be underestimated, and people are and will lose their jobs.
But information is power. Sun Tzu knew this a long time ago. The AI arms race won’t just change job markets - it will change global markets, public opinion, warfare, everything.
The ability to mass produce seemingly reliable information in moments - and the consequent inability to trust or source information in a world flooded by it…
I can’t find the words to express how dangerous it is. The long-term consequences are going to be on par with - and terribly codependent with - the consequences of the industrial revolution.
If you could reliably detect “AI” using an “AI” you could also use an “AI” to make posts that the other “AI” couldn’t detect.
Sure, but then the generator AI is no longer optimised to generate whatever you wanted initially, but to generate text that fools the detector network, thus making the original generator worse at its intended job.
I see no reason why “post right wing propaganda” and "write so you don’t sound like “AI” " should be conflicting goals.
The actual argument why I don’t find such results credible is that the “creator” is trained to sound like humans, so the “detector” has to be trained to find stuff that does not sound like humans. This means, both basically have to solve the same task: Decide if something sounds like a human.
To be able to find the “AI” content, the “detector” would have to be better at deciding what sounds like a human than the “creator”. So for the results to have any kind of accuracy, you’re already banking on the “detector” company having more processing power / better training data / more money than, say, OpenAI or google.
But also, if the “detector” was better at the job, it could be used as a better “creator” itself. Then, how would we distinguish the content it created?
I’m not necessarily saying they’re conflicting goals, merely that they’re not the same goal.
The incentive for the generator becomes “generate propaganda that doesn’t have the language chatacteristics of typical LLMs”, so the incentive is split between those goals. As a simplified example, if the additional incentive were “include the word bamboo in every response”, I think we would both agree that it would do a worse job at its original goal, since the constraint means that outputs that would have been optimal previously are now considered poor responses.
Meanwhile, the detector network has a far simpler task - given some input string, give back a value representing the confidence it was output by a system rather than a person.
I think it’s also worth considering that LLMs don’t “think” in the same way people do - where people construct an abstract thought, then find the best combinations of words to express that thought, an LLM generates words that are likely to follow the preceding ones (including prompts). This does leave some space for detecting these different approaches better than at random, even though it’s impossible to do so reliably.
But I guess really the important thing is that people running these bots don’t really care if it’s possible to find that the content is likely generated, just so long as it’s not so obvious that the content gets removed. This means they’re not really incentivised to spend money training models to avoid detection.