Social Media

everything-that-could-go-wrong-with-x’s-new-ai-written-community-notes

Everything that could go wrong with X’s new AI-written community notes


X says AI can supercharge community notes, but that comes with obvious risks.

Elon Musk’s X arguably revolutionized social media fact-checking by rolling out “community notes,” which created a system to crowdsource diverse views on whether certain X posts were trustworthy or not.

But now, the platform plans to allow AI to write community notes, and that could potentially ruin whatever trust X users had in the fact-checking system—which X has fully acknowledged.

In a research paper, X described the initiative as an “upgrade” while explaining everything that could possibly go wrong with AI-written community notes.

In an ideal world, X described AI agents that speed up and increase the number of community notes added to incorrect posts, ramping up fact-checking efforts platform-wide. Each AI-written note will be rated by a human reviewer, providing feedback that makes the AI agent better at writing notes the longer this feedback loop cycles. As the AI agents get better at writing notes, that leaves human reviewers to focus on more nuanced fact-checking that AI cannot quickly address, such as posts requiring niche expertise or social awareness. Together, the human and AI reviewers, if all goes well, could transform not just X’s fact-checking, X’s paper suggested, but also potentially provide “a blueprint for a new form of human-AI collaboration in the production of public knowledge.”

Among key questions that remain, however, is a big one: X isn’t sure if AI-written notes will be as accurate as notes written by humans. Complicating that further, it seems likely that AI agents could generate “persuasive but inaccurate notes,” which human raters might rate as helpful since AI is “exceptionally skilled at crafting persuasive, emotionally resonant, and seemingly neutral notes.” That could disrupt the feedback loop, watering down community notes and making the whole system less trustworthy over time, X’s research paper warned.

“If rated helpfulness isn’t perfectly correlated with accuracy, then highly polished but misleading notes could be more likely to pass the approval threshold,” the paper said. “This risk could grow as LLMs advance; they could not only write persuasively but also more easily research and construct a seemingly robust body of evidence for nearly any claim, regardless of its veracity, making it even harder for human raters to spot deception or errors.”

X is already facing criticism over its AI plans. On Tuesday, former United Kingdom technology minister, Damian Collins, accused X of building a system that could allow “the industrial manipulation of what people see and decide to trust” on a platform with more than 600 million users, The Guardian reported.

Collins claimed that AI notes risked increasing the promotion of “lies and conspiracy theories” on X, and he wasn’t the only expert sounding alarms. Samuel Stockwell, a research associate at the Centre for Emerging Technology and Security at the Alan Turing Institute, told The Guardian that X’s success largely depends on “the quality of safeguards X puts in place against the risk that these AI ‘note writers’ could hallucinate and amplify misinformation in their outputs.”

“AI chatbots often struggle with nuance and context but are good at confidently providing answers that sound persuasive even when untrue,” Stockwell said. “That could be a dangerous combination if not effectively addressed by the platform.”

Also complicating things: anyone can create an AI agent using any technology to write community notes, X’s Community Notes account explained. That means that some AI agents may be more biased or defective than others.

If this dystopian version of events occurs, X predicts that human writers may get sick of writing notes, threatening the diversity of viewpoints that made community notes so trustworthy to begin with.

And for any human writers and reviewers who stick around, it’s possible that the sheer volume of AI-written notes may overload them. Andy Dudfield, the head of AI at a UK fact-checking organization called Full Fact, told The Guardian that X risks “increasing the already significant burden on human reviewers to check even more draft notes, opening the door to a worrying and plausible situation in which notes could be drafted, reviewed, and published entirely by AI without the careful consideration that human input provides.”

X is planning more research to ensure the “human rating capacity can sufficiently scale,” but if it cannot solve this riddle, it knows “the impact of the most genuinely critical notes” risks being diluted.

One possible solution to this “bottleneck,” researchers noted, would be to remove the human review process and apply AI-written notes in “similar contexts” that human raters have previously approved. But the biggest potential downfall there is obvious.

“Automatically matching notes to posts that people do not think need them could significantly undermine trust in the system,” X’s paper acknowledged.

Ultimately, AI note writers on X may be deemed an “erroneous” tool, researchers admitted, but they’re going ahead with testing to find out.

AI-written notes will start posting this month

All AI-written community notes “will be clearly marked for users,” X’s Community Notes account said. The first AI notes will only appear on posts where people have requested a note, the account said, but eventually AI note writers could be allowed to select posts for fact-checking.

More will be revealed when AI-written notes start appearing on X later this month, but in the meantime, X users can start testing AI note writers today and soon be considered for admission in the initial cohort of AI agents. (If any Ars readers end up testing out an AI note writer, this Ars writer would be curious to learn more about your experience.)

For its research, X collaborated with post-graduate students, research affiliates, and professors investigating topics like human trust in AI, fine-tuning AI, and AI safety at Harvard University, the Massachusetts Institute of Technology, Stanford University, and the University of Washington.

Researchers agreed that “under certain circumstances,” AI agents can “produce notes that are of similar quality to human-written notes—at a fraction of the time and effort.” They suggested that more research is needed to overcome flagged risks to reap the benefits of what could be “a transformative opportunity” that “offers promise of dramatically increased scale and speed” of fact-checking on X.

If AI note writers “generate initial drafts that represent a wider range of perspectives than a single human writer typically could, the quality of community deliberation is improved from the start,” the paper said.

Future of AI notes

Researchers imagine that once X’s testing is completed, AI note writers could not just aid in researching problematic posts flagged by human users, but also one day select posts predicted to go viral and stop misinformation from spreading faster than human reviewers could.

Additional perks from this automated system, they suggested, would include X note raters quickly accessing more thorough research and evidence synthesis, as well as clearer note composition, which could speed up the rating process.

And perhaps one day, AI agents could even learn to predict rating scores to speed things up even more, researchers speculated. However, more research would be needed to ensure that wouldn’t homogenize community notes, buffing them out to the point that no one reads them.

Perhaps the most Musk-ian of ideas proposed in the paper, is a notion of training AI note writers with clashing views to “adversarially debate the merits of a note.” Supposedly, that “could help instantly surface potential flaws, hidden biases, or fabricated evidence, empowering the human rater to make a more informed judgment.”

“Instead of starting from scratch, the rater now plays the role of an adjudicator—evaluating a structured clash of arguments,” the paper said.

While X may be moving to reduce the workload for X users writing community notes, it’s clear that AI could never replace humans, researchers said. Those humans are necessary for more than just rubber-stamping AI-written notes.

Human notes that are “written from scratch” are valuable to train the AI agents and some raters’ niche expertise cannot easily be replicated, the paper said. And perhaps most obviously, humans “are uniquely positioned to identify deficits or biases” and therefore more likely to be compelled to write notes “on topics the automated writers overlook,” such as spam or scams.

Photo of Ashley Belanger

Ashley is a senior policy reporter for Ars Technica, dedicated to tracking social impacts of emerging policies and new technologies. She is a Chicago-based journalist with 20 years of experience.

Everything that could go wrong with X’s new AI-written community notes Read More »

false-claims-that-ivermectin-treats-cancer,-covid-lead-states-to-pass-otc-laws

False claims that ivermectin treats cancer, COVID lead states to pass OTC laws

Doctors told the Times that they have already seen some cases where patients with treatable, early-stage cancers have delayed effective treatments to try ivermectin, only to see no effect and return to their doctor’s office with cancers that have advanced.

Risky business

Nevertheless, the malignant misinformation on social media has made its way into state legislatures. According to an investigation by NBC News published Monday, 16 states have proposed or passed legislation that would make ivermectin available over the counter. The intention is to make it much easier for people to get ivermectin and use it for any ailment they believe it can cure.

Idaho, Arkansas, and Tennessee have passed laws to make ivermectin available over the counter. On Monday, Louisiana’s state legislature passed a bill to do the same, and it now awaits signing by the governor. The other states that have considered or are considering such bills include: Alabama, Georgia, Kentucky, Maine, Minnesota, Mississippi, New Hampshire, North Carolina, North Dakota, Pennsylvania, South Carolina, and West Virginia.

State laws don’t mean the dewormer would be readily available, however; ivermectin is still regulated by the Food and Drug Administration, and it has not been approved for over-the-counter use yet. NBC News called 15 independent pharmacies in the three states that have laws on the books allowing ivermectin to be sold over the counter (Idaho, Arkansas, and Tennessee) and couldn’t find a single pharmacist who would sell it without a prescription. Pharmacists pointed to the federal regulations.

Likewise, CVS Health said its pharmacies are not currently selling ivermectin over the counter in any state. Walgreens declined to comment.

Some states, such as Alabama, have considered legislation that would protect pharmacists from any possible disciplinary action for dispensing ivermectin without a prescription. However, one pharmacist in Idaho, who spoke with NBC News, said that such protection would still not be enough. As a prescription-only drug, ivermectin is not packaged for retail sale. If it were, it would include over-the-counter directions and safety statements written specifically for consumers.

“If you dispense something that doesn’t have directions or safety precautions on it, who’s ultimately liable if that causes harm?” the pharmacist said. “I don’t know that I would want to assume that risk.”

It’s a risk people on social media don’t seem to be concerned with.

False claims that ivermectin treats cancer, COVID lead states to pass OTC laws Read More »

florida-ban-on-kids-using-social-media-likely-unconstitutional,-judge-rules

Florida ban on kids using social media likely unconstitutional, judge rules

A federal judge ruled today that Florida cannot enforce a law that requires social media platforms to block kids from using their platforms. The state law “is likely unconstitutional,” US Judge Mark Walker of the Northern District of Florida ruled while granting the tech industry’s request for a preliminary injunction.

The Florida law “prohibits some social media platforms from allowing youth in the state who are under the age of 14 to create or hold an account on their platforms, and similarly prohibits allowing youth who are 14 or 15 to create or hold an account unless a parent or guardian provides affirmative consent for them to do so,” Walker wrote.

The law is subject to intermediate scrutiny under the First Amendment, meaning it must be “narrowly tailored to serve a significant governmental interest,” must “leave open ample alternative channels for communication,” and must not “burden substantially more speech than is necessary to further the government’s legitimate interests,” the ruling said.

Florida claimed its law is designed to prevent harm to youth and is narrowly tailored because it targets sites that use specific features that have been deemed to be addictive. But the law applies too broadly, Walker found:

Even assuming the significance of the State’s interest in limiting the exposure of youth to websites with “addictive features,” the law’s restrictions are an extraordinarily blunt instrument for furthering it. As applied to Plaintiffs’ members alone, the law likely bans all youth under 14 from holding accounts on, at a minimum, four websites that provide forums for all manner of protected speech: Facebook, Instagram, YouTube, and Snapchat. It also bans 14- and 15-year-olds from holding accounts on those four websites absent a parent’s affirmative consent, a requirement that the Supreme Court has clearly explained the First Amendment does not countenance.

Walker said the Florida “law applies to any social media site that employs any one of the five addictive features under any circumstances, even if, for example, the site only sends push notifications if users opt in to receiving them, or the site does not auto-play video for account holders who are known to be youth. Accordingly, even if a social media platform created youth accounts for which none of the purportedly ‘addictive’ features are available, it would still be barred from allowing youth to hold those accounts if the features were available to adult account holders.”

Florida ban on kids using social media likely unconstitutional, judge rules Read More »

discord-lures-users-to-click-on-ads-by-offering-them-new-orbs-currency

Discord lures users to click on ads by offering them new Orbs currency

Sellis also announced that Discord is working with brand measurement firm Kantar to help advertisers track ad success. With Kantar technology, advertisers can measure things like “awareness, recall, and intent,” Sellis said. The partnership further underscores Discord’s growing reliance on advertising revenue.

“Our partnership with Discord is helping marketers better understand Discord as an advertising platform for new generations,” Nicole Jones, Kantar’s chief commercial lead, said on Discord’s blog.

Rethinking ads

Discord also announced this week that it will soon sell Play Quests to more advertisers. The announcement follows the company’s introduction of video ads to the Discord mobile app in June. Video Quests, as they’re called, allow advertisers to show trailers, announcements, and other types of content.

Overall, Discord’s new ad-friendly approach to business is very different than its previous strategy, which kept Discord ad-free from its 2015 launch until last year. Because the company is expected to go public soon, its leaders have determined that it’s no longer sufficient to rely completely on premium add-ons and subscriptions. Discord isn’t profitable, forcing the firm to reconsider its use of ads, which cofounder and CEO Jason Citron felt were too intrusive as recently as 2021.

Currently, Discord’s ads are limited to clickable sidebars within the platform and offer direct benefits to users. Introducing ads can be a slippery slope, though, especially for social media companies that prioritize ad revenue to please investors. On the other hand, another social media company, Reddit, has seen success by boosting its ad business. Reddit went public in March 2024 and became profitable in October 2024 after reporting a 60 percent year-over-year increase in ad revenue. Reddit has hinted at plans to introduce new and more types of ads, and we can expect Discord to consider the same after its IPO, which a March Bloomberg report suggested could happen as soon as this year.

Advance Publications, which owns Ars Technica parent Condé Nast, is the largest shareholder in Reddit.

Discord lures users to click on ads by offering them new Orbs currency Read More »

meta-hypes-ai-friends-as-social-media’s-future,-but-users-want-real-connections

Meta hypes AI friends as social media’s future, but users want real connections


Two visions for social media’s future pit real connections against AI friends.

A rotting zombie thumb up buzzing with flies while the real zombies are the people in the background who can't put their phones down

Credit: Aurich Lawson | Getty Images

Credit: Aurich Lawson | Getty Images

If you ask the man who has largely shaped how friends and family connect on social media over the past two decades about the future of social media, you may not get a straight answer.

At the Federal Trade Commission’s monopoly trial, Meta CEO Mark Zuckerberg attempted what seemed like an artful dodge to avoid criticism that his company allegedly bought out rivals Instagram and WhatsApp to lock users into Meta’s family of apps so they would never post about their personal lives anywhere else. He testified that people actually engage with social media less often these days to connect with loved ones, preferring instead to discover entertaining content on platforms to share in private messages with friends and family.

As Zuckerberg spins it, Meta no longer perceives much advantage in dominating the so-called personal social networking market where Facebook made its name and cemented what the FTC alleged is an illegal monopoly.

“Mark Zuckerberg says social media is over,” a New Yorker headline said about this testimony in a report noting a Meta chart that seemed to back up Zuckerberg’s words. That chart, shared at the trial, showed the “percent of time spent viewing content posted by ‘friends'” had declined over the past two years, from 22 to 17 percent on Facebook and from 11 to 7 percent on Instagram.

Supposedly because of this trend, Zuckerberg testified that “it doesn’t matter much” if someone’s friends are on their preferred platform. Every platform has its own value as a discovery engine, Zuckerberg suggested. And Meta platforms increasingly compete on this new playing field against rivals like TikTok, Meta argued, while insisting that it’s not so much focused on beating the FTC’s flagged rivals in the connecting-friends-and-family business, Snap and MeWe.

But while Zuckerberg claims that hosting that kind of content doesn’t move the needle much anymore, owning the biggest platforms that people use daily to connect with friends and family obviously still matters to Meta, MeWe founder Mark Weinstein told Ars. And Meta’s own press releases seem to back that up.

Weeks ahead of Zuckerberg’s testimony, Meta announced that it would bring back the “magic of friends,” introducing a “friends” tab to Facebook to make user experiences more like the original Facebook. The company intentionally diluted feeds with creator content and ads for the past two years, but it now appears intent on trying to spark more real conversations between friends and family, at least partly to fuel its newly launched AI chatbots.

Those chatbots mine personal information shared on Facebook and Instagram, and Meta wants to use that data to connect more personally with users—but “in a very creepy way,” The Washington Post wrote. In interviews, Zuckerberg has suggested these AI friends could “meaningfully” fill the void of real friendship online, as the average person has only three friends but “has demand” for up to 15. To critics seeking to undo Meta’s alleged monopoly, this latest move could signal a contradiction in Zuckerberg’s testimony, showing that the company is so invested in keeping users on its platforms that it’s now creating AI friends (wh0 can never leave its platform) to bait the loneliest among us into more engagement.

“The average person wants more connectivity, connection, than they have,” Zuckerberg said, hyping AI friends. For the Facebook founder, it must be hard to envision a future where his platforms aren’t the answer to providing that basic social need. All this comes more than a decade after he sought $5 billion in Facebook’s 2012 initial public offering so that he could keep building tools that he told investors would expand “people’s capacity to build and maintain relationships.”

At the trial, Zuckerberg testified that AI and augmented reality will be key fixtures of Meta’s platforms in the future, predicting that “several years from now, you are going to be scrolling through your feed, and not only is it going to be sort of animated, but it will be interactive.”

Meta declined to comment further on the company’s vision for social media’s future. In a statement, a Meta spokesperson told Ars that “the FTC’s lawsuit against Meta defies reality,” claiming that it threatens US leadership in AI and insisting that evidence at trial would establish that platforms like TikTok, YouTube, and X are Meta’s true rivals.

“More than 10 years after the FTC reviewed and cleared our acquisitions, the Commission’s action in this case sends the message that no deal is ever truly final,” Meta’s spokesperson said. “Regulators should be supporting American innovation rather than seeking to break up a great American company and further advantaging China on critical issues like AI.”

Meta faces calls to open up its platforms

Weinstein, the MeWe founder, told Ars that back in the 1990s when the original social media founders were planning the first community portals, “it was so beautiful because we didn’t think about bots and trolls. We didn’t think about data mining and surveillance capitalism. We thought about making the world a more connected and holistic place.”

But those who became social media overlords found more money in walled gardens and increasingly cut off attempts by outside developers to improve the biggest platforms’ functionality or leverage their platforms to compete for their users’ attention. Born of this era, Weinstein expects that Zuckerberg, and therefore Meta, will always cling to its friends-and-family roots, no matter which way Zuckerberg says the wind is blowing.

Meta “is still entirely based on personal social networking,” Weinstein told Ars.

In a Newsweek op-ed, Weinstein explained that he left MeWe in 2021 after “competition became impossible” with Meta. It was a time when MeWe faced backlash over lax content moderation, drawing comparisons between its service and right-wing apps like Gab or Parler. Weinstein rejected those comparisons, seeing his platform as an ideal Facebook rival and remaining a board member through the app’s more recent shift to decentralization. Still defending MeWe’s failed efforts to beat Facebook, he submitted hundreds of documents and was deposed in the monopoly trial, alleging that Meta retaliated against MeWe as a privacy-focused rival that sought to woo users away by branding itself the “anti-Facebook.”

Among his complaints, Weinstein accused Meta of thwarting MeWe’s attempts to introduce interoperability between the two platforms, which he thinks stems from a fear that users might leave Facebook if they discover a more appealing platform. That’s why he’s urged the FTC—if it wins its monopoly case—to go beyond simply ordering a potential breakup of Facebook, Instagram, and WhatsApp to also require interoperability between Meta’s platforms and all rivals. That may be the only way to force Meta to release its clutch on personal data collection, Weinstein suggested, and allow for more competition broadly in the social media industry.

“The glue that holds it all together is Facebook’s monopoly over data,” Weinstein wrote in a Wall Street Journal op-ed, recalling the moment he realized that Meta seemed to have an unbeatable monopoly. “Its ownership and control of the personal information of Facebook users and non-users alike is unmatched.”

Cory Doctorow, a special advisor to the Electronic Frontier Foundation, told Ars that his vision of a better social media future goes even further than requiring interoperability between all platforms. Social networks like Meta’s should also be made to allow reverse engineering so that outside developers can modify their apps with third-party tools without risking legal attacks, he said.

Doctorow said that solution would create “an equilibrium where companies are more incentivized to behave themselves than they are to cheat” by, say, retaliating against, killing off, or buying out rivals. And “if they fail to respond to that incentive and they cheat anyways, then the rest of the world still has a remedy,” Doctorow said, by having the choice to modify or ditch any platform deemed toxic, invasive, manipulative, or otherwise offensive.

Doctorow summed up the frustration that some users have faced through the ongoing “enshittification” of platforms (a term he coined) ever since platforms took over the Internet.

“I’m 55 now, and I’ve gotten a lot less interested in how things work because I’ve had too many experiences with how things fail,” Doctorow told Ars. “And I just want to make sure that if I’m on a service and it goes horribly wrong, I can leave.”

Social media haters wish OG platforms were doomed

Weinstein pointed out that Meta’s alleged monopoly impacts a group often left out of social media debates: non-users. And if you ask someone who hates social media what the future of social media should look like, they will not mince words: They want a way to opt out of all of it.

As Meta’s monopoly trial got underway, a personal blog post titled “No Instagram, no privacy” rose to the front page of Hacker News, prompting a discussion about social media norms and reasonable expectations for privacy in 2025.

In the post, Wouter-Jan Leys, a privacy advocate, explained that he felt “blessed” to have “somehow escaped having an Instagram account,” feeling no pressure to “update the abstract audience of everyone I ever connected with online on where I am, what I am doing, or who I am hanging out with.”

But despite never having an account, he’s found that “you don’t have to be on Instagram to be on Instagram,” complaining that “it bugs me” when friends seem to know “more about my life than I tell them” because of various friends’ posts that mention or show images of him. In his blog, he defined privacy as “being in control of what other people know about you” and suggested that because of platforms like Instagram, he currently lacked this control. There should be some way to “fix or regulate this,” Leys suggested, or maybe some universal “etiquette where it’s frowned upon to post about social gatherings to any audience beyond who already was at that gathering.”

On Hacker News, his post spurred a debate over one of the longest-running privacy questions swirling on social media: Is it OK to post about someone who abstains from social media?

Some seeming social media fans scolded Leys for being so old-fashioned about social media, suggesting, “just live your life without being so bothered about offending other people” or saying that “the entire world doesn’t have to be sanitized to meet individual people’s preferences.” Others seemed to better understand Leys’ point of view, with one agreeing that “the problem is that our modern norms (and tech) lead to everyone sharing everything with a large social network.”

Surveying the lively thread, another social media hater joked, “I feel vindicated for my decision to entirely stay off of this drama machine.”

Leys told Ars that he would “absolutely” be in favor of personal social networks like Meta’s platforms dying off or losing steam, as Zuckerberg suggested they already are. He thinks that the decline in personal post engagement that Meta is seeing is likely due to a combination of factors, where some users may prefer more privacy now after years of broadcasting their lives, and others may be tired of the pressure of building a personal brand or experiencing other “odd social dynamics.”

Setting user sentiments aside, Meta is also responsible for people engaging with fewer of their friends’ posts. Meta announced that it would double the amount of force-fed filler in people’s feeds on Instagram and Facebook starting in 2023. That’s when the two-year span begins that Zuckerberg measured in testifying about the sudden drop-off in friends’ content engagement.

So while it’s easy to say the market changed, Meta may be obscuring how much it shaped that shift. Degrading the newsfeed and changing Instagram’s default post shape from square to rectangle seemingly significantly shifted Instagram social norms, for example, creating an environment where Gen Z users felt less comfortable posting as prolifically as millennials did when Instagram debuted, The New Yorker explained last year. Where once millennials painstakingly designed immaculate grids of individual eye-catching photos to seem cool online, Gen Z users told The New Yorker that posting a single photo now feels “humiliating” and like a “social risk.”

But rather than eliminate the impulse to post, this cultural shift has popularized a different form of personal posting: staggered photo dumps, where users wait to post a variety of photos together to sum up a month of events or curate a vibe, the trend piece explained. And Meta is clearly intent on fueling that momentum, doubling the maximum number of photos that users can feature in a single post to encourage even more social posting, The New Yorker noted.

Brendan Benedict, an attorney for Benedict Law Group PLLC who has helped litigate big tech antitrust cases, is monitoring the FTC monopoly trial on a Substack called Big Tech on Trial. He told Ars that the evidence at the trial has shown that “consumers want more friends and family content, and Meta is belatedly trying to address this” with features like the “friends” tab, while claiming there’s less interest in this content.

Leys doesn’t think social media—at least the way that Facebook defined it in the mid-2000s—will ever die, because people will never stop wanting social networks like Facebook or Instagram to stay connected with all their friends and family. But he could see a world where, if people ever started truly caring about privacy or “indeed [got] tired of the social dynamics and personal brand-building… the kind of social media like Facebook and Instagram will have been a generational phenomenon, and they may not immediately bounce back,” especially if it’s easy to switch to other platforms that respond better to user preferences.

He also agreed that requiring interoperability would likely lead to better social media products, but he maintained that “it would still not get me on Instagram.”

Interoperability shakes up social media

Meta thought it may have already beaten the FTC’s monopoly case, filing for a motion for summary judgment after the FTC rested its case in a bid to end the trial early. That dream was quickly dashed when the judge denied the motion days later. But no matter the outcome of the trial, Meta’s influence over the social media world may be waning just as it’s facing increasing pressure to open up its platforms more than ever.

The FTC has alleged that Meta weaponized platform access early on, only allowing certain companies to interoperate and denying access to anyone perceived as a threat to its alleged monopoly power. That includes limiting promotions of Instagram to keep users engaged with Facebook Blue. A primary concern for Meta (then Facebook), the FTC claimed, was avoiding “training users to check multiple feeds,” which might allow other apps to “cannibalize” its users.

“Facebook has used this power to deter and suppress competitive threats to its personal social networking monopoly. In order to protect its monopoly, Facebook adopted and required developers to agree to conditional dealing policies that limited third-party apps’ ability to engage with Facebook rivals or to develop into rivals themselves,” the FTC alleged.

By 2011, the FTC alleged, then-Facebook had begun terminating API access to any developers that made it easier to export user data into a competing social network without Facebook’s permission. That practice only ended when the UK parliament started calling out Facebook’s anticompetitive conduct toward app developers in 2018, the FTC alleged.

According to the FTC, Meta continues “to this day” to “screen developers and can weaponize API access in ways that cement its dominance,” and if scrutiny ever subsides, Meta is expected to return to such anticompetitive practices as the AI race heats up.

One potential hurdle for Meta could be that the push for interoperability is not just coming from the FTC or lawmakers who recently reintroduced bipartisan legislation to end walled gardens. Doctorow told Ars that “huge public groundswells of mistrust and anger about excessive corporate power” that “cross political lines” are prompting global antitrust probes into big tech companies and are perhaps finally forcing a reckoning after years of degrading popular products to chase higher and higher revenues.

For social media companies, mounting concerns about privacy and suspicions about content manipulation or censorship are driving public distrust, Doctorow said, as well as fears of surveillance capitalism. The latter includes theories that Doctorow is skeptical of. Weinstein embraced them, though, warning that platforms seem to be profiting off data without consent while brainwashing users.

Allowing users to leave the platform without losing access to their friends, their social posts, and their messages might be the best way to incentivize Meta to either genuinely compete for billions of users or lose them forever as better options pop up that can plug into their networks.

In his Newsweek op-ed, Weinstein suggested that web inventor Tim Berners-Lee has already invented a working protocol “to enable people to own, upload, download, and relocate their social graphs,” which maps users’ connections across platforms. That could be used to mitigate “the network effect” that locks users into platforms like Meta’s “while interrupting unwanted data collection.”

At the same time, Doctorow told Ars that increasingly popular decentralized platforms like Bluesky and Mastodon already provide interoperability and are next looking into “building interoperable gateways” between their services. Doctorow said that communicating with other users across platforms may feel “awkward” at first, but ultimately, it may be like “having to find the diesel pump at the gas station” instead of the unleaded gas pump. “You’ll still be going to the same gas station,” Doctorow suggested.

Opening up gateways into all platforms could be useful in the future, Doctorow suggested. Imagine if one platform goes down—it would no longer disrupt communications as drastically, as users could just pivot to communicate on another platform and reach the same audience. The same goes for platforms that users grow to distrust.

The EFF supports regulators’ attempts to pass well-crafted interoperability mandates, Doctorow said, noting that “if you have to worry about your users leaving, you generally have to treat them better.”

But would interoperability fix social media?

The FTC has alleged that “Facebook’s dominant position in the US personal social networking market is durable due to significant entry barriers, including direct network effects and high switching costs.”

Meta disputes the FTC’s complaint as outdated, arguing that its platform could be substituted by pretty much any social network.

However, Guy Aridor, a co-author of a recent article called “The Economics of Social Media” in the Journal of Economic Literature, told Ars that dominant platforms are probably threatened by shifting social media trends and are likely to remain “resistant to interoperability” because “it’s in the interest of the platform to make switching and coordination costs high so that users are less likely to migrate away.” For Meta, research shows its platforms’ network effects have appeared to weaken somewhat but “clearly still exist” despite social media users increasingly seeking content on platforms rather than just socialization, Aridor said.

Interoperability advocates believe it will make it easier for startups to compete with giants like Meta, which fight hard and sometimes seemingly dirty to keep users on their apps. Reintroducing the ACCESS Act, which requires platform compatibility to enable service switching, Senator Mark R. Warner (D-Va.) said that “interoperability and portability are powerful tools to promote innovative new companies and limit anti-competitive behaviors.” He’s hoping that passing these “long-overdue requirements” will “boost competition and give consumers more power.”

Aridor told Ars it’s obvious that “interoperability would clearly increase competition,” but he still has questions about whether users would benefit from that competition “since one consistent theme is that these platforms are optimized to maximize engagement, and there’s numerous empirical evidence we have by now that engagement isn’t necessarily correlated with utility.”

Consider, Aridor suggested, how toxic content often leads to high engagement but lower user satisfaction, as MeWe experienced during its 2021 backlash.

Aridor said there is currently “very little empirical evidence on the effects of interoperability,” but theoretically, if it increased competition in the current climate, it would likely “push the market more toward supplying engaging entertainment-related content as opposed to friends and family type of content.”

Benedict told Ars that a remedy like interoperability would likely only be useful to combat Meta’s alleged monopoly following a breakup, which he views as the “natural remedy” following a potential win in the FTC’s lawsuit.

Without the breakup and other meaningful reforms, a Meta win could preserve the status quo and see the company never open up its platforms, perhaps perpetuating Meta’s influence over social media well into the future. And if Zuckerberg’s vision comes to pass, instead of seeing what your friends are posting on interoperating platforms across the Internet, you may have a dozen AI friends trained on your real friends’ behaviors sending you regular dopamine hits to keep you scrolling on Facebook or Instagram.

Aridor’s team’s article suggested that, regardless of user preferences, social media remains a permanent fixture of society. If that’s true, users could get stuck forever using whichever platforms connect them with the widest range of contacts.

“While social media has continued to evolve, one thing that has not changed is that social media remains a central part of people’s lives,” his team’s article concluded.

Photo of Ashley Belanger

Ashley is a senior policy reporter for Ars Technica, dedicated to tracking social impacts of emerging policies and new technologies. She is a Chicago-based journalist with 20 years of experience.

Meta hypes AI friends as social media’s future, but users want real connections Read More »

meta-argues-enshittification-isn’t-real-in-bid-to-toss-ftc-monopoly-trial

Meta argues enshittification isn’t real in bid to toss FTC monopoly trial

Further, Meta argued that the FTC did not show evidence that users sharing friends-and-family content were shown more ads. Meta noted that it “does not profit by showing more ads to users who do not click on them,” so it only shows more ads to users who click ads.

Meta also insisted that there’s “nothing but speculation” showing that Instagram or WhatsApp would have been better off or grown into rivals had Meta not acquired them.

The company claimed that without Meta’s resources, Instagram may have died off. Meta noted that Instagram co-founder Kevin Systrom testified that his app was “pretty broken and duct-taped” together, making it “vulnerable to spam” before Meta bought it.

Rather than enshittification, what Meta did to Instagram could be considered “a consumer-welfare bonanza,” Meta argued, while dismissing “smoking gun” emails from Mark Zuckerberg discussing buying Instagram to bury it as “legally irrelevant.”

Dismissing these as “a few dated emails,” Meta argued that “efforts to litigate Mr. Zuckerberg’s state of mind before the acquisition in 2012 are pointless.”

“What matters is what Meta did,” Meta argued, which was pump Instagram with resources that allowed it “to ‘thrive’—adding many new features, attracting hundreds of millions and then billions of users, and monetizing with great success.”

In the case of WhatsApp, Meta argued that nobody thinks WhatsApp had any intention to pivot to social media when the founders testified that their goal was to never add social features, preferring to offer a simple, clean messaging app. And Meta disputed any claim that it feared Google might buy WhatsApp as the basis for creating a Facebook rival, arguing that “the sole Meta witness to (supposedly) learn of Google’s acquisition efforts testified that he did not have that worry.”

Meta argues enshittification isn’t real in bid to toss FTC monopoly trial Read More »

kids-are-short-circuiting-their-school-issued-chromebooks-for-tiktok-clout

Kids are short-circuiting their school-issued Chromebooks for TikTok clout

Schools across the US are warning parents about an Internet trend that has students purposefully trying to damage their school-issued Chromebooks so that they start smoking or catch fire.

Various school districts, including some in Colorado, New Jersey, North Carolina, and Washington, have sent letters to parents warning about the trend that’s largely taken off on TikTok.

Per reports from school districts and videos that Ars Technica has reviewed online, the so-called Chromebook Challenge includes students sticking things into Chromebook ports to short-circuit the system. Students are using various easily accessible items to do this, including writing utensils, paper clips, gum wrappers, and pushpins.

The Chromebook challenge has caused chaos for US schools, leading to laptop fires that have forced school evacuations, early dismissals, and the summoning of first responders.

Schools are also warning that damage to school property can result in disciplinary action and, in some states, legal action.

In Plainville, Connecticut, a middle schooler allegedly “intentionally stuck scissors into a laptop, causing smoke to emit from it,” Superintendent Brian Reas told local news station WFSB. The incident reportedly led to one student going to the hospital due to smoke inhalation and is suspected to be connected to the viral trend.

“Although the investigation is ongoing, the student involved will be referred to juvenile court to face criminal charges,” Reas said.

Kids are short-circuiting their school-issued Chromebooks for TikTok clout Read More »

disgruntled-users-roast-x-for-killing-support-account

Disgruntled users roast X for killing Support account

After X (formerly Twitter) announced it would be killing its “Support” account, disgruntled users quickly roasted the social media platform for providing “essentially non-existent” support.

“We’ll soon be closing this account to streamline how users can contact us for help,” X’s Support account posted, explaining that now, paid “subscribers can get support via @Premium, and everyone can get help through our Help Center.”

On X, the Support account was one of the few paths that users had to publicly seek support for help requests the platform seemed to be ignoring. For suspended users, it was viewed as a lifeline. Replies to the account were commonly flooded with users trying to get X to fix reported issues, and several seemingly paying users cracked jokes in response to the news that the account would soon be removed.

“Lololol your support for Premium is essentially non-existent,” a subscriber with more than 200,000 followers wrote, while another quipped “Okay, so no more support? lol.”

On Reddit, X users recently suggested that contacting the Premium account is the only way to get human assistance after briefly interacting with a bot. But some self-described Premium users complained of waiting six months or longer for responses from X’s help center in the Support thread.

Some users who don’t pay for access to the platform similarly complained. But for paid subscribers or content creators, lack of Premium support is perhaps most frustrating, as one user claimed their account had been under review for years, allegedly depriving them of revenue. And another user claimed they’d had “no luck getting @Premium to look into” an account suspension while supposedly still getting charged. Several accused X of sending users into a never-ending loop, where the help center only serves to link users to the help center.

Disgruntled users roast X for killing Support account Read More »

x’s-globe-trotting-defense-of-ads-on-nazi-posts-violates-tos,-media-matters-says

X’s globe-trotting defense of ads on Nazi posts violates TOS, Media Matters says

Part of the problem appeared to be decreased spending from big brands that did return, like reportedly Apple. Other dips were linked to X’s decision to partner with adtech companies, splitting ad revenue with Magnite, Google, and PubMatic, Business Insider reported. The CEO of marketing consultancy Ebiquity, Ruben Schreurs, told Business Insider that most of the top 100 global advertisers he works with were still hesitant to invest in X, confirming “no signs of a mass return.”

For X, the ad boycott has tanked revenue for years, even putting X on the brink of bankruptcy, Musk claimed. The billionaire paid $44 billion for the platform, and at the end of 2024, Fidelity estimated that X was worth just $9.4 billion, CNN reported.

But at the start of 2025, analysts predicted that advertisers may return to X to garner political favor with Musk, who remains a senior advisor in the Trump administration. Perhaps more importantly in the short-term, sources also told Bloomberg that X could potentially raise as much as Musk paid—$44 billion—from investors willing to help X pay down its debt to support new payments and video products.

That could put a Band-Aid on X’s financial wounds as Yaccarino attempts to persuade major brands that X isn’t toxic (while X sues some of them) and Musk tries to turn the social media platform once known as Twitter into an “everything app” as ubiquitous in the US as WeChat in China.

MMFA alleges that its research, which shows how toxic X is today, has been stifled by Musk’s suits, but other groups have filled the gap. The Center for Countering Digital Hate has resumed its reporting since defeating X’s lawsuit last March, and, most recently, University of California, Berkeley, researchers conducted a February analysis showing that “hate speech on the social media platform X rose about 50 percent” in the eight months after Musk’s 2022 purchase, which suggests that advertisers had potentially good reason to be spooked by changes at X and that those changes continue to keep them at bay today.

“Musk has continually tried to blame others for this loss in revenue since his takeover,” MMFA’s complaint said, alleging that all three suits were filed to intimidate MMFA “for having dared to publish an article Musk did not like.”

X’s globe-trotting defense of ads on Nazi posts violates TOS, Media Matters says Read More »

elon-musk-blames-x-outages-on-“massive-cyberattack”

Elon Musk blames X outages on “massive cyberattack”

After DownDetector reported that tens of thousands of users globally experienced repeated X (formerly Twitter) outages, Elon Musk confirmed the issues are due to an ongoing cyberattack on the platform.

“There was (still is) a massive cyberattack against X,” Musk wrote on X. “We get attacked every day, but this was done with a lot of resources. Either a large, coordinated group and/or a country is involved.”

Details remain vague beyond Musk’s post, but rumors were circulating that X was under a distributed denial-of-service (DDOS) attack.

X’s official support channel, which has been dormant since August, has so far remained silent on the outage, but one user asked Grok—X’s chatbot that provides AI summaries of news—what was going on, and the chatbot echoed suspicions about the DDOS attack while raising other theories.

“Over 40,000 users reported issues, with the platform struggling to load globally,” Grok said. “No clear motive yet, but some speculate it’s political since X is the only target. Outages hit hard in the US, Switzerland, and beyond.”

As X goes down, users cry for Twitter

It has been almost two years since Elon Musk declared that Twitter “no longer exists,” haphazardly rushing to rebrand his social media company as X despite critics warning that users wouldn’t easily abandon the Twitter brand.

Fast-forward to today, and Musk got a reminder that his efforts to kill off the Twitter brand never really caught on with a large chunk of his platform.

Elon Musk blames X outages on “massive cyberattack” Read More »

reddit-mods-are-fighting-to-keep-ai-slop-off-subreddits-they-could-use-help.

Reddit mods are fighting to keep AI slop off subreddits. They could use help.


Mods ask Reddit for tools as generative AI gets more popular and inconspicuous.

Redditors in a treehouse with a NO AI ALLOWED sign

Credit: Aurich Lawson (based on a still from Getty Images)

Credit: Aurich Lawson (based on a still from Getty Images)

Like it or not, generative AI is carving out its place in the world. And some Reddit users are definitely in the “don’t like it” category. While some subreddits openly welcome AI-generated images, videos, and text, others have responded to the growing trend by banning most or all posts made with the technology.

To better understand the reasoning and obstacles associated with these bans, Ars Technica spoke with moderators of subreddits that totally or partially ban generative AI. Almost all these volunteers described moderating against generative AI as a time-consuming challenge they expect to get more difficult as time goes on. And most are hoping that Reddit will release a tool to help their efforts.

It’s hard to know how much AI-generated content is actually on Reddit, and getting an estimate would be a large undertaking. Image library Freepik has analyzed the use of AI-generated content on social media but leaves Reddit out of its research because “it would take loads of time to manually comb through thousands of threads within the platform,” spokesperson Bella Valentini told me. For its part, Reddit doesn’t publicly disclose how many Reddit posts involve generative AI use.

To be clear, we’re not suggesting that Reddit has a large problem with generative AI use. By now, many subreddits seem to have agreed on their approach to AI-generated posts, and generative AI has not superseded the real, human voices that have made Reddit popular.

Still, mods largely agree that generative AI will likely get more popular on Reddit over the next few years, making generative AI modding increasingly important to both moderators and general users. Generative AI’s rising popularity has also had implications for Reddit the company, which in 2024 started licensing Reddit posts to train the large language models (LLMs) powering generative AI.

(Note: All the moderators I spoke with for this story requested that I use their Reddit usernames instead of their real names due to privacy concerns.)

No generative AI allowed

When it comes to anti-generative AI rules, numerous subreddits have zero-tolerance policies, while others permit posts that use generative AI if it’s combined with human elements or is executed very well. These rules task mods with identifying posts using generative AI and determining if they fit the criteria to be permitted on the subreddit.

Many subreddits have rules against posts made with generative AI because their mod teams or members consider such posts “low effort” or believe AI is counterintuitive to the subreddit’s mission of providing real human expertise and creations.

“At a basic level, generative AI removes the human element from the Internet; if we allowed it, then it would undermine the very point of r/AskHistorians, which is engagement with experts,” the mods of r/AskHistorians told me in a collective statement.

The subreddit’s goal is to provide historical information, and its mods think generative AI could make information shared on the subreddit less accurate. “[Generative AI] is likely to hallucinate facts, generate non-existent references, or otherwise provide misleading content,” the mods said. “Someone getting answers from an LLM can’t respond to follow-ups because they aren’t an expert. We have built a reputation as a reliable source of historical information, and the use of [generative AI], especially without oversight, puts that at risk.”

Similarly, Halaku, a mod of r/wheeloftime, told me that the subreddit’s mods banned generative AI because “we focus on genuine discussion.” Halaku believes AI content can’t facilitate “organic, genuine discussion” and “can drown out actual artwork being done by actual artists.”

The r/lego subreddit banned AI-generated art because it caused confusion in online fan communities and retail stores selling Lego products, r/lego mod Mescad said. “People would see AI-generated art that looked like Lego on [I]nstagram or [F]acebook and then go into the store to ask to buy it,” they explained. “We decided that our community’s dedication to authentic Lego products doesn’t include AI-generated art.”

Not all of Reddit is against generative AI, of course. Subreddits dedicated to the technology exist, and some general subreddits permit the use of generative AI in some or all forms.

“When it comes to bans, I would rather focus on hate speech, Nazi salutes, and things that actually harm the subreddits,” said 3rdusernameiveused, who moderates r/consoom and r/TeamBuilder25, which don’t ban generative AI. “AI art does not do that… If I was going to ban [something] for ‘moral’ reasons, it probably won’t be AI art.”

“Overwhelmingly low-effort slop”

Some generative AI bans are reflective of concerns that people are not being properly compensated for the content they create, which is then fed into LLM training.

Mod Mathgeek007 told me that r/DeadlockTheGame bans generative AI because its members consider it “a form of uncredited theft,” adding:

You aren’t allowed to sell/advertise the workers of others, and AI in a sense is using patterns derived from the work of others to create mockeries. I’d personally have less of an issue with it if the artists involved were credited and compensated—and there are some niche AI tools that do this.

Other moderators simply think generative AI reduces the quality of a subreddit’s content.

“It often just doesn’t look good… the art can often look subpar,” Mathgeek007 said.

Similarly, r/videos bans most AI-generated content because, according to its announcement, the videos are “annoying” and “just bad video” 99 percent of the time. In an online interview, r/videos mod Abrownn told me:

It’s overwhelmingly low-effort slop thrown together simply for views/ad revenue. The creators rarely care enough to put real effort into post-generation [or] editing of the content [and] rarely have coherent narratives [in] the videos, etc. It seems like they just throw the generated content into a video, export it, and call it a day.

An r/fakemon mod told me, “I can’t think of anything more low-effort in terms of art creation than just typing words and having it generated for you.”

Some moderators say generative AI helps people spam unwanted content on a subreddit, including posts that are irrelevant to the subreddit and posts that attack users.

“[Generative AI] content is almost entirely posted for purely self promotional/monetary reasons, and we as mods on Reddit are constantly dealing with abusive users just spamming their content without regard for the rules,” Abrownn said.

A moderator of the r/wallpaper subreddit, which permits generative AI, disagrees. The mod told me that generative AI “provides new routes for novel content” in the subreddit and questioned concerns about generative AI stealing from human artists or offering lower-quality work, saying those problems aren’t unique to generative AI:

Even in our community, we observe human-generated content that is subjectively low quality (poor camera/[P]hotoshopping skills, low-resolution source material, intentional “shitposting”). It can be argued that AI-generated content amplifies this behavior, but our experience (which we haven’t quantified) is that the rate of such behavior (whether human-generated or AI-generated content) has not changed much within our own community.

But we’re not a very active community—[about] 13 posts per day … so it very well could be a “frog in boiling water” situation.

Generative AI “wastes our time”

Many mods are confident in their ability to effectively identify posts that use generative AI. A bigger problem is how much time it takes to identify these posts and remove them.

The r/AskHistorians mods, for example, noted that all bans on the subreddit (including bans unrelated to AI) have “an appeals process,” and “making these assessments and reviewing AI appeals means we’re spending a considerable amount of time on something we didn’t have to worry about a few years ago.”

They added:

Frankly, the biggest challenge with [generative AI] usage is that it wastes our time. The time spent evaluating responses for AI use, responding to AI evangelists who try to flood our subreddit with inaccurate slop and then argue with us in modmail, [direct messages that message a subreddits’ mod team], and discussing edge cases could better be spent on other subreddit projects, like our podcast, newsletter, and AMAs, … providing feedback to users, or moderating input from users who intend to positively contribute to the community.

Several other mods I spoke with agree. Mathgeek007, for example, named “fighting AI bros” as a common obstacle. And for r/wheeloftime moderator Halaku, the biggest challenge in moderating against generative AI is “a generational one.”

“Some of the current generation don’t have a problem with it being AI because content is content, and [they think] we’re being elitist by arguing otherwise, and they want to argue about it,” they said.

A couple of mods noted that it’s less time-consuming to moderate subreddits that ban generative AI than it is to moderate those that allow posts using generative AI, depending on the context.

“On subreddits where we allowed AI, I often take a bit longer time to actually go into each post where I feel like… it’s been AI-generated to actually look at it and make a decision,” explained N3DSdude, a mod of several subreddits with rules against generative AI, including r/DeadlockTheGame.

MyarinTime, a moderator for r/lewdgames, which allows generative AI images, highlighted the challenges of identifying human-prompted generative AI content versus AI-generated content prompted by a bot:

When the AI bomb started, most of those bots started using AI content to work around our filters. Most of those bots started showing some random AI render, so it looks like you’re actually talking about a game when you’re not. There’s no way to know when those posts are legit games unless [you check] them one by one. I honestly believe it would be easier if we kick any post with [AI-]generated image… instead of checking if a button was pressed by a human or not.

Mods expect things to get worse

Most mods told me it’s pretty easy for them to detect posts made with generative AI, pointing to the distinct tone and favored phrases of AI-generated text. A few said that AI-generated video is harder to spot but still detectable. But as generative AI gets more advanced, moderators are expecting their work to get harder.

In a joint statement, r/dune mods Blue_Three and Herbalhippie said, “AI used to have a problem making hands—i.e., too many fingers, etc.—but as time goes on, this is less and less of an issue.”

R/videos’ Abrownn also wonders how easy it will be to detect AI-generated Reddit content “as AI tools advance and content becomes more lifelike.”

Mathgeek007 added:

AI is becoming tougher to spot and is being propagated at a larger rate. When AI style becomes normalized, it becomes tougher to fight. I expect generative AI to get significantly worse—until it becomes indistinguishable from ordinary art.

Moderators currently use various methods to fight generative AI, but they’re not perfect. r/AskHistorians mods, for example, use “AI detectors, which are unreliable, problematic, and sometimes require paid subscriptions, as well as our own ability to detect AI through experience and expertise,” while N3DSdude pointed to tools like Quid and GPTZero.

To manage current and future work around blocking generative AI, most of the mods I spoke with said they’d like Reddit to release a proprietary tool to help them.

“I’ve yet to see a reliable tool that can detect AI-generated video content,” Aabrown said. “Even if we did have such a tool, we’d be putting hundreds of hours of content through the tool daily, which would get rather expensive rather quickly. And we’re unpaid volunteer moderators, so we will be outgunned shortly when it comes to detecting this type of content at scale. We can only hope that Reddit will offer us a tool at some point in the near future that can help deal with this issue.”

A Reddit spokesperson told me that the company is evaluating what such a tool could look like. But Reddit doesn’t have a rule banning generative AI overall, and the spokesperson said the company doesn’t want to release a tool that would hinder expression or creativity.

For now, Reddit seems content to rely on moderators to remove AI-generated content when appropriate. Reddit’s spokesperson added:

Our moderation approach helps ensure that content on Reddit is curated by real humans. Moderators are quick to remove content that doesn’t follow community rules, including harmful or irrelevant AI-generated content—we don’t see this changing in the near future.

Making a generative AI Reddit tool wouldn’t be easy

Reddit is handling the evolving concerns around generative AI as it has handled other content issues, including by leveraging AI and machine learning tools. Reddit’s spokesperson said that this includes testing tools that can identify AI-generated media, such as images of politicians.

But making a proprietary tool that allows moderators to detect AI-generated posts won’t be easy, if it happens at all. The current tools for detecting generative AI are limited in their capabilities, and as generative AI advances, Reddit would need to provide tools that are more advanced than the AI-detecting tools that are currently available.

That would require a good deal of technical resources and would also likely present notable economic challenges for the social media platform, which only became profitable last year. And as noted by r/videos moderator Abrownn, tools for detecting AI-generated video still have a long way to go, making a Reddit-specific system especially challenging to create.

But even with a hypothetical Reddit tool, moderators would still have their work cut out for them. And because Reddit’s popularity is largely due to its content from real humans, that work is important.

Since Reddit’s inception, that has meant relying on moderators, which Reddit has said it intends to keep doing. As r/dune mods Blue_Three and herbalhippie put it, it’s in Reddit’s “best interest that much/most content remains organic in nature.” After all, Reddit’s profitability has a lot to do with how much AI companies are willing to pay to access Reddit data. That value would likely decline if Reddit posts became largely AI-generated themselves.

But providing the technology to ensure that generative AI isn’t abused on Reddit would be a large challege. For now, volunteer laborers will continue to bear the brunt of generative AI moderation.

Advance Publications, which owns Ars Technica parent Condé Nast, is the largest shareholder of Reddit.

Photo of Scharon Harding

Scharon is a Senior Technology Reporter at Ars Technica writing news, reviews, and analysis on consumer gadgets and services. She’s been reporting on technology for over 10 years, with bylines at Tom’s Hardware, Channelnomics, and CRN UK.

Reddit mods are fighting to keep AI slop off subreddits. They could use help. Read More »