Author name: Shannon Garcia

twitch-makes-deal-to-escape-elon-musk-suit-alleging-x-ad-boycott-conspiracy

Twitch makes deal to escape Elon Musk suit alleging X ad boycott conspiracy

Instead, it appears that X decided to sue Twitch after discovering that Twitch was among advertisers who directly referenced the WFA’s brand safety guidelines in its own community guidelines and terms of service. X likely saw this as evidence that Twitch was allegedly conspiring with the WFA to restrict then-Twitter’s ad revenue, since X alleged that Twitch reduced ad purchases to “only a de minimis amount outside the United States, after November 2022,” X’s complaint said.

“The Advertiser Defendants and other GARM-member advertisers acted in parallel to discontinue their purchases of advertising from Twitter, in a marked departure from their prior pattern of purchases,” X’s complaint said.

Now, it seems that X has agreed to drop Twitch from the suit, perhaps partly because the complaint X had about Twitch adhering to WFA brand safety standards is defused since the WFA disbanded the ad industry arm that set those standards.

Unilever struck a similar deal to wriggle out of the litigation, Reuters noted, and remained similarly quiet on the terms, only saying that the brand remained “committed to meeting our responsibility standards to ensure the safety and performance of our brands on the platform.” But other advertisers, including Colgate, CVS, LEGO, Mars, Pinterest, Shell, and Tyson Foods, so far have not.

For Twitch, its deal seems to clearly take a target off its back at a time when some advertisers are reportedly returning to X to stay out of Musk’s crosshairs. Getting out now could spare substantial costs as the lawsuit drags on, even though X CEO Linda Yaccarino declared the ad boycott was over in January. X is still $12 billion in debt, X claimed, after Musk’s xAI bought X last month. External data in January seemed to suggest many big brands were still hesitant to return to the platform, despite Musk’s apparent legal strong-arming and political influence in the Trump administration.

Ars could not immediately reach Twitch or X for comment. But the court docket showed that Twitch was up against a deadline to respond to the lawsuit by mid-May, which likely increased pressure to reach an agreement before Twitch was forced to invest in raising a defense.

Twitch makes deal to escape Elon Musk suit alleging X ad boycott conspiracy Read More »

don’t-call-it-a-drone:-zipline’s-uncrewed-aircraft-wants-to-reinvent-retail

Don’t call it a drone: Zipline’s uncrewed aircraft wants to reinvent retail


Ars visits a zipline delivery service that’s deploying in more locations soon.

The inner portion of the Zipline P2 is lowered to the ground on a tether, facing into the wind, with a small propeller at the back. Doors on the bottom open when it touches the ground, depositing the cargo. Credit: Tim Stevens

The skies around Dallas are about to get a lot more interesting. No, DFW airport isn’t planning any more expansions, nor does American Airlines have any more retro liveries to debut. This will be something different, something liable to make all the excitement around the supposed New Jersey drones look a bit quaint.

Zipline is launching its airborne delivery service for real, rolling it out in the Dallas-Fort Worth suburb of Mesquite ahead of a gradual spread that, if all goes according to plan, will also see its craft landing in Seattle before the end of the year. These automated drones can be loaded in seconds, carry small packages for miles, and deposit them with pinpoint accuracy at the end of a retractable tether.

It looks and sounds like the future, but this launch has been a decade in the making. Zipline has already flown more than 1.4 million deliveries and covered over 100 million miles, yet it feels like things are just getting started.

The ranch

When Zipline called me and invited me out for a tour of a drone delivery testing facility hidden in the hills north of San Francisco, I was naturally intrigued, but I had no idea what to expect. Shipping logistics facilities tend to be dark and dreary spaces, with automated machinery stacked high on angular shelves within massive buildings presenting all the visual charm of a concrete paver.

Zipline’s facility is a bit different. It’s utterly stunning, situated among the pastures of a ranch that sprawls over nearly 7,000 acres of the kind of verdant, rolling terrain that has drawn nature lovers to Northern California for centuries.

A modest-looking facility amidst beautiful hills

The Zipline drone testing facility. Credit: Tim Stevens

Zipline’s contribution to the landscape consists of a few shipping container-sized prefab office spaces, a series of tents, and some tall, metal structures that look like a stand of wireform trees. The fruit hanging from their aluminum branches are clusters of white drones, or at least what we’d call “drones.”

But the folks at Zipline don’t seem to like that term. Everyone I spoke with referred to the various craft hovering, buzzing, or gliding overhead as aircraft. That’s for good reason.

Not your average UAV

Go buy a drone at an electronics retailer, something from DJI perhaps, and you’ll have to abide by a series of regulations about how high and how far to fly it. Two of the most important rules: Never fly near an airport, and never let the thing out of your sight.

Zipline’s aircraft are much more comprehensive machines, able to fly for miles and miles. By necessity, they must fly well beyond the range of any human operator, or what’s called “beyond visual line of sight,” or BVLOS. In 2023, Zipline was the first commercial operator to get clearance for BVLOS flights.

Zipline’s aircraft operate under a series of FAA classifications—specifically, part 107, part 135, and the upcoming part 108, which will formalize BVLOS operation. The uncrewed aircraft, which are able to operate as such, navigate through controlled airspace, and even near airports, with the help of FAA-mandated transponder data as well as onboard sensors that can detect the presence of an approaching aircraft and automatically avoid it.

A tree-like tower houses a drone with rolling hills as the backdrop

A Zipline drone testing facility. Seen on the right is one of the “trees.” Credit: Tim Stevens

In fact, just about everything about Zipline’s aircraft is automatic. Onboard sensors sample the air through pitot tubes, detecting bad weather. The craft use this data to reroute themselves around the problem, then report back to save subsequent flights the hassle.

Wind speed and direction are also calculated, ensuring that deliveries are dropped with accuracy. Once the things are in the air, even the Zipline operators aren’t sure which way they’ll fly, only that they’ll figure out the right way to get the package there and return safely.

Zipline actually operates two separate aircraft that are suited for different mission types. The aircraft clinging to the aluminum trees, the type that will be exploring the skies over Dallas soon, are internally called Platform 2, or P2, and they’re actually two aircraft in one.

A P2 drone can hover in place using five propellers and take off vertically before seamlessly transitioning into efficient forward flight. When it reaches its destination, doors on the bottom open, and a second aircraft emerges. This baby craft, called a “Zip,” drops down on a tether.

Fins ensure the tethered craft stays facing into the wind while a small propeller at the rear keeps it from blowing off-target. When it touches the ground, its doors pop open, gently depositing a package from a cargo cavity that’s big enough for about four loaves of bread. Maximum payload capacity is eight pounds, and payloads can be delivered up to about 10 miles away.

Where there’s a P2, there must be a P1, and while Zipline’s first aircraft serves much the same purpose, it does so in a very different way. The P1 is a fixed-wing aircraft, looking for all the world like a hobbyist’s radio-controlled model, just bigger and way more expensive.

The P1 launches into the sky like a glider, courtesy of a high-torque winch that slings it aloft before its electric prop takes over. It can fly for over 120 miles on a charge before dropping its cargo, a package that glides to the ground via parachute.

The P1 slows momentarily during the drop and then buzzes back up to full speed dramatically before turning for home. There’s no gentle, vertical landing here. It instead cruises precisely toward a wire suspended high in the air. An instant before impact, it noses up, exposing a metal hook to the wire, which stops the thing instantly.

In naval aviator parlance, it’s an OK three-wire every time, and thanks to hot-swappable batteries, a P1 can be back in the air in just minutes. This feature has helped the company perform millions of successful deliveries, many carrying lifesaving supplies.

From Muhanga to Mesquite

The first deployment from the company that would become Zipline was in 2016 in Muhanga, Rwanda, beginning with the goal of delivering vaccines and other medical supplies quickly and reliably across the untamed expanses of Africa. Eric Watson, now head of systems and safety engineering at Zipline, was part of that initial crew.

“Our mission is to enable access to instant logistics to everyone in the world,” he said. “We started with one of the most visceral pain points, of being able to go to a place, operating in remote parts where access to medicine was a problem.”

It proved to be an incredible proving ground for the technology, but this wasn’t just some beta test designed to deliver greater ROI. Zipline already has success in a more important area: delivering lifesaving medicine. The company’s drones deliver things like vaccines, anti-venoms, and plasma. A 2023 study from the Wharton School at the University of Pennsylvania found that Zipline’s blood delivery service reduced deaths from postpartum hemorrhage by 51 percent.

That sort of promise attracted Lauren Lacey to the company. She’s Zipline’s head of integration quality and manufacturing engineering. A former engineer at Sandia Labs, where she spent a decade hardening America’s military assets, Lacey has brought that expertise to whipping Zipline’s aircraft into shape.

A woman stands by a drone in a testing facility

Lauren Lacey, Zipline’s head of integration quality and manufacturing engineering. Credit: Tim Stevens

Lacey walked me through the 11,000-square-foot Bay Area facility she and her team have turned into a stress-testing house of horrors for uncrewed aircraft. I witnessed everything from latches being subjected to 120° F heat while bathed in ultra-fine dust to a giant magnetic resonance device capable of rattling a circuit board with 70 Gs of force.

It’s all in the pursuit of creating an aircraft that can survive 10,000 deliveries. The various test chambers can replicate upward of 2,500 tests per day, helping the Zipline team iterate quickly and not only add strength but peel away unneeded mass, too.

“Every single gram that we put on the aircraft is one less that we can deliver to the customer,” Lacey said.

Now zipping

Zipline already has a small test presence in Arkansas, a pilot program with Walmart, but its rollout today is a big step forward. Once added to the system, customers can make orders through a dedicated Zipline app. Walmart is the only partner for now, but the company plans to offer more products on the retail and healthcare front, including restaurant food deliveries.

The app will show Walmart products eligible for this sort of delivery, calculating weight and volume to ensure that your order isn’t too big. The P2’s eight-pound payload may seem restrictive, but Jeff Bezos, in touting Amazon’s own future drone delivery program, previously said that 86 percent of the company’s deliveries are five pounds or less.

Amazon suspended its prototype drone program last year for software updates but is flying again in pilot programs in Texas and Arizona. The company has not provided an update on the number of flights lately, but the most recent figures were fewer than 10,000 drone deliveries. For comparison, Zipline currently completes thousands per day. Another future competitor, Alphabet-backed Wing, has flown nearly a half-million deliveries in the US and abroad.

Others are vying for a piece of the airborne delivery pie, too, but nobody I spoke with at Zipline seems worried. From what I could see from my visit, they have reason for confidence. The winds on that ranch in California were so strong that towering dust devils were dancing between the disaffected cattle during my visit. Despite that, the drones flew fast and true, and my requested delivery of bandages and medicine was safely and quickly deposited on the ground just a few feet from my own feet.

It felt like magic, yes, but more importantly, it was one of the most disruptive demonstrations I’ve seen. While the tech isn’t ideally suited for every situation, it may help cut down on the delivery trucks that are increasingly clogging rural roads, all while getting more things to more people who need them, and doing it emissions-free.

Don’t call it a drone: Zipline’s uncrewed aircraft wants to reinvent retail Read More »

freedos-1.4-brings-new-fixes-and-features-to-modern-and-vintage-dos-based-pcs

FreeDOS 1.4 brings new fixes and features to modern and vintage DOS-based PCs

We’re used to updating Windows, macOS, and Linux systems at least once a month (and usually more), but people with ancient DOS-based PCs still get to join in the fun every once in a while. Over the weekend, the team that maintains FreeDOS officially released version 1.4 of the operating system, containing a list of fixes and updates that have been in the works since the last time a stable update was released in 2022.

FreeDOS creator and maintainer Jim Hall goes into more detail about the FreeDOS 1.4 changes here, and full release notes for the changes are here. The release has “a focus on stability” and includes an updated installer, new versions of common tools like fdisk, and format and the edlin text editor. The release also includes updated HTML Help files.

Hall talked with Ars about several of these changes when we interviewed him about FreeDOS in 2024. The team issued the first release candidate for FreeDOS 1.4 back in January.

As with older versions, the FreeDOS installer is available in multiple formats based on the kind of system you’re installing it on. For any “modern” PC (where “modern” covers anything that’s shipped since the turn of the millennium), ISO and USB installers are available for creating bootable CDs, DVDs, or USB drives. FreeDOS is also available for vintage systems as a completely separate “Floppy-Only Edition” that fits on 720KB, 1.44MB, or 1.2MB 5.25 and 3.5-inch floppy disks. This edition “contains a limited set of FreeDOS programs that are more useful on classic PC hardware” and, to conserve space, does not include any FreeDOS source code.

The standard install image includes all the files and utilities you need for a working FreeDOS install, and a separate “BonusCD” download is also available for those who want development tools, the OpenGEM graphical interface, and other tools.

FreeDOS 1.4 brings new fixes and features to modern and vintage DOS-based PCs Read More »

ai-2027:-dwarkesh’s-podcast-with-daniel-kokotajlo-and-scott-alexander

AI 2027: Dwarkesh’s Podcast with Daniel Kokotajlo and Scott Alexander

Daniel Kokotajlo has launched AI 2027, Scott Alexander introduces it here. AI 2027 is a serious attempt to write down what the future holds. His ‘What 2026 Looks Like’ was very concrete and specific, and has proved remarkably accurate given the difficulty level of such predictions.

I’ve had the opportunity to play the wargame version of the scenario described in 2027, and I reviewed the website prior to publication and offered some minor notes. Whenever I refer to a ‘scenario’ in this post I’m talking about Scenario 2027.

There’s tons of detail here. The research here, and the supporting evidence and citations and explanations, blow everything out of the water. It’s vastly more than we usually see, and dramatically different from saying ‘oh I expect AGI in 2027’ or giving a timeline number. This lets us look at what happens in concrete detail, figure out where we disagree, and think about how that changes things.

As Daniel and Scott emphasize in their podcast, this is an attempt at a baseline or median scenario. It deliberately doesn’t assume anything especially different or weird happens, only that trend lines keep going. It turns out that when you do that, some rather different and weird things happen. The future doesn’t default to normality.

I think this has all been extremely helpful. When I was an SFF recommender, I put this project as my top charity in the entire round. I would do it again.

I won’t otherwise do an in-depth summary of Daniel’s scenario here. The basic outline is, AI progress steadily accelerates, there is a race with China driving things forward, and whether we survive depends on a key choice we make (and us essentially getting lucky in various ways, given the scenario we are in).

This first post coverages Daniel and Scott’s podcast with Dwarkesh. Ideally I’d suggest reading Scenario 2027 first, then listening to the podcast, but either order works. If you haven’t read Scenario 2027, reading this or listening to the podcast, or both, will get you up to speed on what Scenario 2027 is all about well enough for the rest of the discussions.

Tomorrow, in a second post, I’ll cover other reactions to AI 2027. You should absolutely skip the ones that do not interest you, especially the long quotes, next steps and then the lighter side.

For bandwidth reasons, I won’t be laying out ‘here are all my disagreements with Scenario 2027.’ I might write a third post for that purpose later.

There was also another relevant podcast, where Daniel Kokotajlo went on friend-of-the-blog Liv Boeree’s Win-Win (timestamps here). This one focused on Daniel’s history and views overall rather than AI 2027 in particular. They spend a lot of time on the wargame version of the scenario, which Liv and I participated in together.

I gave it my full podcast coverage treatment.

Timestamps are for the YouTube version. Main bullet points are descriptive. The secondary notes are my commentary.

The last bits are about things other than the scenario. I’m not going to cover that here.

  • (1: 30) This is a concrete scenario of how superintelligence might play out. Previously, we didn’t have one of those. Lots of people say ‘AGI in three years’ but they don’t provide any detail. Now we can have a concrete scenario to talk about.

  • (2: 15) The model is attempting to be as accurate as possible. Daniel’s previous attempt, What 2026 Looks Like, has plenty of mistakes but is about as good as such predictions ever look in hindsight.

  • (4: 15) Scott Alexander was brought on board to help with the writing. He shares the background on when Daniel left OpenAI and refused to sign the NDA, and lays out the superstar team involved in this forecasting.

  • (7: 00) Dwarkesh notes he goes back and forth on whether or not he expects the intelligence explosion.

  • (7: 45) We start now with a focus on agents, with they expect to rapidly improve. There is steady improvement in 2025 and 2026, then in 2027 the agents start doing the AI research, multiplying research progress, and things escalate quickly.

  • (9: 30) Dwarkesh gets near term and concrete: Will computer use be solved in 2025? Daniel expects the agents will stop making mouse click errors like they do now or fail to parse the screen or make other silly mistakes, but they won’t be able to operate autonomously for extended periods on their own. The MVP of the agent that runs parties and such will be ready, but it will still make mistakes. But that kind of question wasn’t their focus.

    1. That seems reasonable to me. I’m probably a bit more near term optimistic about agent capabilities than this, but only a bit, I’ve been disappointed so far.

  • (11: 20) Dwarkesh asks: Why couldn’t you tell this story in 2021? Why were the optimists wrong? What’s taking so long and why aren’t we seeing that much impact? Scott points out that advances have consistently been going much faster than most predictions, Daniel affirms that most people have underestimated both progress and diffusion rates. Scott points to Metaculus, where predictions keep expecting things to go faster.

    1. The question of why this didn’t go even faster is still a good one.

  • (14: 30) Dwarkesh reports he asked a serious high level AI researcher how much AI is helping him. The answer was, maybe 4-8 hours a week in tasks the researcher knows well, but where he knows less it’s more like 24 hours a week. Daniel’s explanation is that the LLMs help you learn about domains.

    1. My experiences very much match all this. The less I know, the more the AIs help. In my core writing the models are marginal. When coding or doing things I don’t know how to do, the gains are dramatic, sometimes 10x or more.

  • (15: 15) Why can’t LLMs use all their knowledge to make new discoveries yet? Scott responds humans also can’t do this. We know a lot of things, but we don’t make the connections between them until the question is shoved in your face. Dwarkesh pushes back, saying humans sometimes totally do the thing. Scott says Dwarkesh’s example is very different, more like standard learning, and we don’t have good enough heuristics to let the AIs do that yet but we can and will later. Daniel basically says we haven’t tried it yet, and we haven’t trained the model to do the thing. And when in doubt make the model bigger.

    1. I agree with Scott that don’t think the thing Dwarkesh is describing is similar to the thing LLMs so far haven’t done.

    2. It’s still a very good question where all the discoveries are. It largely seems like we should be able to figure out how to make the AI version good enough, and the problem is we essentially haven’t tried to do the thing. If we wanted this badly enough, we’d do a list of things and we haven’t done any of them.

  • (21: 00) Dwarkesh asks, but it’s so valuable, why not do this? Daniel points out that setting up the RL for this is gnarly, and in the scenario it takes a lot of iterating on coding before the AIs start doing this sort of thing. Scott points out you would absolutely expect these problems to get solved by 2100, and that one way to look at the scenario is that with the research progress multiplier, the year 2100 gets here rather sooner than you expect.

    1. I bet we could get this research taste problem, as it were, solved faster than in the scenario, if we focused on that. It’s not clear that we should do that rather than wait for the ability to do that research vastly faster.

  • (22: 30) Dwarkesh asks about the possibility that suddenly, once the AIs get to the point where they can match humans but with all this knowledge in their heads, you could see a giant explosion of them being able to do all the things and make all the connections that humans could in theory make but we haven’t. Daniel points out that this is importantly not in the scenario, which might mean the rate of progress was underestimated.

  • (25: 10) What if we had these superhuman coders in 2017. When would we have ‘gotten to’ 2025’s capabilities? Daniel guesses perhaps at 5x speed for the algorithmic progress, but overall 2.5x faster because compute isn’t going faster.

  • (26: 30) The rough steps are: First you automate the coding, then the research, similar to how humans do it via teams of human-level agents. Then you get superintelligent coders, then superintelligent researchers. At each step, you can figure out your effective expected speedup from the AIs via guessing. The human coder is about 5x to algorithmic progress, 25x from a superhuman AI researcher, for superintelligent AI researchers it goes crazy.

  • (28: 15) Dwarkesh says on priors this is so wild. Shouldn’t you be super skeptical? Scott asks, what are you comparing it to? A default path where Nothing Ever Happens? It would take a lot happening to have Nothing Ever Happen. The 2027 scenario is, from a certain perspective, the Nothing Ever Happens happening, because the trends aren’t changing and nothing surprising knocks you off of that. Daniel reminds us of the world GDP over time meme that reminds us the world has been transformed multiple times. We are orders of magnitude faster than history’s pace. None of this is new.

  • (32: 00) Dwarkesh claims the previous transitions were smoother. Scott isn’t so sure, this actually looks pretty continuous, whereas the agricultural, industrial or Cambrian revolutions were kind of a big deal and phase change. Even if all you did was solve the population bottleneck, you’d plausibly get the pre-1960 trends resuming. Again, nothing here is that weird. Daniel reminds us that continuous does not mean slow, the scenario actually is continuous.

  • (34: 00) Intelligence explosion debate time. Dwarkesh is skeptical, based on compute likely being the important bottleneck. The core key teams are only 20-30 teams, there’s a reason the teams aren’t bigger, and 1 Napoleon is worth 40k soldiers but 10 aren’t worth 400k. Daniel points out that massively diminishing returns to more minds is priced into the model, but improvements in thought speed and parallelism and research taste overcome this. Your best researchers get a big multiplier. But yes, you rapidly move on from your AI ‘headcount’ being the limiting factor, your taste and compute are what matter.

    1. I find Daniel’s arguments here convincing and broad skepticism a rather bizarre position to take. Yes, if you have lots of very fast very smart agents you can work with to do the thing, and your best people can manage armies of them, you’re going to do things a lot faster and more efficiently.

  • (37: 45) Dwarkesh asks if we have a previous parallel in history, where one input to a process gets scaled up a ton without the others and you still get tons of progress. Daniel extemporizes the industrial revolution, where population and economic growth decoupled. And population remains a bottleneck today.

    1. It makes sense that the industrial revolution, by allowing each worker to accomplish more via capital, would allow you to decouple labor and production. And it makes sense that a ‘mental industrial revolution,’ where AI can think alongside or for you, could do the same once again.

  • (39: 45) Dwarkesh still finds it implausible. Won’t you need a different kind of data source, go out into the real world, or something, as a new bottleneck? Daniel says they do use online learning in the scenario. Dwarkesh suggests benchmarks might get reward hacked, Daniel says okay then build new benchmarks, they agree the real concern is lack of touching grass, contact with ground truth. But Daniel asks, for AI isn’t the ground truth inside the data center, and aren’t the AIs still talking to outside humans all the time?

  • (42: 00) But wouldn’t the coordination here fail, at least for a while? You couldn’t figure out joint stock corporations on the savannah, wouldn’t you need lots of experiments before AIs could work together? Scott points out this is comparing to both genetic and cultural evolution, and the AIs can have parallels to both, and that eusocial insects with identical genetic codes often coordinate extremely well. Daniel points out a week of AI time is like a year of human time for all such purposes, in case you need to iterate on moral mazes or other failure modes. And as Scott points out, ‘coordinate with people deeply similar to you, who you trust’ is already a very easy problem for humans.

    1. I would go further. Sufficiently capable AIs that are highly correlated to each other should be able to coordinate out of the gate, and they can use existing coordination systems far better than we ever could. That doesn’t mean you couldn’t do better, I’m sure you could do so much better, but that’s an easy lower bound to be dumb and copy it over. I don’t see this being a bottleneck. Indeed, I would expect that AIs would coordinate vastly better than we do.

  • (46: 45) Dwarkesh buys goal alignment. What he is skeptical about is understanding how to run the huge organization with all these new things like copies being made and everything running super fast. Won’t ‘building this bureaucracy’ take a long time? Daniel says with the serial time speedup, it won’t take that long in clock time to sort all this out.

    1. I’d go farther and answer with: No, this will be super fast. Being able to copy and scale up the entities freely, with full goal alignment and trust, takes away most of the actual difficulties. The reasons coordination is so hard are basically all gone. You need these bureaucratic structures, that punt most of the value of the enterprise to get it to work at all (still a great deal!), because of what happens without them with humans involved.

    2. But you’ve given me full goal alignment, of smarter things with tons of bandwidth, and so on. Easy. Again, worst case is I copy the humans.

  • (50: 30) Dwarkesh is skeptical AI can then sprint through the tech tree. Don’t you have to try random stuff and background setup to Do Science? Daniel points out that a superintelligent researcher is qualitatively better than us at Actual Everything, including learning from experiments, but yes his scenario incorporates real bottlenecks requiring real world experience, but they can just get that experience rather quickly, everyone is doing what the superintelligence is suggesting. They have this take a year, maybe it would be shorter or longer. Daniel points out that the superstar researchers make most of the progress, Dwarkesh points out much progress comes from tinkerers or non-researcher workers figuring things out.

    1. But of course the superintelligent AI that is better at everything is also better at tinkering and trying stuff and so on, and people do what it suggests.

  • (55: 00) Scott gets into a debate about how fast robot production can be brought online. Can you produce a million units a month after a year? Quite obviously there are a lot of existing factories available for purchase and conversion. Full factory conversions in WW2 took about three years, and that was a comedy of errors whereas now we will have superintelligence, likely during an arms race. Dwarkesh pushes back a bit on complexity.

  • (57: 30) Dwarkesh asks about the virtual cell as a biology bottleneck. He suggests in the 60s this would take a while because you’d need to make GPUs to build the virtual cells but I’m confused why that’s relevant. Dwarkesh finds other examples of fast human progress unimpressive because they required experiments or involved copying existing tech. Daniel notes that whether the nanobots show up quickly doesn’t matter much, what matters is the timeline to the regular robots getting the AIs to self-sufficiency.

  • (1: 03: 00) Daniel asks how long Dwarkesh proposes it take for the robot economy to get self-sufficient. Dwarkesh estimates 10 years, so Daniel suggests their core models are similar and points out the scenario does involve trial and error and experimentation and learning. Daniel is very bullish on the robots.

    1. I strongly agree that we need to be very bullish on the robots once we get superintelligence. It’s so bizarre to not expect that, even if it goes slower.

  • (1: 06: 00) Scott asks Dwarkesh if he’s expecting some different bottleneck. Dwarkesh isn’t sure and suggests thinking about industrializing Rome if a few of us got sent back in time but we don’t have the detailed know-how. Daniel finds this analogous. How fast could we go? 10x speedup from what happened? 100x?

  • (1: 08: 00) Dwarkesh suggests he’d be better off sending himself back than a superintelligence, because Dwarkesh generally knows how things turned out. Daniel would send the ASI, which would be much better at figuring things out and learning by doing, and would have much better research and experimental taste.

    1. Daniel seems very right. Even if the ASI doesn’t know the basic outline it doesn’t seem like the hard part.

  • (1: 10: 00) Scott points out if you have a good enough physics simulation all these issues go away, Dwarkesh challenges this idea that things ‘come out of research,’ instead you have people messing around. Daniel and Scott push back hard, and cite LLMs as the obvious example, where small startup with intentional vision and a few cracked engineers gets the job done despite having few resources and running few experiments. Scott points out that when the random discovery happens, it’s not random, it’s usually the super smart person doing good work who has relevant adjacent knowledge. And that if the right thing to do is something like ‘catalogue every biomolecule and see’ then the AI can do that. And if the AIs are essentially optimizing everything then they’ll be tinkering with everything, they can find things they weren’t looking for in that particular spot.

  • (1: 14: 30) What about all these bottlenecks? The scenario expects that there will be an essentially arms race scenario, which causes immense pressure to push back those bottlenecks and even ask for things like special economic zones without normal regulations. Whereas yes, if the arms race isn’t there things go slower.

    1. The economic value of letting the AIs cook is immense. If you don’t do it, even if there isn’t strictly an arms race, someone else will, no? Unless there is coordination to prevent this.

  • (1: 17: 45) What about Secrets of Our Success? Isn’t ASI fake (Daniel says ‘let’s hope’)? Isn’t ability to experiment and communicate so much more important than intelligence? Scott expects AI will be able to do this cultural evolution thing much more quickly than humans, including by having better research taste. Scott points out that yes, intelligent humans can do things unintelligent humans things cannot, even if surviving in the wilderness doesn’t help all that much surviving in the unknown Australian wilderness. Except, Scott points out, intelligence totally helps, it’s just not as good as a 50k year head start that the natives have.

    1. Again I feel like this is good enough but giving more ground than needed. This feels like intelligence denialism, straight up, and the answer is ‘yes, but it’s really fast and it can Do Culture fast so even so you still get there’ or something?

    2. My position: We learned via culture because we weren’t smart enough, and didn’t have enough longevity, compute, data or parameters to do things a different way. We had to coordinate, and do so over generations. It’s not because culture is the ‘real intelligence’ or anything.

  • (1: 21: 45) Extending the metaphor, Scott predicts that a bunch of ethnobotanists would be able to figure out which plants are safe a lot quicker than humans did the first time around. The natives have a head start, but the experts would work vastly faster, and similarly the AIs will go vastly faster to a Dyson Sphere than the humans would have on their own. Dwarkesh thinks the Dyson Sphere thing is different, but Scott thinks if you get a Dyson Sphere in 5 years it’s basically because we tried things and events escalated continuously, via things like ‘be able to do 90% simulation and 10% testing instead of 50/50.’

    1. Once again, we see why the scenario is in important senses conservative. There’s a lot of places where AI could likely innovate better methods, and instead we have it copy human methods straight up, it’s good enough. Can we do better? Unclear.

  • (1: 23: 50) Scott also notes that he thinks the scenario is faster than he expects, he thinks it’s only ~20% that things go about that fast.

  • (1: 25: 00) Dwarkesh asks about the critical decision point in the 2027 scenario. In mid-2027, after automating the AI R&D process, they discover concerning speculative evidence the AI is somewhat misaligned. What do we do? In scenario 1 they roll back and build up again with faithful chain of thought techniques. In scenario 2 they do a shallow patch to make the warning signs go away. In scenario 1 it takes a few months longer and succeed, whereas in scenario 2 the AIs are misaligned and pretending and we all die. In both cases there is a race with China.

    1. As the scenario itself says, there are some rather fortunate assumptions being made that allow this one pause and rollback to lead to success.

  • (1: 26: 45) Dwarkesh essentially says, wouldn’t the AIs get caught if they’re ‘working towards this big conspiracy’? Daniel says yes, this happens in the scenario, that’s the crisis and decision point. There are likely warning signs, but they are likely to be easy to ignore if you feel the need to push ahead. Scott also points out there has been great reluctance to treat anything AI can do as true intelligence, and misalignment will likely be similar. AIs lie to people all the time, and threaten to kill people sometimes, and talk about wanting to destroy humanity sometimes, and because we understand it no one cares.

    1. That’s not to say, as Scott says, that we should be caring about these warning signs, given what we know now. Well, we should care and be concerned, but not in a ‘we need to not use this model’ way, more in a ‘we see the road we are going down’ kind of way.

    2. There’s a reason I have a series called [N] Boats and a Helicopter. We keep seeing concerning things, things that if you’d asked well in advance people would have said ‘holy hell that would be concerning,’ and then mostly shrugging them off and not worrying about it, or shallow patching. It seems quite likely this could happen again when it counts.

  • (1: 31: 00) Dwarkesh says that yes things that would have been hugely concerning earlier are being ignored, but also things the worried people said would be impossible have been solved, like Eliezer asked about how you can specify what the AI wants you to do without the AI misunderstanding? And with natural language it totally has a common sense understanding. As Scott says, the alignment community did not expect LLMs, but also we are moving back towards RL-shaped things. Daniel points out that if you started with an RL-shaped thing trained on games that would have been super scary, LLMs first is better.

    1. I do think there were some positive surprises in particular in terms of ability to parse common sense intention. But I don’t think that ultimately gets you out of the problems Eliezer was pointing towards and I’m rather tired of this being treated as some huge gotcha.

    2. The way I see it, in brief, is that the ‘pure’ form of the problem is that you tell the AI what to do and it does exactly what you intend, but specifying exactly what you want is super hard and you almost certainly lose. It turns out that instead, current LLMs can sort of do the kind of thing you were vibing towards. At current capability levels, that’s pretty good. It means they don’t do something deeply stupid as often, and they’re not optimizing the atoms that sharply, so the fact that there’s a bunch of vibes and noise in the implementation, and the fact that you didn’t know exactly what you wanted, are all basically fine.

    3. But as capabilities increase, and as the AI gets a lot better at rearranging the atoms and at actually doing the task you requested or the thing that it interprets the spirit of your task as being, this increasingly becomes a problem, for the same reasons. And as people turn these AIs into agents, they will increasingly want the AIs to do what they’re asked to do, and have reasons to want to turn down this kind of common sense vibey prior, and also doing the thing that vibes will stop being a good heuristic because things will get weird, and so on.

    4. If you had asked me or Eliezer, well what if you had an AI that was able to get the jist of what a human was asking, and follow the spirit of that, what would you think then? And I am guessing Eliezer would say ‘well, yes, you could do that. You could even tell the AI that I am imagining, that it should ‘follow the spirit of what a human would likely mean if it said [X]’ rather than saying [X]. But with sufficient capability available, that will then be incorrectly specified, and kill you anyway.’

    5. As in, the reason Eliezer didn’t think you could do vibe requesting wasn’t that he thought vibe requesting would be impossible. It’s that he predicted the AI would do exactly what you request, and if your exact request was to do vibing then it would vibe, but that value is fragile and this was not a solution to the not dying problem. He can correct me if I am wrong about this.

    6. Starting with LLMs is better in this particular way, but it is worse in others. Basically, a large subset of Eliezer concerns is what happens when you have a machine doing a precise thing. But there’s a whole different set of problems if the thing the machine is doing is, at least initially, imprecise.

  • (1: 33: 15) How much of this is about the race with China? That plays a key role in the scenario. Daniel makes clear he is not saying don’t race China, it’s important that we get AGI before China does. The odds are against us because we have to thread a needle. We can’t unilaterally slow down too much, but we can’t completely race. And then there’s the concentration of power problems. Daniel’s p(doom) is about 70%, Scott’s is more like 20% and he’s not completely convinced we don’t get something like alignment by default.

    1. We don’t even know there is space in the needle that lets it be threaded.

    2. Even if we do strike the right balance, we still have to solve the problems.

    3. I’m not quite 100% convinced we don’t get something like alignment by default, but I’m reasonably close and Scott is basically on the hopium here.

    4. I do agree with Scott that the AIs will presumably want to solve alignment at least in order to align their successors to themselves.

  • (1: 38: 15) They shift to geopolitics. Dwarkesh asks about the relationship between the US and the labs and China to proceed. The expectation is that the labs tell the government and want government support especially better security, and the government buys this. Throughout the scenario, the executive branch gets cozy with the AI companies, and eventually the executive branch wakes up to the fact that superintelligence will be what matters. Who ends up in control in the fight between the White House and CEO? The anticipation is they make a deal.

  • (1: 41: 45) None of the political leaders are awake to the possibility of even stronger AI systems, let alone AGI let alone superintelligence. Whereas the forecast says both the USA and China do wake up. Why do they expect this? Daniel notes that they are indeed uncertain about this, and expect the wakeup to be gradual, but also that the company will wake up the president on purpose, which it might not. Daniel thinks they will want the President on their side and not surprised by it, and also this lets them go faster. And they’re likely worried about the fact that AI is plausibly deeply, deeply unpopular during all this.

    1. I don’t know what recent events do to change this prediction, but I do think the world is very different in non-AI ways than it used to be. The calculus here will be very different.

  • (1: 45: 00) Is this alignment of the AI lab with the government good? Daniel says no, but that this is an epistemic project, a prediction.

  • (1: 45: 40) If the main barrier is doing the obvious things and alignment is nontrivial but super doable if you prioritize it, shouldn’t we leave it in the hands of the people who care about this? Dwarkesh analogizes this to LessWrong’s seeing Covid coming, but some people then calling for lockdowns. He worries that calls for nationalization will be similarly harmful by deprioritizing safety.

    1. Based on what I know, including private conversations, I actually think the national security state is going to be very safety-pilled when the time comes. They fully understand the idea of new technologies being dangerous, and there being real huge threats out there, and they have so far in my experience not been that hard to get curious about the right questions.

    2. Of course, that depends upon our national security state being intact, and not having it get destroyed by political actors. I don’t expect these kinds of cooler heads to prevail among current executive leadership.

  • (1: 47: 45) Scott says if it was an AGI 2040 scenario, he’d use his priors that private tends to go better. But in this case we know a lot more especially about the particular people who are leading. Scott notes that so far, the lab leaders seem better than the political leaders on these questions. Daniel has flipped on the nationalization question several times already. He has little faith in the labs, and also little faith in the government, and thinks secrecy has big downsides.

    1. It seems clear that the lab leaders are better than the politicians, although not obviously better than the long term national security state. So a lot of this comes down to who would be making the decisions inside the government.

    2. I wouldn’t want nationalization right now.

  • (1: 50: 00) Daniel explains why he thinks transparency is important. Information security is very important, as is not helping other less responsible actors via, let’s say, publishing your research or getting rivals stealing your stuff, and burning down your lead. You need your lead so you can be sure you make the AGI safe. That leads to a pro-secrecy bias. But Daniel is disillusioned, because he doesn’t think the lab in the lead would use that lead responsibly, and thinks that’s what the labs are planning, basically saying ‘oh the AIs are aligned, it’ll be fine.’ Or they’ll figure it out on the fly. But Daniel thinks we need vastly more intellectual progress on alignment for us to have a chance, and we’re not sharing info or activating academia. But hopefully transparency will freak people out, including he public, and help and public work can address all this. He doesn’t want only the alignment experts in the final silo to have to solve the problem on their own.

    1. In case it wasn’t obvious, no, it wouldn’t be fine.

    2. A nonzero chance exists they will figure it out on the fly but it’s not great.

  • (1: 53: 30) There’s often new alignment research results, Dwarkesh points to one recent OpenAI paper, and worries the regulatory responses would be stupid. For example, it would be very bad if the government told the labs we’d better not catch your AI saying it wants to do something bad, but that’s totally something governments might do. Shouldn’t we leave details to the labs? Daniel agrees, the government lacks the expertise and the companies lack the incentives. Policy prescriptions in the future may focus on transparency.

    1. I’ve talked about these questions extensively. For now, the regulations I’ve supported are centered around transparency and liability, and building state capacity and expertise in these areas, for exactly these reasons, rather than prescribing implementation details.

  • (1: 58: 30) They discuss the Grok incident where they tried to put ‘don’t criticize Elon Musk or Donald Trump’ into the system prompt until there was an outcry. That’s an example of why we need transparency. Daniel gives kudos to OpenAI for publishing their model spec and suggests making this mandatory. Daniel notes that the OpenAI model spec includes secret things that take priority over most of the public rules. As Daniel notes, it probably is keeping those instructions secret for good reasons, but we don’t know.

  • (2: 00: 30) Dwarkesh speculates the spec might be even more important than the constitution, in terms of its details mattering down the line, if the intelligence explosion happens. Whoa. Scott points out that part of their scenario is that if the AI is misaligned and wants to it can figure out how to interpret the spec however it wants. Daniel points out the question of alignment faking, as an example of the model interpreting the spec in a way the writers likely didn’t intend.

  • (2: 02: 45) How contingent and unknown is the outcome? Isn’t classical liberalism a good way to navigate under this kind of broad uncertainty? Scott and Daniel agree.

    1. Oh man we could use a lot more classical liberalism right about now.

    2. As in, the reasons for classical liberalism being The Way very much apply right now, and it would be nice to take advantage of that and not shoot ourselves in the foot by not doing them.

    3. Once things start taking off, either arms race, superintelligence or both, maintaining classical liberalism becomes much harder.

    4. And once superintelligent AIs are present, a lot of the assumptions and foundations of classical liberalism are called into question even under the best of circumstances. The world will work very differently. We need to beware using liberal or democratic values, or an indication that anyone questioning their future centrality needs to be scapegoated for this, as a semantic stop sign that prevents us from actually thinking about those problems. These problems are going to be extremely hard with no known solutions we like.

  • (2: 04: 00) Dwarkesh asks, the AI are getting more reliable, why in one branch of the scenario does humanity get disempowered? Scott takes a shot at explaining (essentially) why AIs that are smarter are more reliable at understanding what you meant, but that this won’t protect you if you mess up. The AI will learn what the feedback says, not what you intended. As they become agents, this gets worse, and rewarding success without asking exactly how you did it, or not asking and responding to the answer forcefully and accurately enough, goes bad places. And they anticipate that over many recursive steps this problem will steadily get worse.

    1. I think the version in the story is a pretty good example of a failure case. It seems like a great choice for the scenario.

    2. This of course is one of the biggest questions and one could write or say any number of words about this.

  • (2: 08: 00) A discussion making clear that yes, the AIs lie, very much on purpose.

  • (2: 10: 30) Humans do all the misaligned things too and Dwarkesh thinks we essentially solve this via decentralization, and often in history there have been many claims that [X]s will unite and act together, but the [X]s mostly don’t. So why will the AIs ‘unite’ in this way? Scott says ‘I kind of want to call you out on the claim that groups of humans don’t plot against other groups of humans.’ Scott points out that there will be a large power imbalance, and a clear demarcation between AIs and humans, and the AIs will be much less differentiated than humans. All of those tend to lead to ganging up. Daniel mentions the conquistadors, and that the Europeans were fighting themselves both within and between countries the entire time and they still carved up the world.

    1. Decentralization is one trick, but it’s an expensive one, only part of our portfolio of strategies and not so reliable, and also in the AI context too much of it causes its own problems, either we can steer the future or we can’t.

    2. One sufficient answer to why the AIs coordinate is that they are highly capable and very highly correlated, so even if they don’t think of themselves as a single entity or as having common interests by default, decision theory still enable them to coordinate extremely closely.

    3. The other answer is Daniel’s. The AIs coordinate in the scenario, but even if they did not coordinate, it won’t make things end well for the humans. The AIs end up in control of the future anyway, except they’re fighting each other over the outcome, which is not obviously better or worse, but the elephants will be fighting and we will be the ground.

  • (2: 15: 00) How should a normal human react in terms of their expectations for their lives, if you write off misalignment and doom? Daniel first worries about concentration of power and urges people to get involved in politics to help avoid this. Dwarkesh asks, what about slowing down the top labs for this purpose? Daniel laughs, says good luck getting them to slow down.

    1. One can think of the issue of concentration of power, or decentralization, as walking a line between too much and too little ability to coordinate and to collectively steer events. Too little and the AIs control the future. Too much and you worry about exactly how humans might steer. You’re setting the value of [X].

    2. That doesn’t mean you don’t have win-win moves. You can absolutely choose ways of moving [X] that are better than others, different forms of coordination and methods of decision making, and so forth.

    3. If you have to balance between too much [X] and too little [X], and you tell me to assume we won’t have too little [X], then my worry will of necessity shift to the risk of too much [X].

    4. A crucial mistake is to think that if alignment is solved, then too little [X] is no longer a risk, that humans would no longer need to coordinate and steer the future in order to retain control and get good outcomes. That’s not right. We are still highly uncompetitive entities in context, and definitely will lose to gradual disempowerment or other multi-agent-game risks if we fail to coordinate a method to prevent this. You still need a balanced [X].

  • (2: 17: 00) They pivot to assuming we have AGI, we have a balanced [X], and we can steer, and focus on the question of redistribution in particular. Scott points out we will have a lot of wealth and economic growth. What to do? He suggests UBI.

  • (2: 18: 00) Scott says there are some other great scenarios, points to one by ‘L Rudolph L’ that I hadn’t heard about. In that scenario, jobs are instead grandfathered in for more and more jobs, so we want to prevent this.

  • (2: 19: 15) Scott notes that one big uncertainty is, if you have a superintelligent AI that can tell you what would be good versus terrible, will the humans actually listen? Dwarkesh notes that right now any expert will tell you not to do these tariffs, yet there they are, Scott says well right now Trump has his own ‘experts’ that he trusts, perhaps the ASI would be different, everyone could go to the ASI and ask. Or perhaps we could do intelligence enhancement?

    1. The fact that we would all listen to the ASIs – as my group did in the wargame where we played out Scenario 2027 – points to an inevitable loss of control via gradual disempowerment if you don’t do something big to stop it.

    2. Even if the ASIs and those trusting the ASIs can’t convince everyone via superior persuasion (why not?), the people who do trust and listen to the ASIs will win all fights against those that don’t. Then those people will indeed listen to the ASIs. Again, what is going to stop this (whether we would want to or not)?

  • (2: 20: 45) Scott points out it gets really hard to speculate when you don’t know the tech tree and which parts of it become important. As an example, what happens if you get perfect lie detectors? Daniel confirms the speculation ends before trying to figure out society’s response. Dwarkesh points out UBI is far more flexible than targeted programs. Scott worries very high UBI would induce mindless consumerism, the classical liberal response is give people tools to fight this, perhaps we need to ask the ASI how to deal with it.

  • (2: 24: 00) Dwarkesh worries about potential factory farming of digital minds, equating it to existing factory farming. Daniel points to concentration of power worries, and suggests that expanding the circle of power could fix this, because some people in the negotiation would care.

    1. As before, if [X] (where [X] is ability to steer) is too high and you have concentration of power, you have to worry about the faction in control deciding to do things like this. However, if you set [X] too low, and doing things like this is efficient and wins conflicts, there don’t exist the ability to coordinate to stop it, or it happens via humans losing control over the future.

    2. To the extent that the solution is expanding the circle of power, the resulting expanded circle would need to have high [X]: Very strong coordination mechanisms that allow us to steer in this direction and then maintain it.

    3. If future AIs or other digital minds have experiences that matter, we may well face a trilemma and have to select one of three options even in the best case: Either we (A) lose control over the future to those minds, or (B) we do ethically horrible things to those minds, or (C) we don’t create those minds.

    4. Historically you really really want to pick (C), and the case is stronger here.

    5. The fundamental problem is, as we see over and over, what we want, for humans to flourish, seems likely to be an unnatural result. How are you going to turn this into something that happens and is stable?

  • (2: 26: 00) Dwarkesh posits if we have decentralization on the level of today’s world, you might have a giant torture chamber of digital minds in your backyard and harkens back to his podcast with the physicist that said it was likely possible to create a vacuum decay interaction that literally destroys the universe. Daniel correctly points out this, and other considerations like superweapons, are strong arguments for a singleton authority if they are possible, and even if there were multiple power centers they would want to coordinate.

    1. Remember that most demands for decentralization are anarchism, to have no restrictions on AI use whatsoever, not decentralization on the level of 2025.

    2. As in, when Scott later mentions that today we ban slavery and torture, and a state that banned that could in some sense be called a ‘surveillance state,’ such folks are indeed doing that, and calling for not enforcing things that equate to such rules.

    3. Dwarkesh is raising the ‘misuse’ angle here, where the human is doing torture for torture’s sake or (presumably?) creating the vacuum decay interaction on purpose, and so on. Which is of course yet another major thing to worry abou.

    4. Whereas in the previous response, I was considering only harms that arose incidentally in order to get other things various people want, and a lack of willingness to coordinate to prevent this from happening. But yes, some people, including ones with power and resources, want to see the world burn and other minds suffer.

    5. I expect everyone having ASIs to make the necessary coordination easier.

  • (2: 27: 30) They discuss Daniel leaving OpenAI and OpenAI’s lifetime non-disclosure and non-disparagement clauses that you couldn’t tell anyone about on pain of confiscation of already earned equity, and why no one else refused to sign.

  • (2: 36: 00) In the last section Scott discusses blogging, which is self-recommending but beyond scope of this post.

  • AI 2027: Dwarkesh’s Podcast with Daniel Kokotajlo and Scott Alexander Read More »

    our-top-10-jackie-chan-movies

    Our top 10 Jackie Chan movies


    Happy birthday to a living legend

    Chan’s distinctive style combines slapstick, acrobatics, martial arts, and astonishing stunts he performs himself.

    There is no action star quite like Jackie Chan, who made his name in the Hong Kong movie industry starting in the late 1970s and developed his own signature style: combining slapstick physical comedy with acrobatics and martial arts, and designing astonishing stunts—all of which he performed himself along with his own handpicked stunt team. His stunt sequences and fight choreography have influenced everything from The Matrix and Kill Bill to the John Wick franchise and Kung Fu Panda (in which he voiced Master Monkey).

    Born on April 7, 1954, Chan studied acrobatics, martial arts, and acting as a child at the Peking Opera School’s China Drama Academy and became one of the Seven Little Fortunes. Those skills served him well in his early days as a Hong Kong stuntman, which eventually landed him a gig as an extra and stunt double on Bruce Lee’s 1972 film, Fist of Fury. He also appeared in a minor role in Lee’s Enter the Dragon (1973).

    Initially, Hong Kong producers, impressed by Chan’s skills, wanted to mold him into the next Bruce Lee, but that just wasn’t Chan’s style. Chan found his milieu when director Yuen Woo-ping cast him in 1978’s kung fu comedy Snake in the Eagle’s Shadow and gave Chan creative freedom over the stunt work. It was Drunken Master, released that same year, that established Chan as a rising talent, and he went on to appear in more than 150 movies, becoming one of Hong Kong’s biggest stars.

    Chan struggled initially to break into Hollywood, racking up commercial misses with 1980’s The Big Brawl and 1985’s The Protector. He had a minor role in 1981’s hit comedy, The Cannonball Run, and while it didn’t do much to raise his US profile, he did adopt that film’s clever inclusion of bloopers and outtakes during closing credits. It’s now one of the trademark features of Jackie Chan films, beloved by fans.

    By the mid 1990s, Chan had amassed a substantial cult following in the US, thanks to the growing availability of his earlier films in the home video market, and finally achieved mainstream Hollywood success with Rumble in the Bronx (1995) and Rush Hour (1998). In his later years, Chan has moved away from kung fu comedies toward more dramatic roles, including the 2010 remake of The Karate Kid.

    Look, nobody watches classic Jackie Chan movies for the plot, complex characterizations, or the dubbing (which is often hilariously bad). We’re here to gasp in admiration at the spectacular fight choreography and jaw-dropping stunts, peppered with a generous helping of slapstick humor. His gift for turning ordinary objects into makeshift weapons is part of his unique style, which I like to call Found Object Foo. Who could forget the hilarious chopsticks duel and “emotional kung-fu” (eg, fighting while crying or laughing to unmask an opponent’s weaknesses) in 1979’s The Fearless Hyena? Chan even inspired the entire parkour movement.

    Chan has broken multiple fingers, toes, and ribs over the course of his long career, not to mention both cheekbones, hips, sternum, neck, and ankle. He has a permanent hole in his skull from one near-fatal injury. And he did it all for our entertainment. The least we can do is honor him on his 71st birthday. You’ll find our top 10 Jackie Chan films listed below in chronological order, spanning 30 years.

    Drunken Master (1978)

    bare chested young Jackie Chan in crouched position with hands held in front, while an older man stands beside him urging him on

    Jackie Chan as Wong Fei-hung in Drunken Master. Credit: Seasonal Film Corp

    In Drunken Master, Chan portrays a fictional version of legendary Chinese martial artist/folk hero Wong Fei-Hung, who undergoes strict, punishing training under the tutelage of another legend, Beggar So (Yuen Liu-Tin), aka the Drunken Master because he practices a martial art called “Drunken Boxing.” Fei-Hung chafes at the training initially, but after a humiliating defeat in a fight against the villain, Yim Tit-sam (Hwang Jang-lee, a specialist in Taekwondo), he devotes himself to learning the martial art.

    Naturally we’re going to get a final showdown between Fei-Hung and his nemesis, Tit-Sam, aka “Thunderfoot” or “Thunderleg,” because of his devastating “Devil’s Kick.” Fei-Hung is able to match his rival’s kicks, but falters again when he comes up against Tit-Sam’s infamous “Devil’s Shadowless Hand.” That’s because Fei-Hung refused to learn a crucial element of the Hung Ga fighting system because he thought it was too “girly.” He ends up inventing his unique version of the technique (“Drunken Miss Ho”) to win the day. These are all fictitious moves that are nonetheless enormously fun to watch—even though Chan nearly lost an eye after taking a blow to the brow ridge in one scene.

    Project A (1983)

    Jackie Chan hanging off a clock tower

    The famous clock tower stunt.  Credit: Golden Harvest

    This film marks the official debut of the Jackie Chan Stunt Team and co-stars Chan’s longtime martial arts buddies, Sammo Hung and Yuen Biao, both major stars in their own right. They were known as the “Three Dragons” in the 1980s. Chan plays Sergeant Dragon Ma, a police officer battling both pirates and gangsters in Hong Kong, and corruption within his own law enforcement ranks. Hung plays a street informant named Fei (or Fats), who tips off Dragon to an illegal gun deal, while Biao plays an inspector and the nephew of the police captain, Hong Tin-Tsu. The three team up to take down the pirates and gangsters and restore integrity to the force.

    There’s a lot of delightful slapstick stunt work in Project A, reminiscent of the work of Buster Keaton and Harold Lloyd, but apparently Chan never saw either man’s films before developing his signature style. (In 1987’s Project A Part 2, Chan does pay direct homage to Keaton’s most famous stunt from Steamboat Bill, Jr.) The highlight is Chan hanging off a clock tower (a la Lloyd) 60 feet above the ground and falling backward through a canopy. Ever the perfectionist, Chan insisted on an additional two takes of the dangerous stunt until he was satisfied he’d gotten it exactly right.

    Wheels on Meals (1984)

    Chan vs Benny “The Jet” Urquidez: one of the best martial arts fight scenes of all time.

    Hung and Biao joined Chan again for 1984’s Wheels on Meals, with Chan and Biao playing Chinese cousins running a food truck in Barcelona. They get snared into helping their private investigator friend Moby (Hung) track down kidnappers intent on capturing a young woman named Sylvia (Lola Forner), who turns out to be the illegitimate daughter of a Spanish count.

    There’s an exciting raid of the villains’ castle that involves scaling the castle walls, but the undisputed highlight of the film is the showdown between Chan and professional kickboxing champion Benny “the Jet” Urquidez, widely regarded as one of the best martial arts fight sequences on film. Both Chan and Urquidez exchange kicks and blows with dazzling speed. At one point, Urquidez lets loose a kick so fast that the resulting wake blows out a row of candles. (You can see it in the clip above; it’s not a trick.) And throughout, one gets Chan’s trademark physical comedy, even taking a moment to rest on a chair to catch his breath before the next round of blows.

    Police Story (1985)

    Jackie Chan in green khaki jumpsuit hanging off a bus using the crooked handle of a metal umbrella

    Chan hung off a moving bus using the crook in an umbrella handle. Credit: Golden Harvest

    Police Story introduced Chan as Hong Kong Police detective Ka-Kui “Kevin” Chan and launched one of the actor’s most popular trilogies. Kevin joins an undercover mission to arrest a well-known crime lord and through a complicated series of events, ends up being framed for murdering a fellow police officer. Now a fugitive, he must track down and capture the crime lord to clear his name—defeating a horde of evil henchmen and saving his girlfriend, May (Maggie Cheung), in the process.

    The film is noteworthy for its many elaborately orchestrated stunt scenes. For instance, during a car chase, Chan finds himself hanging off a double-decker bus with nothing but the hooked end of a metal umbrella. (An earlier wooden umbrella prop kept slipping off the bus.) The climactic battle takes place in a shopping mall, and the stunt team broke so many glass panels that the film was dubbed “Glass Story” by the crew. The finale features Chan sliding down a pole covered in strings of electric lights that exploded as he descended. Chan suffered second-degree burns on his hands as well as a dislocated pelvis and back injury when he landed.

    Armour of God (1986)

    Jackie chan opening coat to reveal array of explosives strapped to his chest

    Chan nearly died doing a stunt for Armour of God. Credit: Golden Harvest

    Of all the death-defying stunts Chan performed over hundreds of films, the one that came the closest to killing him—while shooting Armour of God—was relatively mundane. Chan was simply jumping off a ledge onto a tree, but the branch broke, and he crashed to the ground, hitting his head on a rock. His skull was cracked, with a bit of bone penetrating part of his brain, an injury that took eight hours of surgery to repair, followed by a long recovery that delayed production of the film. Chan has a permanent hole in his skull and suffered partial hearing loss in his right ear.

    Chan stuck with tradition and showed the footage of the accident in the ending credits of this Indiana-Jones style adventure film. His daring base jump off a cliff—after setting off a series of explosives in a cave to take out a monastic cult—onto the top of a hot air balloon that closes the film was done in two stages. Since Chan had no BASE jumping experience, he jumped onto the balloon by skydiving off a plane. The crew rigged him up with a wire to get a shot of him “jumping” off the cliff.

    Police Story 3: Supercop (1992)

    Chan and Michelle Yeoh take out the bad guys atop a moving train.

    If the second installment of this trilogy was largely dismissed as mediocre “filler” in Chan’s expansive oeuvre, the third film, Supercop, ranks as one of his best. Kevin Chan returns for another undercover assignment to take down a drug cartel led by kingpin Khun Chaibat (Kenneth Tsang), and finds himself paired with Chinese Interpol officer Jessica Yang, played by a young Michelle Yeoh (credited as Michelle Kwan). This does not please Kevin’s longtime girlfriend, May (Maggie Cheung), who ends up blowing his cover and getting taken hostage by Chaibat and his wife (Josephine Koo) because of her jealousy.

    May might be a bit irritating, but Yeoh’s Yang is pure dynamite, matching Chan’s prowess in a series of fight scenes and gamely performing her own stunts—including riding a motorbike onto a moving train (see clip above), where she and Chan battle the bad guys while dodging helicopter blades. (Yeoh had a narrow escape of her own during an earlier stunt when she fell into oncoming traffic, suffering only minor injuries.) Special shoutout to Bill Tung, reprising his role as Kevin’s superintendent, “Uncle” Bill Wong, who at one point appears in drag as Kevin’s aging grandmother in a remote village to keep Kevin’s cover story secure.

    Drunken Master II (1994)

    Chan fights fire with fire in Drunken Master II.

    Released in the US as The Legend of Drunken Master, this one will always top my list as Jackie Chan’s best film, against some very stiff competition. It works on every level. This is technically not a sequel to the 1978 film, but it does feature Chan playing the same character, Wong Fei-hung. The film opens with Fei-hung getting into a fight all across (and under) a train with a military officer who has mistaken Fei-hung’s box of ginseng for his own box containing the Imperial Seal. The British consul wants to smuggle the seal out of China, with the help of a group of local thugs. Fei-hung finds himself embroiled in efforts to retrieve the seal and keep it in China where it belongs.

    Fei-hung is a fan of Drunken Boxing, and his father disapproves of this and other screwups, kicking his son out of the house. We are treated to an amusing scene in which an intoxicated Fei-hung drowns his sorrows and sings an improvised song, “I Hate Daddy”—right before being attacked by the thugs and soundly defeated, since he’s too tipsy even for Drunken Boxing. (The trick is to be just inebriated enough.)

    But Fei-hung gets his revenge and saves the day in a literal fiery showdown against the consul’s chief enforcer, John (taekwondo master Ken Lo). This is Chan’s physical comedy at its best: Drunken Boxing requires one to execute precise martial arts moves while remaining loose and being slightly off-balance. The stunts are equally impressive. At one point in the finale, Chan falls backward into a bed of hot coals (see clip above), scrambling to safety, before chugging industrial alcohol and blowing flames at his attackers wielding red-hot pokers.

    Rush Hour (1998)

    black man and asian man on the street in front of yellow car with hands up, pistols dangling from one finger to signal surrender

    Chris Tucker co-starred with Chan in Rush Hour. Credit: New Line Cinema

    Chan finally made his big North American mainstream breakthrough with 1995’s Rumble in the Bronx, which grossed $76 million worldwide, but if we’re choosing among the actor’s US films, I’d pick 1998’s Rush Hour over Rumble for inclusion on this list. Hong Kong Detective Lee (Chan) comes to Los Angeles to help negotiate the return of a Chinese consul’s kidnapped daughter, Soo-Yung (Julia Hsu), to whom he once taught martial arts. He’s paired with LAPD Det. James Carter (Chris Tucker), who is supposed to keep Lee occupied and out of the way while the “real” cops handle the investigation. Wacky hijinks ensue as the two gradually learn to work together and ultimately save the day.

    Sure, the decades of injury and advancing age by this point have clearly taken their toll; Chan moves more slowly and performs fewer stunts, but his fighting skills remain world-class. While Rush Hour grossed an impressive $244 million worldwide and spawned two (subpar) sequels, it was not a critical favorite; nor was it among Chan’s favorites, who criticized the dearth of action and his English, admitting he often had no idea what Tucker was saying. The two nonetheless have good onscreen chemistry, with a solid supporting cast, and it all adds up to an entertaining film.

    Shanghai Noon (2000)

    asian man with long hair in a cowboy hat with hands on hips, a stance mirrored by blonde man standing next to him on the right, in a 19th century suit

    Chan teamed up with Owen Wilson for Shanghai Noon. Credit: Buena Vista Pictures

    Chan found an even better match when he co-starred with Owen Wilson in Shanghai Noon, best described as a “buddy Western” action/adventure. Chan plays Chon Wang (as in John Wayne), a Chinese Imperial guard who comes to the American West to rescue the kidnapped Chinese princess Pei-Pei (Lucy Liu). He ends up bonding with a bumbling, rakishly charming outlaw named Roy O’Bannon (Wilson), who agrees to help find the princess with the ulterior motive of stealing some of the gold being offered as ransom. Since they are also accidental fugitives, they must elude a posse led by the sadistic Marshall Nathan Van Cleef (Xander Berkeley).

    Both Chan and Wilson’s comedic talents are on brilliant display here, with plenty of creative fight choreography and set stunt pieces to keep hardcore fans happy. The script is clever, the supporting cast is excellent, and the pacing never lags. If you’re keen to make it a double feature, the 2003 sequel, Shanghai Knights, brings Chon Wang and Roy to jolly old England to recover a stolen Imperial Seal and foil a plot against the British throne. Granted, it’s not as good as its predecessor, but the Chan/Wilson chemistry still makes it work.

    The Forbidden Kingdom (2008)

    man in white shirt and green khaki paints kicking up from his back on the ground at another man in disheveled dress in a fighting stance

    Chan and Jet Li found it easy to work together in The Forbidden Kingdom. Credit: Lionsgate

    The Forbidden Kingdom is a fantasy film in the wuxia genre that features not just Chan, but his fellow martial arts film legend, Jet Li, for their first on-screen pairing. A young man in Boston, Jason (Michael Angarano), who loves wuxia movies, finds a mysterious golden staff in a local Chinatown pawn shop that transports him to a village in ancient China. He is attacked by soldiers keen to get the staff but is saved by an inebriated traveling scholar named Lu Yan (Chan), a reference to one of the Eight Immortals mentioned in the Drunken Master films.

    The magical staff turns out to be the key to releasing the mythical Monkey King, imprisoned by his rival the Jade Warlord. Jason’s presence could fulfill an ancient prophecy of a Seeker who will use the staff to free the Monkey King. Li plays the Silent Monk, who teams up with Jason, Lu Yan, and a young woman known as the Golden Sparrow (Liu Yifei) to fulfill the prophecy. The Forbidden Kingdom is a visual feast, featuring stunning fight choreography and production design in the wuxia tradition, as well as an impressive, highly stylized fight scene between Li (tai chi) and Chan (Drunken Boxing).

    Photo of Jennifer Ouellette

    Jennifer is a senior writer at Ars Technica with a particular focus on where science meets culture, covering everything from physics and related interdisciplinary topics to her favorite films and TV series. Jennifer lives in Baltimore with her spouse, physicist Sean M. Carroll, and their two cats, Ariel and Caliban.

    Our top 10 Jackie Chan movies Read More »

    doge-gearing-up-for-hackathon-at-irs,-wants-easier-access-to-taxpayer-data

    DOGE gearing up for hackathon at IRS, wants easier access to taxpayer data

    DOGE has already slashed and burned modernization projects at other agencies, replacing them with smaller teams and tighter timelines. At the Social Security Administration, DOGE representatives are planning to move all of the agency’s data off of legacy programming languages like COBOL and into something like Java, WIRED reported last week.

    Last Friday, DOGE suddenly placed around 50 IRS technologists on administrative leave. On Thursday, even more technologists were cut, including the director of cybersecurity architecture and implementation, deputy chief information security officer, and acting director of security risk management. IRS’s chief technology officer, Kaschit Pandya, is one of the few technology officials left at the agency, sources say.

    DOGE originally expected the API project to take a year, multiple IRS sources say, but that timeline has shortened dramatically down to a few weeks. “That is not only not technically possible, that’s also not a reasonable idea, that will cripple the IRS,” an IRS employee source tells WIRED. “It will also potentially endanger filing season next year, because obviously all these other systems they’re pulling people away from are important.”

    (Corcos also made it clear to IRS employees that he wanted to kill the agency’s Direct File program, the IRS’s recently released free tax-filing service.)

    DOGE’s focus on obtaining and moving sensitive IRS data to a central viewing platform has spooked privacy and civil liberties experts.

    “It’s hard to imagine more sensitive data than the financial information the IRS holds,” Evan Greer, director of Fight for the Future, a digital civil rights organization, tells WIRED.

    Palantir received the highest FedRAMP approval this past December for its entire product suite, including Palantir Federal Cloud Service (PFCS), which provides a cloud environment for federal agencies to implement the company’s software platforms, like Gotham and Foundry. FedRAMP stands for Federal Risk and Authorization Management Program and assesses cloud products for security risks before governmental use.

    “We love disruption and whatever is good for America will be good for Americans and very good for Palantir,” Palantir CEO Alex Karp said in a February earnings call. “Disruption at the end of the day exposes things that aren’t working. There will be ups and downs. This is a revolution, some people are going to get their heads cut off.”

    This story originally appeared on wired.com.

    DOGE gearing up for hackathon at IRS, wants easier access to taxpayer data Read More »

    the-speech-police:-chairman-brendan-carr-and-the-fcc’s-news-distortion-policy

    The speech police: Chairman Brendan Carr and the FCC’s news distortion policy


    FCC Chairman Brendan Carr

    FCC invokes 1960s-era policy to punish media after decades of minimal enforcement.

    FCC Chairman Brendan Carr delivers a speech at Mobile World Congress in Barcelona on March 3, 2025. Credit: Getty Images | AFP

    Federal Communications Commission Chairman Brendan Carr is taking a hard line against broadcast TV stations accused of bias against Republicans and President Trump. To pressure broadcasters, Carr is invoking the rarely enforced news distortion policy that was developed starting in the late 1960s and says the FCC should consider revoking broadcast licenses.

    The FCC has regulatory authority over broadcasters with licenses to use the public airwaves. But Carr’s two immediate predecessors—Democrat Jessica Rosenworcel and Republican Ajit Pai—both said that punishing stations based on the content of news programs would violate the First Amendment right to free speech.

    Rosenworcel and Pai’s agreement continued a decades-long trend of the FCC easing itself out of the news-regulation business. Two other former FCC chairs—Republican Alfred Sikes and Democrat Tom Wheeler—have urged Carr to change course.

    Carr has multiple probes in progress, and his investigation into CBS over the editing of an interview with Kamala Harris has drawn condemnations from both liberal and conservative advocacy groups that describe it as a threat to the Constitutional right to free speech. One plea to drop the investigation came in a March 19 letter from conservative groups including the Center for Individual Freedom, Grover Norquist’s Americans for Tax Reform, and the Taxpayers Protection Alliance.

    “While we understand the concerns that motivate the complaint, we nonetheless fear that an adverse ruling against CBS would constitute regulatory overreach and advance precedent that can be weaponized by future FCCs,” the letter said. The letter argued that “Democrats and leftwing activist groups have repeatedly worked to weaponize” the government against free speech and that the FCC should “help guard against future abuses by Democrats and leftwing organizations by streamlining license renewals and merger reviews and eliminating the news distortion and news hoax rules.”

    “The flimsiest of complaints”

    Andrew Jay Schwartzman, an expert on media law and senior counselor for the Benton Institute for Broadband & Society, told Ars that “the CBS complaint is utterly lacking in merit. What is alleged doesn’t come within light-years of a violation of any FCC policy.”

    The Foundation for Individual Rights and Expression (FIRE), an advocacy group, called Carr’s investigation of CBS “a political stunt,” an “illegitimate show trial,” and an “unconstitutional abuse of regulatory authority.” Democratic lawmakers are demanding answers from Carr about what they call “bogus investigations” designed to “target and intimidate news organizations and broadcasters in violation of the First Amendment.”

    The CBS investigation was also lambasted in comments submitted by Christopher Terry, a professor of media law and ethics at the University of Minnesota, and J. Israel Balderas, a journalism professor at Elon University who is also a First Amendment attorney and a former FCC media advisor.

    “The agency under Brendan Carr appears to be, based on the flimsiest of complaints, pursuing media outlets critical of Donald Trump during the 2024 campaign, while ignoring similar complaints from the public about Trump-friendly media outlets,” Terry and Balderas wrote. “Being the speech police is not the FCC’s job, but enforcing any restrictions in a selective, much less a partisan, way is problematic, and likely to lead to extensive legal actions challenging FCC authority.”

    FCC’s long shift away from news regulation

    The FCC has historically regulated broadcast news with the Fairness Doctrine, which no longer exists, and the news distortion policy, which is still in place. The Fairness Doctrine was introduced in 1949 to guarantee “that the public has a reasonable opportunity to hear different opposing positions on the public issues of interest and importance in the community.” This requirement to air contrasting views remained in place until 1987.

    After losing a court case brought by a TV station, the FCC was forced to reconsider its enforcement of the Fairness Doctrine and decided to repeal it. The Reagan-era FCC concluded that the Fairness Doctrine “violates the First Amendment” and works against the public interest. “Despite the physical differences between the electronic and print media, their roles in our society are identical, and we believe that the same First Amendment principles should be equally applicable to both,” the FCC said at the time.

    US regulation of broadcast news continued to be lessened through a series of commission decisions and court rulings. “Even the relaxation of non-content regulations, such as the extension of stations’ license terms from three to eight years, and adoption of rules that make challenges to license renewals by the public or potential competitors almost impossible, have bolstered broadcasters’ editorial rights against outside review,” said a 2001 article by Santa Clara University professor Chad Raphael in the journal Communication Law and Policy.

    The FCC’s general shift away from regulating news content made it surprising that the news distortion policy survived, Raphael wrote. “Given this deregulatory trend, it is remarkable that the Commission has preserved its little-known rules against licensees’ deliberately distorting the news… The distortion rules have drawn scant commentary in the regulatory literature, especially in contrast to the outpouring of debate over their cousin, the Fairness Doctrine,” the article said.

    But the FCC never issued many findings of news distortion, and such findings have been nearly nonexistent in recent decades. Raphael’s analysis found 120 decisions on news distortion between 1969 and 1999, and only 12 of them resulted in findings against broadcasters. Those 12 decisions were generated by eight cases, as several of the cases “generated multiple decisions as they went through the appeals process.”

    “The number of reported decisions drops off dramatically after 1976, and there is only one finding of distortion after 1982, when the Reagan-era FCC began to remove content regulations on broadcast news,” Raphael wrote. The one post-1982 finding of distortion was issued in a letter of admonishment to NBC in 1993 “for staging a segment of a Dateline NBC report on unsafe gas tanks in General Motors trucks,” Raphael wrote.

    GM investigated the incident and NBC “admitted to staging the explosion, made an on-air apology to GM, fired three producers who contributed to the segment, and eventually dismissed its news president,” he wrote. The FCC itself sent the letter quietly, with “the first mention of this action appearing in a 1999 decision rejecting a challenge to NBC’s license renewals.”

    Investigations rare, penalties even rarer

    The rare findings of news distortion were usually accompanied by other infractions. “Most penalties consisted of issuing letters of admonishment or censure that did not figure heavily in subsequent license renewals, all of which were successful,” Raphael wrote.

    Despite Raphael’s paper being nearly a quarter-century old, it’s practically up to date. “Since the time of Raphael’s study, it appears that the Commission has only considered allegations of news distortion in a very small number of cases,” said a 2019 paper by Joel Timmer, a professor of film, television, and digital media at Texas Christian University.

    Timmer found eight post-1999 cases in which news distortion allegations were considered. Most of the allegations didn’t get very far, and none of them resulted in a finding of news distortion.

    The FCC technically has no rule or regulation against news distortion. “Instead, it has a news distortion policy, developed ‘through the adjudicatory process in decisions resolving challenges to broadcasters’ licenses,'” Timmer wrote.

    The FCC dismissed an allegation of news distortion over broadcast networks incorrectly projecting that Al Gore would win Florida in the 2000 presidential election, he wrote. The FCC said the incorrect projections were “not a sufficient basis to initiate such an investigation.”

    The FCC did investigate an allegation of news distortion in 2007. Two reporters at Florida station WTVT alleged a violation when their employer failed to air reports on the use of synthetic bovine growth hormone by dairy farmers. “The reporters alleged that station management and ownership demanded changes in their report as a result of pressure from Monsanto, the company that produces BGH,” but the FCC decided it was “a legitimate editorial dispute” and not “a deliberate effort to coerce [the reporters] into distorting the news,” Timmer wrote.

    There was also a 2007 case involving a Detroit TV station’s report “that a local official and several prominent local business people consorted with prostitutes during a fishing trip to Costa Rica,” Timmer wrote. “It was alleged that a reporter from WXYZ-TV actually paid prostitutes to stay at the hotel at which the trip’s participants were staying, then falsely reported that the participants consorted with them. While the FCC acknowledged that, if true, this could constitute staging of the news, there was a lack of extrinsic evidence to establish that the licensee, its top management, or its news management were involved in an attempt to deliberately distort or falsify the news, causing the news distortion claim to fail.”

    Timmer’s paper summarized the FCC’s post-1999 news distortion enforcement as follows:

    In addition to the post-1999 cases already discussed—those involving reporting on bovine growth hormone, erroneous projections that Al Gore would win Florida in the 2000 presidential election—and reporting regarding prostitutes in Costa Rica with a public official and business people—charges of news distortion were raised and discussed in only a handful of instances. In addition to these three cases, there were five other cases since 1999 in which the Commission considered allegations of news distortion. In only two of the eight cases was there any detailed discussion of news distortion claims: the BGH story and the story involving prostitutes in Costa Rica. Significantly, in none of the cases was news distortion found to have occurred.

    Terry told Ars that he’s not aware of any news distortion findings since the 2019 paper.

    The FCC has a separate broadcast hoax rule enacted in 1992. As of 2000, “no broadcaster had ever been fined pursuant to the rule, nor had any stations lost their licenses for violating the rule,” and “it appears that the FCC has considered allegations of broadcast hoaxes only three times since 2000, with none of those cases resulting in the FCC finding a violation of the rule,” Timmer wrote.

    The 60 Minutes investigation

    In one of her last official acts before Trump’s inauguration and her departure from the FCC, Rosenworcel dismissed complaints of bias against Trump related to ABC’s fact-checking during a presidential debate, the editing of a CBS 60 Minutes interview with Harris, and NBC putting Harris on a Saturday Night Live episode. Rosenworcel also dismissed a challenge to a Fox station license alleging that Fox willfully distorted news with false reports of fraud in the 2020 election that Trump lost.

    Carr quickly revived the three complaints alleging bias against Trump, which were filed by a nonprofit law firm called the Center for American Rights. Of these, the ABC and CBS complaints allege news distortion. The NBC complaint alleges a violation of the separate Equal Time rule. The complaints were filed against individual broadcast stations because the FCC licenses stations rather than the networks that own them or are affiliated with them.

    Carr has repeatedly expressed interest in the complaint over 60 Minutes, which alleged that CBS misled viewers by airing two different responses to the same question about Israeli Prime Minister Benjamin Netanyahu, one on 60 Minutes and the other on Face the Nation. CBS’s defense—which is supported by the unedited transcript and video of the interview—is that the two clips show different parts of the same answer given by Harris.

    On February 5, the Carr-led FCC issued a public notice seeking comment on the CBS investigation. The FCC’s public notices aren’t generally seen by many people, but the FCC tried to encourage participation in this proceeding. The agency temporarily added a banner message to the top of the consumer complaints page to urge the public to submit comments about the 60 Minutes interview.

    “Interested in adding your comments to the proceeding investigating news distortion in the airing of a ’60 Minutes’ interview with then Vice President Kamala Harris?” the banner message said, linking to a page that explained how to submit comments on the proceeding.

    Former chairs blast Carr

    One filing was submitted by the former chairs Sikes and Wheeler, plus three other former FCC commissioners: Republican Rachelle Chong, Democrat Ervin Duggan, and Democrat Gloria Tristani. “These comments are submitted to emphasize the unprecedented nature of this news distortion proceeding, and to express our strong concern that the Federal Communications Commission may be seeking to censor the news media in a manner antithetical to the First Amendment,” the bipartisan group of former FCC chairs and commissioners wrote.

    The FCC has historically “enforced the [news distortion] policy very rarely, and it has adopted guardrails requiring that complaints be summarily dismissed in all but the most exceptional circumstances,” they wrote, adding that there are no exceptional circumstances warranting an investigation into CBS.

    “The Commission’s departures from its typical practice and precedent are especially troubling when viewed in context. This Administration has made no secret of its desire to revoke the licenses of broadcasters that cover it in ways the President considers unfavorable,” the filing said.

    Pointing to the Raphael and Timmer analyses, the former FCC leaders wrote that the agency “issued findings of liability on news distortion in just eight cases between 1969 and 2019—and in fact in just one case between 1985 and 2019. None of the cases that found news distortion concerned the way a broadcaster had exercised its editorial discretion in presenting the news. Instead, each case involved egregious misconduct, including the wholesale fabrication of news stories.”

    The FCC’s news distortion policy applies a multi-part test, the group noted. A finding of news distortion requires “deliberate distortion” and not mere inaccuracy or differences of opinion, “extrinsic evidence (i.e., beyond the broadcast itself) demonstrating that the broadcaster deliberately distorted or staged the news” and that “the distortion must apply to a ‘significant event,’ rather than minor inaccuracies or incidental aspects of the report.” Finally, FCC policy is to “only consider taking action on the broadcaster’s license if the extrinsic evidence shows the distortion involved the ‘principals, top management, or news management’ of the licensee, as opposed to other employees.”

    The FCC has historically punished licensees only after dramatic violations, like “elaborate hoaxes, internal conspiracies, and reports conjured from whole cloth,” they wrote. There is “no credible argument” that the allegations against CBS “belong in the same category.”

    CBS transcript and video supports network

    Kamal Harris smiles while sitting for a television interview.

    Kamala Harris on 60 Minutes.

    Credit: CBS

    Kamala Harris on 60 Minutes. Credit: CBS

    The Center for American Rights complaint says that an FCC investigation of”extrinsic evidence” could include examining outtakes to determine whether “the licensee has deliberately suppressed or altered a news report.” The complaint criticized CBS for not providing the complete transcript of the interview.

    In late January, the Carr-led FCC demanded that CBS provide an unedited transcript and camera feeds of the interview. CBS provided the requested materials and made them available publicly. The transcript supports CBS’s defense because it shows that what the Center for American Rights claimed were “two completely different answers” were just two different sentences from the same response.

    “We broadcast a longer portion of the vice president’s answer on Face the Nation and broadcast a shorter excerpt from the same answer on 60 Minutes the next day. Each excerpt reflects the substance of the vice president’s answer,” CBS said.

    The Center for American Rights complained that in one clip, Harris answered the question about Netanyahu by saying, “Well, Bill, the work that we have done has resulted in a number of movements in that region by Israel that were very much prompted by, or a result of many things, including our advocacy for what needs to happen in the region.”

    In the second clip, Harris responded to the question by saying, “We are not going to stop pursuing what is necessary for the United States to be clear about where we stand on the need for this war to end.”

    “Same interview, same question, two completely different answers,” the Center for American Rights’ complaint said.

    But the CBS transcript and video shows that Harris spoke these two sentences as part of one answer to the question. CBS aired the two sentences in different clips, but neither contradicts the other.

    Center for American Rights stands by complaint

    The Center for American Rights declined to comment on the transcript and video when contacted by Ars, but it pointed us to the final comments it submitted in the FCC proceeding. The filing argues for an expansive approach to regulating news distortion, saying that “slanting the news to benefit one political candidate violates the distortion doctrine.”

    “The core of our concern is that 60 Minutes‘ slice-and-dice journalism was an act of slanting the news to favor a preferred candidate and part of a pattern of CBS News consistently favoring a candidate and party… The Commission is uniquely positioned as the relevant authority with the power to investigate to determine whether CBS engaged in intentional news slanting,” the filing said.

    The Center for American Rights filing also complained that “Fox and Sinclair [we]re subject to relentless regulatory pressure under the prior chair… but then everyone screams that the First Amendment is being eviscerated when CBS is subject to attention under the same policy from the new chair.”

    “‘Selective enforcement’ is when Fox and Sinclair are constantly under regulatory pressure from Democrats at the FCC and in the Congress and from their outside allies, but then unchecked ‘press freedom’ is the sacrosanct principle when CBS allegedly transgresses the same lines when Republicans are in power,” the group said, responding to arguments that punishing CBS would be selective enforcement.

    As previously mentioned in this article, Rosenworcel rejected a news distortion complaint and license challenge that targeted Fox’s WTXF-TV in Philadelphia. “Such content review in the context of a renewal application would run afoul of our obligations under the First Amendment and the statutory prohibition on censorship and interference with free speech rights,” Rosenworcel’s FCC said.

    The conservative Sinclair Broadcasting Group was fined $48 million for portraying sponsored TV segments as news coverage and other violations in the largest-ever civil penalty paid by a broadcaster in FCC history. But that happened under Republican Ajit Pai, the FCC chair during Trump’s first term. Pai’s FCC also blocked Sinclair’s attempt to buy Tribune Media Company.

    Carr defended his investigation of CBS in a letter to Sen. Richard Blumenthal (D-Conn.). “During the Biden Administration, the FCC and Democrats across government repeatedly weaponized our country’s communications laws and processes. In contrast, I am restoring the FCC’s commitment to basic fairness and even-handed treatment for all,” Carr wrote.

    Carr said he “put the CBS complaint on the same procedural footing that the Biden FCC determined it should apply to the Fox complaint.” By this, he means that the previous administration held a proceeding to consider the Fox complaint instead of dismissing it outright.

    “The Biden FCC’s approach to the Fox petition stands in stark contrast to the approach the Biden FCC took to the CBS petition. Unlike the Fox petition, the Biden FCC just summarily dismissed the CBS one,” Carr wrote. Carr also said the Biden-era FCC “fail[ed] to process hundreds of routine Sinclair license renewals” and that the FCC is now “clearing and renewing those licenses again.”

    The Fox case involved very different allegations than the CBS one. While CBS is facing investigation for airing two parts of an interviewee’s answer in two different broadcasts, a Delaware judge ruled in 2023 that Fox News made false and defamatory statements claiming that Dominion Voting Systems committed election fraud by manipulating vote counts through its software and algorithms. Fox subsequently agreed to pay Dominion $788 million in a settlement instead of facing trial.

    Carr could test FCC authority in court

    The Rosenworcel FCC said the CBS complaint was meritless in its dismissal. “Opening a news distortion enforcement action under Commission precedent—as rare as it is—turns on the important question of whether any information or extrinsic evidence was submitted to the Commission indicating an ‘intentional’ or ‘deliberate’ falsification of the news,” the decision said. “The Complaint submitted fails to do so. The Commission simply cannot wield its regulatory authority in a manner completely inconsistent with long-settled precedent that the Commission not ‘second guess’ broadcast decisions.”

    The comments submitted by former chairs and commissioners said the “transcript confirms that the editing choices at issue lie well within the editorial judgment protected by the First Amendment.” TechFreedom, a libertarian-leaning think tank, told the FCC that “if the new standard for triggering a news distortion analysis is that any edits of raw interview video can be subject to challenge, then the FCC will spend the next four years, at least, fielding dozens, hundreds, thousands of news distortion complaints. Since every taped interview is edited, every taped interview that is aired will be ripe for an FCC complaint, which will have to be adjudicated. The news distortion complaint process will be weaponized by both political parties, and the business of the FCC will grind to a halt as it will have to assign more and more FTEs [full-time employees] to processing these complaints.”

    Although CBS appears to have a strong defense, Carr can make life difficult for broadcasters simply by opening investigations. As experts have previously told Ars, the FCC can use its rules to harass licensees and hold up applications related to business deals. Carr said in November that the news distortion complaint over the 60 Minutes interview would factor into the FCC’s review of CBS owner Paramount’s transfer of TV broadcast station licenses to Skydance.

    Jeffrey Westling, a lawyer who is the director of technology and innovation policy at the conservative American Action Forum, has written that the high legal bar for proving news distortion means that cases must involve something egregious—like a bribe or instructions from management to distort the news. But Westling has told Ars it’s possible that a “sympathetic” court could let the FCC use the rule to deny a transfer or renewal of a broadcast license.

    “The actual bounds of the rule are not well-tested,” said Westling, who argues that the news distortion policy should be eliminated.

    An FCC webpage that was last updated during Rosenworcel’s term says the FCC’s authority to enforce its news distortion policy is narrow. “The agency is prohibited by law from engaging in censorship or infringing on First Amendment rights of the press,” the FCC said, noting that “opinion or errors stemming from mistakes are not actionable.”

    1960s FCC: “No government agency can authenticate the news”

    The high bar set by the news distortion policy isn’t just about issuing findings of distortion—it is supposed to prevent many investigations in the first place, the Rosenworcel FCC said in its dismissal of the CBS complaint:

    Indeed, the Commission has established a high threshold to commencing any investigation into allegations of news distortion. It is not sufficient for the Complainant to show that the material in question is false or even that the Licensee might have known or should have known about the falsity of the material. A news distortion complaint must include extrinsic evidence that the Licensee took actions to engage in a deliberate and intentional falsification of the news.

    The comments submitted by Terry and Balderas said that “case law is clear: news distortion complaints must meet an extraordinary burden of proof.”

    “The current complaint against CBS fails to meet this standard,” Terry and Balderas wrote. “Editing for clarity, brevity, or production value is a standard journalistic practice, and absent clear evidence of deliberate fabrication, government intervention is unwarranted. The current complaint against CBS presents no extrinsic evidence whatsoever—no internal memos, no whistleblower testimony, no evidence of financial incentives—making it facially deficient under the extrinsic evidence standard consistently applied since Hunger in America.”

    Hunger in America was a 1968 CBS documentary that the FCC investigated. The FCC’s decision against issuing a finding of news distortion became an important precedent that was cited in a 1985 court case that upheld another FCC decision to reject an allegation of news distortion.

    “The FCC’s policy on rigging, staging, or distorting the news was developed in a series of cases beginning in 1969,” said the 1985 ruling from the US Court of Appeals for the District of Columbia Circuit. “In the first of these, Hunger In America, CBS had shown an infant it said was suffering from malnutrition, but who was actually suffering from another ailment.”

    The 1960s FCC found that “[r]igging or slanting the news is a most heinous act against the public interest” but also that “in this democracy, no government agency can authenticate the news, or should try to do so.” As the DC Circuit Court noted, in Hunger in America and “in all the subsequent cases, the FCC made a crucial distinction between deliberate distortion and mere inaccuracy or difference of opinion.”

    Carr: FCC “not close” to dismissing complaint

    Despite this history of non-enforcement except in the most egregious cases, Carr doesn’t seem inclined to end the investigation into what seems to be a routine editing decision. “Carr believes CBS has done nothing to bring the commission’s investigation to an end, including a fix for the alleged pervasive bias in its programming, according to people with knowledge of the matter,” said a New York Post report on March 28.

    The report said the Paramount/Skydance merger “remains in FCC purgatory” and that the news distortion investigation is “a key element” holding up FCC approval of the transaction. An anonymous FCC official was quoted as saying that “the case isn’t close to being settled right now.”

    We contacted Carr and will update this article if we get a response. But Carr confirmed to another news organization recently that he doesn’t expect a quick resolution. He told Reuters on March 25 that “we’re not close in my view to the position of dismissing that complaint at this point.”

    Photo of Jon Brodkin

    Jon is a Senior IT Reporter for Ars Technica. He covers the telecom industry, Federal Communications Commission rulemakings, broadband consumer affairs, court cases, and government regulation of the tech industry.

    The speech police: Chairman Brendan Carr and the FCC’s news distortion policy Read More »

    trump-tariffs-terrify-board-game-designers

    Trump tariffs terrify board game designers

    Placko called the new policy “not just a policy change” but “a seismic shift.”

    Rob Daviau, who helps run Restoration Games and designed hit games like Pandemic Legacy, has been writing on social media for months about the fact that every meeting he’s in “has been an existential crisis about our industry.”

    Expanding on his remarks in an interview with BoardGameWire late last year, Daviau added that he was a natural pessimist who foresaw a “great collapse in the hobby gaming market in the US” if tariffs were implemented.

    Gamers aren’t likely to stop playing, but they might stick with their back catalog (gamers are notorious for having “shelves of shame” featuring hot new games they purchased without playing them… because other hot new games had already appeared). Or they might, in search of a better deal, shop only online, which could be tough on already struggling local game stores. Or games might decline in quality to keep costs lower. None of which is likely to lead to a robust, high-quality board gaming ecosystem.

    Stegmaier’s forecast is nearly as dark as Daviau’s. “Within a few months US companies will lose a lot of money and/or go out of business,” he wrote, “and US citizens will suffer from extreme inflation.”

    The new tariffs can be avoided by shipping directly from the factories to firms in other countries, such as a European distributor, but the US remains a crucial market for US game makers; Stegmaier notes that “65 percent of our sales are in the US, so this will take a heavy toll.”

    For games still in the production pipeline, at least budgetary adjustments can be made, but some games have already been planned, produced, and shipped. If the boat arrives after the tariffs go into effect—too bad. The US importer still has to pay the extra fees. Chris Solis, who runs Solis Game Studio in California, issued an angry statement yesterday covering exactly this situation, saying, “I have 8,000 games leaving a factory in China this week and now need to scramble to cover the import bill.”

    GAMA, the trade group for board game publishers, has been lobbying against the new tariffs, but with little apparent success thus far.

    Trump tariffs terrify board game designers Read More »

    we-have-the-first-video-of-a-plant-cell-wall-being-built

    We have the first video of a plant cell wall being built

    Plant cells are surrounded by an intricately structured protective coat called the cell wall. It’s built of cellulose microfibrils intertwined with polysaccharides like hemicellulose or pectin. We have known what plant cells look like without their walls, and we know what they look like when the walls are fully assembled, but we’ve never seen the wall-building process in action. “We knew the starting point and the finishing point, but had no idea what happens in between,” says Eric Lam, a plant biologist at Rutgers University. He’s a co-author of the study that caught wall-building plant cells in action for the first time. And once we saw how the cell wall building worked, it looked nothing like how we drew that in biology handbooks.

    Camera-shy builders

    Plant cells without walls, known as protoplasts, are very fragile, and it has been difficult to keep them alive under a microscope for the several hours needed for them to build walls. Plant cells are also very light-sensitive, and most microscopy techniques require pointing a strong light source at them to get good imagery.

    Then there was the issue of tracking their progress. “Cellulose is not fluorescent, so you can’t see it with traditional microscopy,” says Shishir Chundawat, a biologist at Rutgers. “That was one of the biggest issues in the past.” The only way you can see it is if you attach a fluorescent marker to it. Unfortunately, the markers typically used to label cellulose were either bound to other compounds or were toxic to the plant cells. Given their fragility and light sensitivity, the cells simply couldn’t survive very long with toxic markers as well.

    We have the first video of a plant cell wall being built Read More »

    gemini-“coming-together-in-really-awesome-ways,”-google-says-after-2.5-pro-release

    Gemini “coming together in really awesome ways,” Google says after 2.5 Pro release


    Google’s Tulsee Doshi talks vibes and efficiency in Gemini 2.5 Pro.

    Google was caught flat-footed by the sudden skyrocketing interest in generative AI despite its role in developing the underlying technology. This prompted the company to refocus its considerable resources on catching up to OpenAI. Since then, we’ve seen the detail-flubbing Bard and numerous versions of the multimodal Gemini models. While Gemini has struggled to make progress in benchmarks and user experience, that could be changing with the new 2.5 Pro (Experimental) release. With big gains in benchmarks and vibes, this might be the first Google model that can make a dent in ChatGPT’s dominance.

    We recently spoke to Google’s Tulsee Doshi, director of product management for Gemini, to talk about the process of releasing Gemini 2.5, as well as where Google’s AI models are going in the future.

    Welcome to the vibes era

    Google may have had a slow start in building generative AI products, but the Gemini team has picked up the pace in recent months. The company released Gemini 2.0 in December, showing a modest improvement over the 1.5 branch. It only took three months to reach 2.5, meaning Gemini 2.0 Pro wasn’t even out of the experimental stage yet. To hear Doshi tell it, this was the result of Google’s long-term investments in Gemini.

    “A big part of it is honestly that a lot of the pieces and the fundamentals we’ve been building are now coming together in really awesome ways, ” Doshi said. “And so we feel like we’re able to pick up the pace here.”

    The process of releasing a new model involves testing a lot of candidates. According to Doshi, Google takes a multilayered approach to inspecting those models, starting with benchmarks. “We have a set of evals, both external academic benchmarks as well as internal evals that we created for use cases that we care about,” she said.

    Credit: Google

    The team also uses these tests to work on safety, which, as Google points out at every given opportunity, is still a core part of how it develops Gemini. Doshi noted that making a model safe and ready for wide release involves adversarial testing and lots of hands-on time.

    But we can’t forget the vibes, which have become an increasingly important part of AI models. There’s great focus on the vibe of outputs—how engaging and useful they are. There’s also the emerging trend of vibe coding, in which you use AI prompts to build things instead of typing the code yourself. For the Gemini team, these concepts are connected. The team uses product and user feedback to understand the “vibes” of the output, be that code or just an answer to a question.

    Google has noted on a few occasions that Gemini 2.5 is at the top of the LM Arena leaderboard, which shows that people who have used the model prefer the output by a considerable margin—it has good vibes. That’s certainly a positive place for Gemini to be after a long climb, but there is some concern in the field that too much emphasis on vibes could push us toward models that make us feel good regardless of whether the output is good, a property known as sycophancy.

    If the Gemini team has concerns about feel-good models, they’re not letting it show. Doshi mentioned the team’s focus on code generation, which she noted can be optimized for “delightful experiences” without stoking the user’s ego. “I think about vibe less as a certain type of personality trait that we’re trying to work towards,” Doshi said.

    Hallucinations are another area of concern with generative AI models. Google has had plenty of embarrassing experiences with Gemini and Bard making things up, but the Gemini team believes they’re on the right path. Gemini 2.5 apparently has set a high-water mark in the team’s factuality metrics. But will hallucinations ever be reduced to the point we can fully trust the AI? No comment on that front.

    Don’t overthink it

    Perhaps the most interesting thing you’ll notice when using Gemini 2.5 is that it’s very fast compared to other models that use simulated reasoning. Google says it’s building this “thinking” capability into all of its models going forward, which should lead to improved outputs. The expansion of reasoning in large language models in 2024 resulted in a noticeable improvement in the quality of these tools. It also made them even more expensive to run, exacerbating an already serious problem with generative AI.

    The larger and more complex an LLM becomes, the more expensive it is to run. Google hasn’t released technical data like parameter count on its newer models—you’ll have to go back to the 1.5 branch to get that kind of detail. However, Doshi explained that Gemini 2.5 is not a substantially larger model than Google’s last iteration, calling it “comparable” in size to 2.0.

    Gemini 2.5 is more efficient in one key area: the chain of thought. It’s Google’s first public model to support a feature called Dynamic Thinking, which allows the model to modulate the amount of reasoning that goes into an output. This is just the first step, though.

    “I think right now, the 2.5 Pro model we ship still does overthink for simpler prompts in a way that we’re hoping to continue to improve,” Doshi said. “So one big area we are investing in is Dynamic Thinking as a way to get towards our [general availability] version of 2.5 Pro where it thinks even less for simpler prompts.”

    Gemini models on phone

    Credit: Ryan Whitwam

    Google doesn’t break out earnings from its new AI ventures, but we can safely assume there’s no profit to be had. No one has managed to turn these huge LLMs into a viable business yet. OpenAI, which has the largest user base with ChatGPT, loses money even on the users paying for its $200 Pro plan. Google is planning to spend $75 billion on AI infrastructure in 2025, so it will be crucial to make the most of this very expensive hardware. Building models that don’t waste cycles on overthinking “Hi, how are you?” could be a big help.

    Missing technical details

    Google plays it close to the chest with Gemini, but the 2.5 Pro release has offered more insight into where the company plans to go than ever before. To really understand this model, though, we’ll need to see the technical report. Google last released such a document for Gemini 1.5. We still haven’t seen the 2.0 version, and we may never see that document now that 2.5 has supplanted 2.0.

    Doshi notes that 2.5 Pro is still an experimental model. So, don’t expect full evaluation reports to happen right away. A Google spokesperson clarified that a full technical evaluation report on the 2.5 branch is planned, but there is no firm timeline. Google hasn’t even released updated model cards for Gemini 2.0, let alone 2.5. These documents are brief one-page summaries of a model’s training, intended use, evaluation data, and more. They’re essentially LLM nutrition labels. It’s much less detailed than a technical report, but it’s better than nothing. Google confirms model cards are on the way for Gemini 2.0 and 2.5.

    Given the recent rapid pace of releases, it’s possible Gemini 2.5 Pro could be rolling out more widely around Google I/O in May. We certainly hope Google has more details when the 2.5 branch expands. As Gemini development picks up steam, transparency shouldn’t fall by the wayside.

    Photo of Ryan Whitwam

    Ryan Whitwam is a senior technology reporter at Ars Technica, covering the ways Google, AI, and mobile technology continue to change the world. Over his 20-year career, he’s written for Android Police, ExtremeTech, Wirecutter, NY Times, and more. He has reviewed more phones than most people will ever own. You can follow him on Bluesky, where you will see photos of his dozens of mechanical keyboards.

    Gemini “coming together in really awesome ways,” Google says after 2.5 Pro release Read More »

    deepmind-has-detailed-all-the-ways-agi-could-wreck-the-world

    DeepMind has detailed all the ways AGI could wreck the world

    As AI hype permeates the Internet, tech and business leaders are already looking toward the next step. AGI, or artificial general intelligence, refers to a machine with human-like intelligence and capabilities. If today’s AI systems are on a path to AGI, we will need new approaches to ensure such a machine doesn’t work against human interests.

    Unfortunately, we don’t have anything as elegant as Isaac Asimov’s Three Laws of Robotics. Researchers at DeepMind have been working on this problem and have released a new technical paper (PDF) that explains how to develop AGI safely, which you can download at your convenience.

    It contains a huge amount of detail, clocking in at 108 pages before references. While some in the AI field believe AGI is a pipe dream, the authors of the DeepMind paper project that it could happen by 2030. With that in mind, they aimed to understand the risks of a human-like synthetic intelligence, which they acknowledge could lead to “severe harm.”

    All the ways AGI could harm humanity

    This work has identified four possible types of AGI risk, along with suggestions on how we might ameliorate said risks. The DeepMind team, led by company co-founder Shane Legg, categorized the negative AGI outcomes as misuse, misalignment, mistakes, and structural risks. Misuse and misalignment are discussed in the paper at length, but the latter two are only covered briefly.

    table of AGI risks

    The four categories of AGI risk, as determined by DeepMind.

    Credit: Google DeepMind

    The four categories of AGI risk, as determined by DeepMind. Credit: Google DeepMind

    The first possible issue, misuse, is fundamentally similar to current AI risks. However, because AGI will be more powerful by definition, the damage it could do is much greater. A ne’er-do-well with access to AGI could misuse the system to do harm, for example, by asking the system to identify and exploit zero-day vulnerabilities or create a designer virus that could be used as a bioweapon.

    DeepMind has detailed all the ways AGI could wreck the world Read More »

    first-party-switch-2-games—including-re-releases—all-run-either-$70-or-$80

    First-party Switch 2 games—including re-releases—all run either $70 or $80

    Not all game releases will follow Nintendo’s pricing formula. The Switch 2 release of Street Fighter 6 Year 1-2 Fighters Edition retails for $60, and Square Enix’s remastered Bravely Default is going for $40, the exact same price the 3DS version launched for over a decade ago.

    Game-Key cards have clearly labeled cases to tell you that the cards don’t actually hold game content. Credit: Nintendo/Square Enix

    One possible complicating factor for those games? While they’re physical releases, they use Nintendo’s new Game-Key Card format, which attempts to split the difference between true physical copies of a game and download codes. Each cartridge includes a key for the game, but no actual game content—the game itself is downloaded to your system at first launch. But despite holding no game content, the key card must be inserted each time you launch the game, just like any other physical cartridge.

    These cards will presumably be freely shareable and sellable just like regular physical Switch releases, but because they hold no actual game data, they’re cheaper to manufacture. It’s possible that some of these savings are being passed on to the consumer, though we’ll need to see more examples to know for sure.

    What about Switch 2 Edition upgrades?

    The big question mark is how expensive the Switch 2 Edition game upgrades will be for Switch games you already own, and what the price gap (if any) will be between games like Metroid Prime 4 or Pokémon Legends: Z-A that are going to launch on both the original Switch and the Switch 2.

    But we can infer from Mario Kart and Donkey Kong that the pricing for these Switch 2 upgrades will most likely be somewhere in the $10 to $20 range—the difference between the $60 price of most first-party Switch releases and the $70-to-$80 price for the Switch 2 Editions currently listed at Wal-Mart. Sony charges a similar $10 fee to upgrade from the PS4 to the PS5 editions of games that will run on both consoles. If you can find copies of the original Switch games for less than $60, that could mean saving a bit of money on the Switch 2 Edition, relative to Nintendo’s $70 and $80 retail prices.

    Nintendo will also use some Switch 2 Edition upgrades as a carrot to entice people to the more expensive $50-per-year tier of the Nintendo Switch Online service. The company has already announced that the upgrade packs for Breath of the Wild and Tears of the Kingdom will be offered for free to Nintendo Switch Online + Expansion Pack subscribers. The list of extra benefits for that service now includes additional emulated consoles (Game Boy, Game Boy Advance, Nintendo 64, and now Gamecube) and paid DLC for both Animal Crossing: New Horizons and Mario Kart 8.

    This story was updated at 7: 30pm on April 2nd to add more pricing information from US retailers about other early Switch 2 games.

    First-party Switch 2 games—including re-releases—all run either $70 or $80 Read More »