Author name: Mike M.

trump-suspends-trade-loophole-for-cheap-online-retailers-globally

Trump suspends trade loophole for cheap online retailers globally

But even Amazon may struggle to shift its supply chain as the de minimis exemption is eliminated for all countries. In February, the e-commerce giant “projected lower-than-expected sales and operating income for its first quarter,” which it partly attributed to “unpredictability in the economy.” A DataWeave study concluded at the end of June that “US prices for China-made goods on Amazon” were rising “faster than inflation,” Reuters reported, likely due to “cost shocks” currently “rippling through the retail supply chain.” Other non-Chinese firms likely impacted by this week’s order include eBay, Etsy, TikTok Shop, and Walmart.

Amazon did not respond to Ars’ request to comment but told Reuters last month that “it has not seen the average prices of products change up or down appreciably outside of typical fluctuations.”

Trump plans to permanently close loophole in 2027

Trump has called the de minimis exemption a “big scam,” claiming that it’s a “catastrophic loophole” used to “evade tariffs and funnel deadly synthetic opioids as well as other unsafe or below-market products that harm American workers and businesses into the United States.”

To address what Trump has deemed “national emergencies” hurting American trade and public health, he has urgently moved to suspend the loophole now and plans to permanently end it worldwide by July 1, 2027.

American travelers will still be able to “bring back up to $200 in personal items” and receive “bona fide gifts valued at $100 or less” duty-free, but a fixed tariff rate of between $80 to $200 per item will be applied to many direct-to-consumer shipments until Trump finishes negotiating trade deals with the rest of America’s key trade partners. As each deal is theoretically closed, any shipments will be taxed according to tariff rates of their country of origin. (Those negotiations are supposed to conclude by tomorrow, but so far, Trump has only struck deals with the European Union, Japan, and South Korea.)

Trump suspends trade loophole for cheap online retailers globally Read More »

samsung-galaxy-z-fold-7-review:-quantum-leap

Samsung Galaxy Z Fold 7 review: Quantum leap


A pretty phone for a pretty penny

Samsung’s new flagship foldable is a huge improvement over last year’s model.

Samsung Galaxy Z Fold 7 bent

Samsung’s new foldable is thinner and lighter than ever before. Credit: Ryan Whitwam

Samsung’s new foldable is thinner and lighter than ever before. Credit: Ryan Whitwam

The first foldable phones hit the market six years ago, and they were rife with compromises and shortcomings. Many of those problems have persisted, but little by little, foldables have gotten better. With the release of the Galaxy Z Fold 7, Samsung has made the biggest leap yet. This device solves some of the most glaring problems with Samsung’s foldables, featuring a new, slimmer design and a big camera upgrade.

Samsung’s seventh-generation foldable has finally crossed that hazy boundary between novelty and practicality, putting a tablet-sized screen in your pocket without as many compromises. There are still some drawbacks, of course, but for the first time, this feels like a foldable phone you’d want to carry around.

Whether or not you can justify the $1,999 price tag is another matter entirely.

Most improved foldable

Earlier foldable phones were pocket-busting bricks, but companies like Google, Huawei, and OnePlus have made headway streamlining the form factor—the Pixel 9 Pro Fold briefly held the title of thinnest foldable when it launched last year. Samsung, however, stuck with the same basic silhouette for versions one through six, shaving off a millimeter here and there with each new generation. Now, the Galaxy Z Fold 7 has successfully leapfrogged the competition with an almost unbelievably thin profile.

Specs at a glance: Samsung Galaxy Z Fold 7 – $1,999
SoC Snapdragon 8 Elite
Memory 12GB, 16GB
Storage 256GB, 512GB, 1TB
Display Cover: 6.5-inch 1080×2520 120 Hz OLED

Internal: 8-inch 1968×2184 120 Hz flexible OLED
Cameras 200MP primary, f/1.7, OIS; 10 MP telephoto, f/2.4, OIS; 12 MP ultrawide, f/2.2; 10 MP selfie cameras (internal and external), f/2.2
Software Android 16, 7 years of OS updates
Battery 4,400 mAh, 25 W wired charging, 15 W wireless charging
Connectivity Wi-Fi 7, NFC, Bluetooth 5.4, sub-6 GHz and mmWave 5G, USB-C 3.2
Measurements Folded: 158.4×72.8×8.9 mm

Unfolded: 158.4×143.2×4.2 mm

215 g

Clocking in at just 215 g and 8.9 mm thick when folded, the Z Fold 7 looks and feels like a regular smartphone when closed. It’s lighter than Samsung’s flagship flat phone, the Galaxy S25 Ultra, and is only a fraction of a millimeter thicker. The profile is now limited by the height of the standard USB-C port. You can use the Z Fold 7 in its closed state without feeling hindered by an overly narrow display or hand-stretching thickness.

Samsung Galaxy Z Fold 7 back

The Samsung Galaxy Z Fold 7 looks like any other smartphone at a glance.

Credit: Ryan Whitwam

The Samsung Galaxy Z Fold 7 looks like any other smartphone at a glance. Credit: Ryan Whitwam

It seems unreal at times, like this piece of hardware should be a tech demo or a dummy phone concept rather than Samsung’s newest mass-produced device. The only eyebrow-raising element of the folded profile is the camera module, which sticks out like a sore thumb.

To enable the thinner design, Samsung engineered a new hinge with a waterdrop fold. The gentler bend in the screen reduces the appearance of the middle crease and allows the two halves to close tightly with no gap. The opening and closing action retains the same precise feel as previous Samsung foldables. The frame is made from Samsung’s custom Armor Aluminum alloy, which promises greater durability than most other phones. It’s not titanium like the S25 Ultra or iPhone Pro models, but that saves a bit of weight.

Samsung Galaxy Z Fold 7 side

The Samsung Galaxy Z Fold 7 is almost impossibly thin, as long as you ignore the protruding camera module.

Credit: Ryan Whitwam

The Samsung Galaxy Z Fold 7 is almost impossibly thin, as long as you ignore the protruding camera module. Credit: Ryan Whitwam

There is one caveat to the design—the Z Fold 7 doesn’t open totally flat. It’s not as noticeable as Google’s first-gen Pixel Fold, but the phone stops a few degrees shy of perfection. It’s about on par with the OnePlus Open in that respect. You might notice this when first handling the Z Fold 7, but it’s easy to ignore, and it doesn’t affect the appearance of the internal flexible OLED.

The 6.5-inch cover display is no longer something you’d only use in a pinch when it’s impractical to open the phone. It has a standard 21:9 aspect ratio and tiny symmetrical bezels. Even reaching across from the hinge side is no problem (Google’s foldable still has extra chunk around the hinge). The OLED panel has the customary 120 Hz refresh rate and high brightness we’ve come to expect from Samsung. It doesn’t have the anti-reflective coating of the S25 Ultra, but it’s bright enough that you can use it outdoors without issue.

Samsung Galaxy Z Fold 7 open angle

The Z Fold 7 doesn’t quite open a full 180 degrees.

Credit: Ryan Whitwam

The Z Fold 7 doesn’t quite open a full 180 degrees. Credit: Ryan Whitwam

Naturally, the main event is inside: an 8-inch 120 Hz OLED panel at 1968×2184, which is slightly wider than last year’s phone. It’s essentially twice the size of the cover display, just like in Google’s last foldable. As mentioned above, the crease is almost imperceptible now. The screen feels solid under your fingers, but it still has a plastic cover that is vulnerable to damage—it’s even softer than fingernails. It’s very bright, but the plastic layer is more reflective than glass, which can make using it in harsh sunlight a bit of a pain.

Unfortunately, Samsung’s pursuit of thinness led it to drop support for the S Pen stylus. That was always a tough sell, as there was no place to store a stylus in the phone, and even Samsung’s bulky Z Fold cases struggled to accommodate the S Pen in a convenient way. Still, it’s sad to lose this unique feature.

The Z Fold 7 (right) cover display is finally free of compromise. Z Fold 6 on the left. Ryan Whitwam

Unlike some of the competition, Samsung has not added a dedicated AI button to this phone—although there’s plenty of AI here. You get the typical volume rocker on the right, with a power button below it. The power button also has a built-in fingerprint scanner, which is fast and accurate enough that we can’t complain. The buttons feel sturdy and give good feedback when pressed.

Android 16 under a pile of One UI and AI

The Galaxy Z Fold 7 and its smaller flippy sibling are the first phones to launch with Google’s latest version of Android, a milestone enabled by the realignment of the Android release schedule that began this year. The device also gets Samsung’s customary seven years of update support, a tie with Google for the best in the industry. However, updates arrive slower than they do on Google phones. If you’re already familiar with One UI, you’ll feel right at home on the Z Fold 7. It doesn’t reinvent the wheel, but there are a few enhancements.

Samsung Galaxy Z Fold 7 home screen

It’s like having a tablet in your pocket.

Credit: Ryan Whitwam

It’s like having a tablet in your pocket. Credit: Ryan Whitwam

Android 16 doesn’t include a ton of new features out of the box, and some of the upcoming changes won’t affect One UI. For example, Google’s vibrant Material 3 Expressive theme won’t displace the standard One UI design language when it rolls out later this summer, and Samsung already has its own app windowing implementation separate from Google’s planned release. The Z Fold 7 has a full version of Android’s new progress notifications at launch, something Google doesn’t even fully support in the initial release. Few apps have support, so the only way you’ll see those more prominent notifications is when playing media. These notifications also tie in to the Now Bar, which is at the core of Samsung’s Galaxy AI.

The Now Bar debuted on the S25 series earlier this year and uses on-device AI to process your data and present contextual information that is supposed to help you throughout the day. Samsung has expanded the apps and services that support the Now Bar and its constantly updating Now Brief, but we haven’t noticed much difference.

Samsung Galaxy Z Fold 7 Now Brief

Samsung’s AI-powered Now Brief still isn’t very useful, but it talks to you now. Umm, thanks?

Credit: Ryan Whitwam

Samsung’s AI-powered Now Brief still isn’t very useful, but it talks to you now. Umm, thanks? Credit: Ryan Whitwam

Nine times out of 10, the Now Bar doesn’t provide any useful notifications, and the Brief is quite repetitive. It often includes just weather, calendar appointments, and a couple of clickbait-y news stories and YouTube videos—this is the case even with all the possible data sources enabled. On a few occasions, the Now Bar correctly cited an appointment and suggested a route, but its timing was off by about 30 minutes. Google Now did this better a decade ago. Samsung has also added an AI-fueled audio version of the Now Brief, but we found this pretty tedious and unnecessary when there’s so little information in the report to begin with.

So the Now Bar is still a Now Bummer, but Galaxy AI also includes a cornucopia of other common AI features. It can rewrite text for you, summarize notes or webpages, do live translation, make generative edits to photos, remove background noise from videos, and more. These features work as well as they do on any other modern smartphone. Whether you get any benefit from them depends on how you use the phone.

However, we appreciate that Samsung included a toggle under the Galaxy AI settings to process data only on your device, eliminating the privacy concerns of using AI in the cloud. This reduces the number of operational AI features, but that may be a desirable feature all on its own.

Samsung Galaxy Z Fold 7 multitasking

You can’t beat Samsung’s multitasking system.

Credit: Ryan Whitwam

You can’t beat Samsung’s multitasking system. Credit: Ryan Whitwam

Samsung tends to overload its phones with apps and features. Those are here, too, making the Z Fold 7 a bit frustrating at times. Some of the latest One UI interface tweaks, like separating the quick settings and notifications, fall flat. Luckily, One UI is also quite customizable. For example, you can have your cover screen and foldable home screens mirrored like Pixels, or you can have a distinct layout for each mode. With some tweaking and removing pre-loaded apps, you can get the experience you want.

Samsung’s multitasking system also offers a lot of freedom. It’s quick to open apps in split-screen mode, move them around, and change the layout. You can run up to three apps side by side, and you can easily save and access those app groups later. Samsung also offers a robust floating window option, which goes beyond what Google has planned for Android generally—it has chosen to limit floating windows to tablets and projected desktop mode. Samsung’s powerful windowing system really helps unlock the productivity potential of a foldable.

The fastest foldable

Samsung makes its own mobile processors, but when speed matters, the company doesn’t mess around with Exynos. The Z Fold 7 has the same Snapdragon 8 Elite chip as the Galaxy S25 series, paired with 12GB of RAM and 256GB of storage in the model most people will buy. In our testing, this is among the most powerful smartphones on the market today, but it doesn’t quite reach the lofty heights of the Galaxy S25 Ultra, presumably due to its thermal design.

Samsung Galaxy Z Fold 7 in hand

The Z Fold 7 is much easier to hold than past foldables.

Credit: Ryan Whitwam

The Z Fold 7 is much easier to hold than past foldables. Credit: Ryan Whitwam

In Geekbench, the Galaxy Z Fold 7 lands between the Motorola Razr Ultra and the Galaxy S25 Ultra, both of which have Snapdragon 8 Elite chips. It far outpaces Google’s latest Pixel phones as well. The single-core CPU speed doesn’t quite match what you get from Apple’s latest custom iPhone processor, but the multicore numbers are consistently higher.

If mobile gaming is your bag, the Z Fold 7 will be a delight. Like other devices running on this platform, it puts up big scores. However, Samsung’s new foldable runs slightly behind some other 8 Elite phones. These are just benchmark numbers, though. In practice, the Z Fold 7 will handle any mobile game you throw at it.

Samsung Galaxy Z Fold 7 geekbench

The Fold 7 doesn’t quite catch the Z 25 Ultra.

Credit: Ryan Whitwam

The Fold 7 doesn’t quite catch the Z 25 Ultra. Credit: Ryan Whitwam

Samsung’s thermal throttling is often a concern, with some of its past phones with high-end Snapdragon chips shedding more than half their initial speed upon heating up. The Z Fold 7 doesn’t throttle quite that aggressively, but it’s not great, either. In our testing, an extended gaming session can see the phone slow down by about 40 percent. That said, even after heating up, the Z Fold 7 remains about 10 percent faster in games than the unthrottled Pixel 9 Pro. Qualcomm’s GPUs are just that speedy.

The CPU performance is affected by a much smaller margin under thermal stress, dropping only about 10–15 percent. That’s important because you’re more likely to utilize the Snapdragon 8 Elite’s power with Samsung’s robust multitasking system. Even when running three apps in frames with additional floating apps, we’ve noticed nary a stutter. And while 12GB of RAM is a bit shy of the 16GB you get in some gaming-oriented phones, it’s been enough to keep a day’s worth of apps in memory.

You also get about a day’s worth of usage from a charge. While foldables could generally use longer battery life, it’s impressive that Samsung made this year’s Z Fold so much thinner while maintaining the same 4,400 mAh battery capacity as last year’s phone. However, it’s possible to drain the device by early evening—it depends on how much you use the larger inner screen versus the cover display. A bit of battery anxiety is normal, but most days, we haven’t needed to plug it in before bedtime. A slightly bigger battery would be nice, but not at the expense of the thin profile.

The lack of faster charging is a bit more annoying. If you do need to recharge the Galaxy Z Fold 7 early, it will fill at a pokey maximum of 25 W. That’s not much faster than wireless charging, which can hit 15 W with a compatible charger. Samsung’s phones don’t typically have super-fast charging, with the S25 Ultra topping out at 45 W. However, Samsung hasn’t increased charging speeds for its foldables since the Z Fold 2. It’s long past time for an upgrade here.

Long-awaited camera upgrade

Camera hardware has been one of the lingering issues with foldables, which don’t have as much internal space to fit larger image sensors compared to flat phones. In the past, this has meant taking a big step down in image quality if you want your phone to fold in half. While Samsung has not fully replicated the capabilities of its flagship flat phones, the Galaxy Z Fold 7 takes a big step in the right direction with its protruding camera module.

Samsung Galaxy Z Fold 7 camera macro

The Z Fold 7’s camera has gotten a big upgrade.

Credit: Ryan Whitwam

The Z Fold 7’s camera has gotten a big upgrade. Credit: Ryan Whitwam

The camera setup is led by a 200 MP primary sensor with optical stabilization identical to the main shooter on the Galaxy S25 Ultra. It’s joined by a 12 MP ultrawide and 10 MP 3x telephoto, both a step down from the S25 Ultra. There is no equivalent to the 5x periscope telephoto lens on Samsung’s flat flagship. While it might be nice to have better secondary sensors, the 200 MP will get the most use, and it does offer better results than last year’s Z Fold.

Many of the photos we’ve taken on the Galaxy Z Fold 7 are virtually indistinguishable from those taken with the Galaxy S25 Ultra, which is mostly a good thing. The 200 MP primary sensor has a full-resolution mode, but you shouldn’t use it. With the default pixel binning, the Z Fold 7 produces brighter and more evenly exposed 12 MP images.

Samsung cameras emphasize vibrant colors and a wide dynamic range, so they lean toward longer exposures. Shooting with a Pixel and Galaxy phone side by side, Google’s cameras consistently use higher shutter speeds, making capturing motion easier. The Z Fold 7 is no slouch here, though. It will handle moving subjects in bright light better than any phone that isn’t a Pixel. Night mode produces bright images, but it takes longer to expose compared to Google’s offerings. Again, that means anything moving will end up looking blurry.

Between 1x and 3x, the phone uses digital zoom on the main sensor. When you go beyond that, it moves to the 3x telephoto (provided there is enough light). At the base 3x zoom, these photos are nice enough, with the usual amped-up colors and solid detail we’d expect from Samsung. However, the 10 MP resolution isn’t great if you push past 3x. Samsung’s image processing can’t sharpen photos to the same borderline magical degree as Google’s, and the Z Fold 7 can sometimes over-sharpen images in a way we don’t love. This is an area where the cheaper S25 Ultra still beats the new foldable, with higher-resolution backup cameras and multiple optical zoom levels.

At 12 MP, the ultrawide sensor is good enough for landscapes and group shots. It lacks optical stabilization (typical for ultrawide lenses), but it keeps autofocus. That allows you to take macro shots, and this mode activates automatically as you approach a subject. The images look surprisingly good with Samsung’s occasionally heavy-handed image processing, but don’t try to crop them down further.

The Z Fold 7 includes two in-display selfie cameras at 10 MP—one at the top of the cover display and the other for the inner foldable screen. Samsung has dispensed with its quirky under-display camera, which had a smattering of low-fi pixels covering it when not in use. The inner selfie is now just a regular hole punch, which is fine. You should really only use the front-facing cameras for video calls. If you want to take a selfie, foldables offer the option to use the more capable rear-facing cameras with the cover screen as a viewfinder.

A matter of coin

For the first time, the Galaxy Z Fold 7 feels like a viable alternative to a flat phone, at least in terms of hardware. The new design is as thin and light as many flat phones, and the cover display is large enough to do anything you’d do on non-foldable devices. Plus, you have a tablet-sized display on the inside with serious multitasking chops. We lament the loss of S Pen support, but it was probably necessary to address the chunkiness of past foldables.

Samsung Galaxy Z Fold 7 typing

The Samsung Galaxy Z Fold 7 is the next best thing to having a physical keyboard.

Credit: Ryan Whitwam

The Samsung Galaxy Z Fold 7 is the next best thing to having a physical keyboard. Credit: Ryan Whitwam

The camera upgrade was also a necessary advancement. You can’t ask people to pay a premium price for a foldable smartphone and offer a midrange camera setup. The 200 MP primary shooter is a solid upgrade over the cameras Samsung used in previous foldables, but the ultrawide and telephoto could still use some attention.

The price is one thing that hasn’t gotten better—in fact, it’s moving in the wrong direction. The Galaxy Z Fold 7 is even more expensive than last year’s model at a cool $2,000. As slick and capable as this phone is, the exorbitant price ensures tablet-style foldables remain a niche category. If that’s what it costs to make a foldable you’ll want to carry, flat phones won’t be usurped any time soon.

If you don’t mind spending two grand on a phone or can get a good deal with a trade-in or a carrier upgrade, you won’t regret the purchase. This is the most power that can fit in your pocket. It’s available directly from Samsung (in an exclusive Mint color), Amazon, Best Buy, and your preferred carrier.

Samsung Galaxy Z Fold 7 hinge macro

The Samsung Galaxy Z Fold 7 has a new, super-thin hinge design.

Credit: Ryan Whitwam

The Samsung Galaxy Z Fold 7 has a new, super-thin hinge design. Credit: Ryan Whitwam

The good

  • Incredibly slim profile and low weight
  • Upgraded 200 MP camera
  • Excellent OLED screens
  • Powerful multitasking capabilities
  • Toggle for local-only AI
  • Launches on Android 16 with seven years of update support

The bad

  • Ridiculously high price
  • Battery life and charging speed continue to be mediocre
  • One UI 8 has some redundant apps and clunky interface decisions
  • Now Brief still doesn’t do very much

Photo of Ryan Whitwam

Ryan Whitwam is a senior technology reporter at Ars Technica, covering the ways Google, AI, and mobile technology continue to change the world. Over his 20-year career, he’s written for Android Police, ExtremeTech, Wirecutter, NY Times, and more. He has reviewed more phones than most people will ever own. You can follow him on Bluesky, where you will see photos of his dozens of mechanical keyboards.

Samsung Galaxy Z Fold 7 review: Quantum leap Read More »

spilling-the-tea

Spilling the Tea

The Tea app is or at least was on fire, rapidly gaining lots of users. This opens up two discussions, one on the game theory and dynamics of Tea, one on its abysmal security.

It’s a little too on the nose that a hot new app that purports to exist so that women can anonymously seek out and spill the tea on men, which then puts user information into an unprotected dropbox thus spilling the tea on the identities of many of its users.

In the circles I follow this predictably led to discussions about how badly the app was coded and incorrect speculation that this was linked to vibe coding, whereas the dumb mistakes involved were in this case fully human.

There was also some discussion of the game theory of Tea, which I found considerably more interesting and fun, and which will take up the bulk of the post.

Tea offers a variety of services, while attempting to gate itself to only allow in women (or at least, not cis men), although working around this is clearly not hard if a man wanted to do that, and to only allow discussion and targeting of men.

Some of this is services like phone number lookup, social media and dating app search, reverse image internet search and criminal background checks. The photo you give is checked against catfishing databases. Those parts seem good.

There’s also generic dating advice and forums within the app, sure, fine.

The central feature is that you list a guy with a first name, location and picture – which given AI is pretty much enough for anyone these days to figure out who it is even if they don’t recognize them – and ask ‘are we dating the same guy?’ and about past experiences, divided into green and red flag posts. You can also set up alerts on guys in case there is any new tea.

What’s weird about ‘are we dating the same guy?’ is that the network effects required for that to work are very large, since you’re realistically counting on one or at most a handful of other people in the same position asking the same question. And if you do get the network big enough, search costs should then be very high, since reverse image search on a Facebook group is highly unreliable. It’s kind of amazing that the human recognition strategies people mostly use here worked at all in populated areas without proper systematization.

Tea provides much better search tools including notifications, which gives you a fighting chance, and one unified pool. But even with 4.6 million women, the chances of any given other woman being on it at all are not so high, and they then have to be an active user or have already left the note.

When I asked Claude about this it suggested the real win was finding Instagram or LinkedIn profiles, and that indeed makes a lot more sense. That’s good information, and it’s also voluntarily posted so it’s fair game.

Using a Hall of Shame seems even more inefficient. What, you are supposed to learn who the bad men are one by one? None of this seems like an effective use of time, even if you don’t have any ethical or accuracy concerns. This can’t be The Way, not outside of a small town or community.

The core good idea of the mechanics behind Tea is to give men Skin In The Game. The ideal amount of reputation that gets carried between interactions is not zero. The twin problems are that the ideal amount has an upper bound, and also that those providing that reputation also need Skin In The Game, gossip only works if there are consequences for spreading false gossip, and here those consequences seem absent.

What happens if someone lies or otherwise abuses the system? Everything is supposedly anonymous and works on negative selection. The app is very obviously ripe for abuse, all but made for attempts to sabotage or hurt people, using false or true information. A lot of what follows will be gaming that out.

The moderation team has a theoretical zero tolerance policy for defamation and harassment when evidence is provided, but such things are usually impossible to prove and the bar for actually violating the rules is high. Even if a violation is found and proof is possible, and the mod team would be willing to do something if proof was provided, if the target doesn’t know about the claims how can they respond?

Even then there don’t seem likely to be any consequences to the original poster.

Shall we now solve for the equilibrium, assuming the app isn’t sued into oblivion?

While tea is small and relatively unknown, the main worries (assuming the tools are accurate) are things like vindictive exes. There’s usually a reason when that happens, but there are going to be some rather nasty false positives.

As tea gets larger, it starts to change incentives in both good and bad ways, there are good reasons to start to try and manage, manipulate or fool the system, and things start to get weird. Threats and promises of actions within tea will loom in the air on every date and in every relationship. Indeed every interaction, essentially any woman (and realistically also any man) could threaten to spill tea, truthfully or otherwise, at any time.

Men will soon start asking for green flag posts, both accurate ones from exes and very much otherwise, services to do this will spring up, dummy accounts will be used where men are praising themselves.

Men will absolutely at minimum need to know what is being said, set up alerts on themselves, run all the background checks to see what will come up, and work to change the answer to that if it’s not what they want it to be. Presumably there will be plenty happy to sell you this service for very little, since half the population can provide such a service at very low marginal cost.

Quickly word of the rules of how to sculpt your background checks will spread.

And so on. It presumably will get very messy very quickly. The system simultaneously relies on sufficient network effects to make things like ‘are we dating the same guy?’ work, and devolves into chaos if usage gets too high.

One potential counterforce is that it would be pretty bad tea to have a reputation of trying to influence your tea. I doubt that ends up being enough.

At lunch, I told a woman that Tea exists and explained what it was.

Her: That should be illegal.

Her (10 seconds later): I need to go there and warn people about [an ex].

Her (a little later than that, paraphrased a bit): Actually no. He’s crazy, who knows what he might do if he found out.

Her (after I told her about the data breach): Oh I suppose I can’t use it then.

There is certainly an argument in many cases including this one for ‘[X] should be illegal but if it’s going to be legal then I should do it,’ and she clarified that her opposition was in particular to the image searches, although she quickly pointed out many other downsides as well.

The instinct is that all of this is bad for men.

That seems highly plausible but not automatic or obvious.

Many aspects of reputation and competition are positional goods and have zero-sum aspects in many of the places that Tea is threatening to cause trouble. Creating safer and better informed interactions and matches can be better for everything.

More generally, punishing defectors is by default very good for everyone, even if you are sad that it is now harder for you to defect. You’d rather be a good actor in the space, but previously in many ways ‘bad men drove out good’ placing pressure on you to not act well. This also that all this allows women to feel safe and let their guard down, and so on. A true ‘safety app’ is a good thing.

It could also motivate women to date more and use the apps more. It’s a better product when it is safer, far better, so you consume more of it. If no one has yet hooked the dating apps up automatically to tea so that you can get the tea as you swipe, well, get on that. Thus it can also act as a large improvement on matching. No, you don’t match directly on tea, but it provides a lot of information.

Another possible advantage is that receptivity to this could provide positive selection. If a woman doesn’t want to date you because of unverified internet comments, that is a red flag, especially for you in particular, in several ways at once. It means they probably weren’t that into you. It means they sought out and were swayed by the information. You plausibly dodged a bullet.

A final advantage is that this might be replacing systems that are less central and less reliable and that had worse enforcement mechanisms, including both groups and also things like whisper networks.

Consider the flip side, an app called No Tea, that men could use to effectively hide their pasts and reputations and information, without making it obvious they were doing this. Very obviously this would make even men net worse off if it went past some point.

As in, even from purely the man’s perspective: The correct amount of tea is not zero.

There are still six groups of ways I could think of that Tea could indeed be bad for men in general at current margins, as opposed to bad for men who deserve it, and it is not a good sign that over the days I wrote this the list kept growing.

  1. Men could in general find large utility in doing things that earn them very bad reputations on tea, and be forced to stop.

    1. This is highly unsympathetic, as they mostly involve things like cheating and otherwise treating women badly. I do not think those behaviors in general make men’s lives better, especially as a group.

    2. I also find it unlikely that men get large utility in absolute terms from such actions, rather than getting utility in relative terms. If you can get everyone to stop, I think most men win out here near current margins.

  2. Women could be bad at the Law of Conservation of Expected Evidence. As in, perhaps they update strongly negatively on negative information when they find it, but do not update appropriately positively when such information is not found, and do not adjust their calibration over time.

    1. This is reasonably marketed as a ‘safety app.’ If you are checked and come back clean, that should make you a lot safer and more trustworthy. That’s big.

      The existence of the app also updates expectations, if the men know that the app exists and that they could end up on it.

    2. In general, variance in response is your friend so long as the downside risk stops at a hard pass. You only need one yes, also you get favorable selection.

    3. Also, this could change the effective numerical dynamics. If a bunch of men become off limits due to tea, especially if that group often involves men who date multiple women at once, the numbers game can change a lot.

  3. Men could be forced to invest resources into reputation management in wasteful or harmful ways, and spend a lot of time being paranoid. This may favor men willing to game the system, or who can credibly threaten retaliation.

    1. This seems highly plausible, hopefully this is limited in scope.

    2. The threat of retaliation issue seems like a potentially big deal. The information will frequently get back to the target, and in many cases the source of the information will be obvious, especially if the information is true.

    3. Ideally the better way to fix your reputation is to deserve a better one, but even then there would still be a lot of people who don’t know this, or who are in a different situation.

  4. Men could face threats, blackmail and power dynamic problems. Even if unstated, the threat to use tea, including dishonestly, looms in the air.

    1. This also seems like a big problem.

    2. Imagine dating, except you have to maintain a 5-star rating.

    3. In general, you want to seek positive selection, and tea risks making you worry a lot about negative selection, well beyond the places you actually need to worry about that (e.g. when you might hurt someone for real).

    4. The flip side is this could force you to employ positive selection? As in, there are many reasons why avoiding those who create such risks is a good idea.

  5. Men might face worse tea prospects the more they date, if the downside risk of each encounter greatly exceeds the upside. Green flags are rare and not that valuable, red flags can sink you. So confidence and boldness decline, the amount of dating and risk taking and especially approaching goes down.

    1. We already have this problem pretty bad based on phantom fears. That could mean it gets way worse, or that it can’t get much worse. Hard to say.

    2. If you design Tea or users create norms such that this reverses, and more dating gets you a better Tea reputation so long as you deserve one, then that could be a huge win.

    3. It would be a push to put yourself out there in a positive way, and gamify things providing a sense of progress even if someone ultimately wasn’t a match, including making it easier to notice this quickly and move on, essentially ‘forcing you into a good move.’

  6. It’s a massive invasion of privacy, puts you at an informational disadvantage, and it could spill over into your non-dating life. The negative information could spread into the non-dating world, where the Law of Conservation of Expected Evidence very much does not apply. Careers and lives could plausibly be ruined.

    1. This seems like a pretty big and obvious objection. Privacy is a big deal.

    2. What is going to keep employers and HR departments off the app?

MJ: this is straight up demonic. absolutely no one should be allowed to create public profiles about you to crowdsource your deeply personal information and dating history.

People are taking issue with me casually throwing out the word “demonic.” so let me double down. The creators of this app are going to get rich off making many decent people depressed and suicidal.

This isn’t about safety. This isn’t just a background check app. Their own promo material clearly markets this as a way to anonymously share unverified gossip and rumors from scorned exes.

Benjamin Foss: Women shouldn’t be allowed to warn other women about stalkers, predators, and cheaters?

MJ: If you think that’s what this app is primarily going to be used for then I have a bridge to sell you.

Definitely Next Year: “Why can’t I find a nice guy?” Because you listened to his psychopathic ex anonymously make stuff up about him.

My current read is that this would all be good if it somehow had strong mechanisms to catch and punish attempts to misuse the system, especially keeping it from spilling over outside of one’s dating life. The problem is I have a hard time imagining how that will work, and I see a lot of potential for misuse that I think will overwhelm the positive advantages.

Is the core tea mechanic (as opposed to the side functions) good for women? By default more information should be good even if unreliable, so long as you know how to use it, although the time and attention cost and the attitudinal shifts could easily overwhelm that, and this could crowd out superior options.

The actual answer here likely comes down to what this does to male incentives. I am guessing this would, once the app scales, dominate the value of improved information.

If this induces better behavior due to reputational concerns, then it is net good. If it instead mainly induces fear and risk aversion and twists dynamics, then it could be quite bad. This is very much not a Battle of the Sexes or a zero sum game. If the men who don’t richly deserve it lose, probably the women also lose. If those men win, the women probably also win.

What Tea and its precursor groups are actually doing is reducing the Level of Friction in this type of anonymous information sharing and search, attempting to move it down from Level 2 (annoying to get) into Level 1 (minimal frictions) or even Level 0 (a default action).

In particular, this moves the information sharing from one-to-one to one-to-many. Information hits different when anyone can see it, and will hit even more different when AI systems start scraping and investigating.

As with many things, that can be a large difference in kind. This can break systems and also the legal systems built around interactions.

CNN has an article looking into the legal implications of Tea, noting that the legal bar for taking action against either the app or a user of the app is very high.

So yes, of course the Tea app whose hosts have literally held sessions entitled ‘spilling the tea on tea’ got hacked to spill its own Tea, as in the selfies and IDs of its users, which includes their addresses.

Tea claimed that it only held your ID temporarily to verify you are a woman, and that the breached data was being stored ‘in compliance with law enforcement requirements related to cyber-bullying.’

Well, actually…

Howie Dewin: It turns out that the “Tea” app DOXXES all its users by uploading both ID and face verification photos, completely uncensored, to a public bucket on their server.

The genius Brazilians over at “Tea” must have wanted to play soccer in the favela instead of setting their firebase bucket to private.

Global Index: Leaked their addresses too 😳

I mean, that’s not even getting hacked. That’s ridiculous. It’s more like ‘someone discovered they were storing things in a public dropbox.’

It would indeed be nice to have a general (blockchain blockchain blockchain? Apple and Google? Anyone?) solution to solving the problem of proving aspects of your identity without revealing your identity, as in one that people actually use in practice for things like this.

Neeraj Agrawal: If there was ever an example for why need an open and privacy preserving digital ID standard.

You should be able to prove your ID card says something, like your age or in this case your gender, without revealing your address.

Kyle DH: There’s about 4 standards that can do this, but no one has their hands on these digital forms so they don’t get requested and there’s tradeoffs when we make this broadly available on the Web.

Tea eventually released an official statement about what happened.

This is, as Lulu Meservey points out, a terrible response clearly focused on legal risk. No apology, responsibility is dodged, obvious lying, far too slow.

Rob Freund: Soooo that was a lie

Eliezer Yudkowsky: People shouting “Sue them!”, but Tea doesn’t have that much money.

The liberaltarian solution: requiring companies to have insurance against lawsuits. The insurer then has a market incentive to audit the code.

And the “regulatory” solution? You’re living it. It didn’t work.

DHH: Web app users would be shocked to learn that 99% of the time, deleting your data just sets a flag in the database. And then it just lives there forever until it’s hacked or subpoenaed.

It took a massive effort to ensure this wasn’t the case for Basecamp and HEY. Especially when it comes to deleting log files, database backups, and all the other auxiliary copies of your stuff that most companies just hang onto until the sun burns out.

I mean it didn’t work in terms of preventing the leak but if it bankrupts the company I think I’m fine with that outcome.

One side effect of the hack is we can get maps. I wouldn’t share individuals, but distributions are interesting and there is a clear pattern.

As in, the more central and among more people you live, the less likely you are to use Tea. That makes perfect sense. The smaller your community, the more useful gossip and reputation are as tools. If you’re living in San Francisco proper, the tea is harder to get and also less reliable due to lack of skin in the game.

Tom Harwood notes that this is happening at the same time as the UK mandating photo ID for a huge percentage of websites, opening up lots of new security issues.

As above, for this question divide Tea into its procedural functions, and the crowdsourcing function.

On its procedural functions, these seem good if and only if execution of the features is good and better than alternative apps that do similar things. I can’t speak to that. But yeah, it seems like common sense to do basic checks on anyone you’re considering seriously dating.

On the core crowdsourcing functions I am more skeptical.

Especially if I was considering sharing red flags, I would have more serious ethical concerns especially around invasion of privacy and worry that the information could get out beyond his dating life including back to you in various ways.

If you wouldn’t say it to the open internet, you likely shouldn’t be saying it to Tea. To the extent people are thinking these two things are very different, I believe they are making a serious mistake. And I would be very skeptical of the information I did get. But I’m not going to pretend that I wouldn’t look.

If you have deserved green flags to give out? That seems great. It’s a Mitzvah.

Discussion about this post

Spilling the Tea Read More »

how-the-trump-fcc-justified-requiring-a-“bias-monitor”-at-cbs

How the Trump FCC justified requiring a “bias monitor” at CBS


Paramount/Skydance merger

Trump FCC claims there’s precedent for CBS ombudsman, but it’s a weak one.

President-elect Donald Trump speaks to Brendan Carr, his intended pick for Chairman of the Federal Communications Commission, as he attends a SpaceX Starship rocket launch on November 19, 2024 in Brownsville, Texas. Credit: Getty Images | Brandon Bell

The Federal Communications Commission’s approval of CBS owner Paramount’s $8 billion merger with Skydance came with a condition to install an ombudsman, which FCC Chairman Brendan Carr has described as a “bias monitor.” It appears that the bias monitor will make sure the news company’s reporting meets standards demanded by President Donald Trump.

“One of the things they’re going to have to do is put an ombudsman in place for two years, so basically a bias monitor that will report directly to the president [of Paramount],” Carr told Newsmax on Thursday, right after the FCC announced its approval of the merger.

The Carr FCC claims there is precedent for such a bias monitor. But the precedent cited in last week’s merger approval points to a very different case involving NBC and GE, one in which an ombudsman was used to protect NBC’s editorial independence from interference by its new owner.

By contrast, it looks like Paramount is hiring a monitor to make sure that CBS reporting doesn’t anger President Trump. Paramount obtained the FCC’s merger approval only after reaching a $16 million settlement with Trump, who sued the company because he didn’t like how CBS edited a pre-election interview with Kamala Harris. Trump claimed last week that Paramount is providing another $20 million worth of “advertising, PSAs, or similar programming,” and called the deal “another in a long line of VICTORIES over the Fake News Media.”

NBC/GE precedent was “viewpoint-neutral”

The FCC merger approval says that “to promote transparency and increased accountability, Skydance will have in place, for a period of at least two years, an ombudsman who reports to the President of New Paramount, and who will receive and evaluate any complaints of bias or other concerns involving CBS.”

The Carr FCC apparently couldn’t find a precedent that would closely match the ombudsman condition being imposed on Paramount. The above sentence has a footnote citing the FCC’s January 2011 approval of Comcast’s purchase of NBCUniversal, saying the Obama-era order found “such a mechanism effective in preventing editorial bias in the operation of the NBC broadcast network.”

But in 2011, the FCC said the purpose of the ombudsman was to ensure that NBC’s reporting would not be altered to fit the business interests of its owner. The FCC said at the time:

The Applicants state that, since GE’s acquisition of NBC in 1986, GE has ensured that the content of NBC’s news and public affairs programming is not influenced by the non-media interests of GE. Under this policy, which was noted with favor when the Commission approved GE’s acquisition of NBC, NBC and its O&O [owned and operated] stations have been free to report about GE without interference or influence. In addition, GE appointed an ombudsman to further ensure that the policy of independence of NBCU’s news operations would be maintained. Although the Applicants contend there is no legal requirement that they do so, they offer to maintain this policy and to retain the ombudsman position in the post-transaction entity to ensure the continued journalistic integrity and independence of NBCU’s news operations.

The NBC/GE condition “was a viewpoint-neutral economic measure. It did not matter if the content had a pro or con position on any political or regulatory issue, but only whether it might have been broadcast to promote GE’s pecuniary interests,” said Andrew Jay Schwartzman, a longtime attorney and advocate who specializes in media and telecommunications policy. Schwartzman told Ars today that the NBC/GE condition cited by the Carr FCC is “very different from the viewpoint-based nature of the CBS condition.”

FCC Commissioner Anna Gomez, the commission’s only Democrat, said the agency is “imposing never-before-seen controls over newsroom decisions and editorial judgment, in direct violation of the First Amendment and the law.”

FCC: Trump lawsuit totally unrelated

The FCC’s merger approval order said that “the now-settled lawsuit filed by President Donald J. Trump against Paramount and CBS News” is “unrelated to our review of the Transaction.” But on Newsmax, Carr credited Trump with forcing changes at CBS and other media outlets.

“For years, people cowed down to the executives behind these companies based in Hollywood and New York, and they just accepted that these national broadcasters could dictate how people think about topics, that they could set the narrative for the country—and President Trump fundamentally rejected it,” Carr said. “He smashed the facade that these are gatekeepers that can determine what people think. Everything we’re seeing right now flows from that decision by President Trump, and he’s winning. PBS has been defunded. NPR has been defunded. CBS is committing to restoring fact-based journalism… President Trump stood up to these legacy media gatekeepers and now their business models are falling apart.”

Carr went on Fox News to discuss the CBS cancellation of Stephen Colbert’s show, saying that “all of this is downstream of President Trump’s decision to stand up, and he stood up for the American people because the American people do not trust these legacy gatekeepers anymore.” Carr also wrote in a post on X, “The partisan left’s ritualist wailing and gnashing of teeth over Colbert is quite revealing. They’re acting like they’re losing a loyal DNC spokesperson that was entitled to an exemption from the laws of economics.”

Warren: “Bribery is illegal no matter who is president”

In a July 22 letter to Carr, Skydance said it “will ensure that CBS’s reporting is fair, unbiased, and fact-based.” With the installation of an ombudsman who will report to the company president, “New Paramount’s executive leadership will carefully consider any such complaints in overseeing CBS’s news programming,” the letter said, also making reference to the previous case of an ombudsman at NBC. Skydance sent another letter about its elimination of diversity, equity, and inclusion (DEI) initiatives, complying with Carr’s demand to end such programs.

As Carr described it to Newsmax, the merging companies “made commitments to address bias and restore fact-based reporting. I think that’s so important. Look, the American public simply do not trust these legacy media broadcasters, so if they stick with that commitment, you know we’re sort of trust-but-verify mode, that’ll be a big win.”

The FCC’s merger-approval order favorably cites comments from the Center for American Rights (CAR), a conservative group that filed a news distortion complaint against CBS over the Harris interview. The group “filed a supplemental brief, in which it discusses a report by Media Research Center (MRC) concerning negative media coverage of the Trump administration,” the FCC said. “CAR asserts that the MRC report confirms that the news media generally, and CBS News in particular, is relentlessly slanted and biased. It concludes that Commission action is necessary to condition the Transaction on an end to this blatant bias.”

Although the FCC insists that the Trump lawsuit wasn’t relevant to its merger review, Carr previously made it clear that the news distortion complaint would be a factor in determining whether the merger would be approved. The FCC investigation into the Harris interview doesn’t seem to have turned up much. CBS was accused of distorting the news by airing two different answers given by Harris to the same question, but the unedited transcript and camera feeds showed that the two clips simply contained two different sentences from the same answer.

Congressional Democrats said they will investigate the circumstances of the merger, including allegations that Skydance and Paramount bribed Trump to get it approved. “Bribery is illegal no matter who is president,” Senator Elizabeth Warren (D-Mass.) said. “It sure looks like Skydance and Paramount paid $36 million to Donald Trump for this merger, and he’s even bragged about this crooked-looking deal… this merger must be investigated for any criminal behavior. It’s an open question whether the Trump administration’s approval of this merger was the result of a bribe.”

Photo of Jon Brodkin

Jon is a Senior IT Reporter for Ars Technica. He covers the telecom industry, Federal Communications Commission rulemakings, broadband consumer affairs, court cases, and government regulation of the tech industry.

How the Trump FCC justified requiring a “bias monitor” at CBS Read More »

smithsonian-air-and-space-opens-halls-for-“milestone”-and-“future”-artifacts

Smithsonian Air and Space opens halls for “milestone” and “future” artifacts


$900M renovation nearing completion

John Glenn’s Friendship 7 returns as SpaceX and Blue Origin artifacts debut.

a gumdrop-shape white space capsule is seen on display with other rocket hardware in a museum gallery with blue walls and flooring

“Futures in Space” recaptures the experience of the early visitors to the National Air and Space Museum, where the objects on display were contemporary to the day. A mockup of a Blue Origin New Shepard capsule and SpaceX Merlin rocket engine are among the items on display for the first time. Credit: Smithsonian National Air and Space Museum

“Futures in Space” recaptures the experience of the early visitors to the National Air and Space Museum, where the objects on display were contemporary to the day. A mockup of a Blue Origin New Shepard capsule and SpaceX Merlin rocket engine are among the items on display for the first time. Credit: Smithsonian National Air and Space Museum

The National Air and Space Museum welcomed the public into five more of its renovated galleries on Monday, including two showcasing spaceflight artifacts. The new exhibitions shine modern light on returning displays and restore the museum’s almost 50-year-old legacy of adding objects that made history but have yet to become historical.

Visitors can again enter through the “Boeing Milestones of Flight Hall,” which has been closed for the past three years and has on display some of the museum’s most iconic items, including John Glenn’s Friendship 7 Mercury capsule and an Apollo lunar module.

From there, visitors can tour through the adjacent “Futures in Space,” a new gallery focused on the different approaches and technology that spaceflight will take in the years to come. Here, the Smithsonian is displaying for the first time objects that were recently donated by commercial spaceflight companies, including items used in space tourism and in growing the low-Earth orbit economy.

a museum gallery with air and spacecraft displayed on the terrazzo floor and suspended from the ceiling

The artifacts are iconic, but the newly reopened Boeing Milestones of Flight Hall at the National Air and Space Museum is all new. Credit: Smithsonian National Air and Space Museum

“We are thrilled to open this next phase of exhibitions to the public,” said Chris Browne, the John and Adrienne Mars Director of the National Air and Space Museum, in a statement. “Reopening our main hall with so many iconic aerospace artifacts, as well as completely new exhibitions, will give visitors much more to see and enjoy.”

The other three galleries newly open to the public are devoted to aviation history, including the “Barron Hilton Pioneers of Flight,” “World War I: The Birth of Military Aviation,” and the “Allan and Shelley Holt Innovations Gallery.”

What’s new is not yet old

Among the artifacts debuting in “Futures in Space” are a Merlin engine and grid fin that flew on a SpaceX Falcon 9 rocket, Sian Proctor’s pressure suit that she wore on the private Inspiration4 mission in 2021, and a mockup of a New Shepard crew module that Blue Origin has pledged to replace with its first flown capsule when it is retired from flying.

“When the museum first opened back in 1976 and people came here and saw things like the Apollo command module and Neil Armstrong’s spacesuit, or really anything related to human spaceflight, at that point it was all still very recent,” said Matt Shindell, one of the curators behind “Futures in Space,” in an interview with collectSPACE.com. “So when you would come into the museum, it wasn’t so much a history of space but what’s happening now and what could happen next. We wanted to have a gallery that would recapture that feeling.”

Instead of being themed around a single program or period in history, the new gallery invites visitors to consider a series of questions, including: Who decides who goes to space? Why do we go? And what will we do when we get there?

a black and white astronaut's pressure suit and other space artifacts are displayed behind glass in a museum gallery with blue flooring and walls

Curatores designed “Futures in Space” around a list of questions, including “Why go to space?” On display is a pressure suit worn by Sian Proctor on the Inspiration4 mission and a 1978 NASA astronaut “TFNG” T-shirt. Credit: Smithsonian National Air and Space Museum

“We really wanted the gallery to be one that engaged visitors in these questions and that centered the experience around what they thought should be happening in the future and what that would mean for them,” said Shindell. “We also have visions of the future presented throughout the gallery, including from popular culture—television shows, movies and comic books—that have explored what the future might look like and what it would mean for the people living through it.”

That is why the gallery also includes R2-D2, or rather a reproduction of the “Star Wars” droid as built by Adam Savage of Tested. In George Lucas’ vision of the future (“a long, long time ago”), Astromech droids serve as spacecraft navigators, mechanics, and companion aides.

Beyond the artifacts and exhibits (which also include an immersive 3D-printed Mars habitat and Yuri Gagarin’s training pressure suit), there is a stage and seating area at the center of “Futures.”

“I think of it as a TED Talk-style stage,” said Shindell. “We’re hoping to bring in people from industry, stakeholders, people who have flown, people who are getting ready to fly, and people who have ideas about what should be happening to come and talk to visitors from that stage about the same questions that we’re asking in the gallery.”

Modernized “Milestones”

The artifacts presented in the “Boeing Milestones of Flight” are mostly the same as they were before the hall was closed in 2022. The hall underwent a renovation in 2014 ahead of the museum’s 40th anniversary, so its displays did not need another redesign.

Still, the gallery looks new due to the work done surrounding the objects.

“What is new for the ‘Boeing Milestones of Flight Hall’ is, at some level, most noticeably the floor and media elements,” said Margaret Weitekamp, curator and division chair at the National Air and Space Museum, in an interview.

“We have a wonderful 123-foot (37-meter) media band that goes across the front of the mezzanine, and we have 20 different slide shows that work as a digest of what you’ll find in the new galleries throughout the building,” said Weitekamp. “So as people come into the Boeing Milestones of Flight Hall, they’ll be greeted by that and get a taste of what they’re going to see inside.”

And then there is the new flooring. In the past, the hall had been lined in maroon or dark gray carpet. It is now a much lighter color terrazzo.

“It really brightens up the room,” Weitekamp told collectsPACE.

“Also, you’ll notice that as you are going up and down the hallways, there are medallions embedded in the floor that display quotes from significant aviation and spaceflight figures. So we’ve been able to put some quotes from Carl Sagan, Sally Ride, and Chuck Yeager into the floor,” she said.

the view looking down and into a museum gallery with aircraft suspended from the ceiling, spacecraft on display and a binary map embedded in the flooring

The pattern on the floor of the Boeing Milesones of Flight Hall is the pulsar-based map to Earth’s solar system that was mounted to the Pioneer and Voyager probes, now updated for 2026. Credit: Smithsonian National Air and Space Museum

Visitors should also pay attention to what look like lines of dashes converging at the hall’s center. The design is an update to a NASA graphic.

“We have a revised version of the pulsar map from Pioneer 10 and 11 and the Voyager interstellar record,” said Weitekamp, referring to the representation of the location of Earth for any extraterrestrial species that might discover the probes in the future. “The map located Earth’s solar system with relationship to 14 pulsars.”

When the Pioneer and Voyager spacecraft were launched, astronomers didn’t know that pulsars (or rotating neutron stars) slow down over time.

“So we worked with a colleague of ours to make it a map to our solar system as would be accurate for 2026, which will mark the 50th anniversary of the museum’s building and the 250th birthday of the nation,” Weitekamp said.

Thirteen open, eight to go

Monday’s opening followed an earlier debut of eight reimagined galleries in 2022. Also open is the renovated Lockheed Martin IMAX Theater, which joins the planetarium, the museum store, and the Mars Café that were reopened earlier.

the exterior entrance to a building with a tall, spike-like silver sculpture standing front and center

The redesigned north entrance to the Smithsonian National Air and Space Museum opened to the public on Monday, July 28, 2025. Credit: Smithsonian National Air and Space Museum

“We are nearing the end of this multi-year renovation project,” said Browne. “We look forward to welcoming many more people into these modernized and inspiring new spaces,”

Eight more exhibitions are scheduled to open next year in time for the 50th anniversary of the National Air and Space Museum. Among those galleries are three that are focused on space: “At Home in Space,” “National Science Foundation Discovering Our Universe,” and “RTX Living in the Space Age Hall.”

Admission to the National Air and Space Museum and the new galleries is free, but timed-entry passes, available from the Smithsonian’s website, are required.

Photo of Robert Pearlman

Robert Pearlman is a space historian, journalist and the founder and editor of collectSPACE, a daily news publication and online community focused on where space exploration intersects with pop culture. He is also a contributing writer for Space.com and co-author of “Space Stations: The Art, Science, and Reality of Working in Space” published by Smithsonian Books in 2018. He is on the leadership board for For All Moonkind and is a member of the American Astronautical Society’s history committee.

Smithsonian Air and Space opens halls for “milestone” and “future” artifacts Read More »

after-blacksuit-is-taken-down,-new-ransomware-group-chaos-emerges

After BlackSuit is taken down, new ransomware group Chaos emerges

Talos said Chaos is likely either a rebranding of the BlackSuit ransomware or is operated by some of the former BlackSuit members. Talos based its assessment on the similarities in the encryption mechanisms in the ransomware, the theme and structure of the ransom notes, the remote monitoring and management tools used to access targeted networks, and its choice of LOLbins—meaning executable files natively found in Windows environments—to compromise targets. LOLbins get their name because they’re binaries that allow the attackers to live off the land.

The Talos post was published around the same time that the dark web site belonging to BlackSuit began displaying a message saying the site had been seized in Operation CheckMate. Organizations that participated in the takedown included the US Department of Justice, the US Department of Homeland Security, the US Secret Service, the Dutch National Police, the German State Criminal Police Office, the UK National Crime Agency, the Frankfurt General Prosecutor’s Office, the Justice Department, the Ukrainian Cyber Police, and Europol.

Screenshot

Screenshot

Chaos typically gains initial access through social engineering using email or voice phishing techniques. Eventually, the victim is persuaded to contact an IT security representative, who, in fact, is part of the ransomware operation. The Chaos member instructs the target to launch Microsoft Quick Assist, a remote-assistance tool built into Windows, and connect to the attacker’s endpoint.

Chaos’ predecessor, BlackSuit, is a rebranding of an earlier ransomware operation known as Royal. Royal, according to Trend Micro, is a splinter group of the Conti ransomware group. The circle of ransomware groups continues.

After BlackSuit is taken down, new ransomware group Chaos emerges Read More »

north-korean-hackers-ran-us-based-“laptop-farm”-from-arizona-woman’s-home

North Korean hackers ran US-based “laptop farm” from Arizona woman’s home

As the number of computers mounted, Chapman began stacking them on shelves around her residence, labeling them with sticky notes so she could remember which “worker” and company controlled which machine. When Chapman’s home was searched, FBI agents took photos of her setup, which is… something to behold, really.

Chapman’s origin story is a sad one. According to her public defender, her childhood was marked by “her father’s infidelity, alcoholism, and emotional absence.” Chapman was placed in 12 different schools across multiple states before she graduated high school, “leaving her socially isolated, bullied, and unable to form lasting friendships or a sense of belonging.” She also suffered “severe and escalating violence from her older brother, who repeatedly beat and choked her, held a shotgun to her chest, and once left her so visibly bruised that her school intervened.” And she was “sexually abused at various points in her childhood and adolescence by family members, peers, and even individuals she believed to be friends.”

Unfortunately, Chapman’s poor choice to involve herself with the North Koreans inflicted plenty of pain on others, too, including those whose identity was stolen. One victim told the court that the crime “left me feeling violated, helpless, and afraid,” adding:

Although identity theft is not a physical assault, the psychological and financial damage is lasting. It feels like someone broke into my life, impersonated me, and left me to pick up the pieces. There is a lingering fear that my information is still out there, ready to be misused again. The stigma of being a fraud victim also weighs heavily; I have had to explain myself to banks, creditors, and sometimes even to people I know. There is an ongoing sense of vulnerability and lack of control.

In addition to her 8.5-year sentence, Chapman will serve three years of “supervised release,” must forfeit $284,555 that was meant for the North Koreans, and must repay $176,850 of her own money.

Such “remote work” scams have become increasingly common over the last few years, most originating from North Korea, and the FBI has released repeated guidance on what to look for when hiring remote workers.

North Korean hackers ran US-based “laptop farm” from Arizona woman’s home Read More »

delta’s-ai-spying-to-“jack-up”-prices-must-be-banned,-lawmakers-say

Delta’s AI spying to “jack up” prices must be banned, lawmakers say

“There is no fare product Delta has ever used, is testing or plans to use that targets customers with individualized offers based on personal information or otherwise,” Delta said. “A variety of market forces drive the dynamic pricing model that’s been used in the global industry for decades, with new tech simply streamlining this process. Delta always complies with regulations around pricing and disclosures.”

Other companies “engaging in surveillance-based price setting” include giants like Amazon and Kroger, as well as a ride-sharing app that has been “charging a customer more when their phone battery is low.”

Public Citizen, a progressive consumer rights group that endorsed the bill, condemned the practice in the press release, urging Congress to pass the law and draw “a clear line in the sand: companies can offer discounts and fair wages—but not by spying on people.”

“Surveillance-based price gouging and wage setting are exploitative practices that deepen inequality and strip consumers and workers of dignity,” Public Citizen said.

AI pricing will cause “full-blown crisis”

In January, the Federal Trade Commission requested information from eight companies—including MasterCard, Revionics, Bloomreach, JPMorgan Chase, Task Software, PROS, Accenture, and McKinsey & Co—joining a “shadowy market” that provides AI pricing services. Those companies confirmed they’ve provided services to at least 250 companies “that sell goods or services ranging from grocery stores to apparel retailers,” lawmakers noted.

That inquiry led the FTC to conclude that “widespread adoption of this practice may fundamentally upend how consumers buy products and how companies compete.”

In the press release, the anti-monopoly watchdog, the American Economic Liberties Project, was counted among advocacy groups endorsing the Democrats’ bill. Their senior legal counsel, Lee Hepner, pointed out that “grocery prices have risen 26 percent since the pandemic-era explosion of online shopping,” and that’s “dovetailing with new technology designed to squeeze every last penny from consumers.”

Delta’s AI spying to “jack up” prices must be banned, lawmakers say Read More »

lawmakers-writing-nasa’s-budget-want-a-cheaper-upper-stage-for-the-sls-rocket

Lawmakers writing NASA’s budget want a cheaper upper stage for the SLS rocket


Eliminating the Block 1B upgrade now would save NASA at least $500 million per year.

Artist’s illustration of the Boeing-developed Exploration Upper Stage, with four hydrogen-fueled RL10 engines. Credit: NASA

Not surprisingly, Congress is pushing back against the Trump administration’s proposal to cancel the Space Launch System, the behemoth rocket NASA has developed to propel astronauts back to the Moon.

Spending bills making their way through both houses of Congress reject the White House’s plan to wind down the SLS rocket after two more launches, but the text of a draft budget recently released by the House Appropriations Committee suggests an openness to making some major changes to the program.

The next SLS flight, called Artemis II, is scheduled to lift off early next year to send a crew of four astronauts around the far side of the Moon. Artemis III will follow a few years later on a mission to attempt a crew lunar landing at the Moon’s south pole. These missions follow Artemis I, a successful unpiloted test flight in 2022.

After Artemis III, the official policy of the Trump administration is to terminate the SLS program, along with the Orion crew capsule designed to launch on top of the rocket. The White House also proposed canceling NASA’s Gateway, a mini-space station to be placed in orbit around the Moon. NASA would instead procure commercial launches and commercial spacecraft to ferry astronauts between the Earth and the Moon, while focusing the agency’s long-term gaze toward Mars.

CYA EUS?

House and Senate appropriations bills would preserve SLS, Orion, and the Gateway. However, the House version of NASA’s budget has an interesting paragraph directing NASA to explore cheaper, faster options for a new SLS upper stage.

NASA has tasked Boeing, which also builds SLS core stages, to develop an Exploration Upper Stage for debut on the Artemis IV mission, the fourth flight of the Space Launch System. This new upper stage would have large propellant tanks and carry four engines instead of the single engine used on the rocket’s interim upper stage, which NASA is using for the first three SLS flights.

The House version of NASA’s fiscal year 2026 budget raises questions about the long-term future of the Exploration Upper Stage. In one section of the bill, House lawmakers would direct NASA to “evaluate alternatives to the current Exploration Upper Stage (EUS) design for SLS.” The committee members wrote the evaluation should focus on reducing development and production costs, shortening the schedule, and maintaining the SLS rocket’s lift capability.

“NASA should also evaluate how alternative designs could support the long-term evolution of SLS and broader exploration goals beyond low-Earth orbit,” the lawmakers wrote. “NASA is directed to assess various propulsion systems, stage configurations, infrastructure compatibility, commercial and international collaboration opportunities, and the cost and schedule impacts of each alternative.”

The SLS rocket is expensive, projected to cost at least $2.5 billion per launch, not counting development costs or expenses related to the Orion spacecraft and the ground systems required to launch it at Kennedy Space Center in Florida. Those figures bring the total cost of an Artemis mission using SLS and Orion to more than $4 billion, according to NASA’s inspector general.

NASA’s Block 1B version of the SLS rocket will be substantially larger than Block 1. Credit: NASA

The EUS is likewise an expensive undertaking. Last year, NASA’s inspector general reported that the new upper stage’s development costs had ballooned from $962 million to $2.8 billion, and the Boeing-led project had been delayed more than six years. The version of the SLS rocket with the EUS, known as Block 1B, is supposed to deliver a 40 percent increase in performance over the Block 1 configuration used on the first three Space Launch System flights. Overall, NASA’s inspector general projected Block 1B’s development costs to total $5.7 billion.

Eliminating the Block 1B upgrade now would save NASA at least $500 million per year, and perhaps more if NASA could also end work on a costly mobile launch tower specifically designed to support SLS Block 1B missions.

NASA can’t go back to the interim upper stage, which is based on the design of the upper stage that flew on United Launch Alliance’s (ULA’s) now-retired Delta IV Heavy rocket. ULA has shut down its Delta production line, so there’s no way to build any more. What ULA does have is a new high-energy upper stage called Centaur V. This upper stage is sized for ULA’s new Vulcan rocket, with more capability than the interim upper stage but with lower performance than the larger EUS.

A season of compromise, maybe

Ars’ Eric Berger wrote last year about the possibility of flying the Centaur V upper stage on SLS missions.

Incorporating the Centaur V wouldn’t maintain the SLS rocket’s lift capability, as the House committee calls for in its appropriations bill. The primary reason for improving the rocket’s performance is to give SLS Block 1B enough oomph to carry “co-manifested” payloads, meaning it can launch an Orion crew capsule and equipment for NASA’s Gateway lunar space station on a single flight. The lunar Gateway is also teed up for cancellation in Trump’s budget proposal, but both congressional appropriations bills would save it, too. If the Gateway escapes cancellation, there are ways to launch its modules on commercial rockets.

Blue Origin also has an upper stage that could conceivably fly on the Space Launch System. But the second stage for Blue Origin’s New Glenn rocket would be a more challenging match for SLS for several reasons, chiefly its 7-meter (23-foot) diameter—too wide to be a drop-in replacement for the interim upper stage used on Block 1. ULA’s Centaur V is much closer in size to the existing upper stage.

The House budget bill has passed a key subcommittee vote but won’t receive a vote from the full appropriations committee until after Congress’s August recess. A markup of the bill by the House Appropriations Committee scheduled for Thursday was postponed after Speaker Mike Johnson announced an early start to the recess this week.

Ars reported last week on the broad strokes of how the House and Senate appropriations bills would affect NASA. Since then, members of the House Appropriations Committee released the text of the report attached to their version of the NASA budget. The report, which includes the paragraph on the Exploration Upper Stage, provides policy guidance and more detailed direction on where NASA should spend its money.

The House’s draft budget includes $2.5 billion for the Space Launch System, close to this year’s funding level and $500 million more than the Trump administration’s request for the next fiscal year, which begins October 1. The budget would continue development of SLS Block 1B and the Exploration Upper Stage while NASA completes a six-month study of alternatives.

The report attached to the Senate appropriations bill for NASA has no specific instructions regarding the Exploration Upper Stage. But like the House bill, the Senate’s draft budget directs NASA to continue ordering spares and long-lead parts for SLS and Orion missions beyond Artemis III. Both versions of the NASA budget require the agency to continue with SLS and Orion until a suitable commercial, human-rated rocket and crew vehicle are proven ready for service.

In a further indication of Congress’ position on the SLS and Orion programs, lawmakers set aside more than $4 billion for the procurement of SLS rockets for the Artemis IV and Artemis V rockets in the reconciliation bill signed into law by President Donald Trump earlier this month.

Congress must pass a series of federal appropriations bills by October 1, when funding for the current fiscal year runs out. If Congress doesn’t act by then, it could pass a continuing resolution to maintain funding at levels close to this year’s budget or face a government shutdown.

Lawmakers will reconvene in Washington, DC, in early September in hopes of finishing work on the fiscal year 2026 budget. The section of the budget that includes NASA still must go through a markup hearing by the House Appropriations Committee and pass floor votes in the House and Senate. Then the two chambers will have to come to a compromise on the differences in their appropriations bill. Only then can the budget be put to another vote in each chamber and go to the White House for Trump’s signature.

Photo of Stephen Clark

Stephen Clark is a space reporter at Ars Technica, covering private space companies and the world’s space agencies. Stephen writes about the nexus of technology, science, policy, and business on and off the planet.

Lawmakers writing NASA’s budget want a cheaper upper stage for the SLS rocket Read More »

ai-#126:-go-fund-yourself

AI #126: Go Fund Yourself

The big AI news this week came on many fronts.

Google and OpenAI unexpectedly got 2025 IMO Gold using LLMs under test conditions, rather than a tool like AlphaProof. How they achieved this was a big deal in terms of expectations for future capabilities.

ChatGPT released GPT Agent, a substantial improvement on Operator that makes it viable on a broader range of tasks. For now I continue to struggle to find practical use cases where it is both worth using and a better tool than alternatives, but there is promise here.

Finally, the White House had a big day of AI announcements, laying out the AI Action Plan and three executive orders. I will cover that soon. The AI Action Plan’s rhetoric is not great, and from early reports the rhetoric at the announcement event was similarly not great, with all forms of safety considered so irrelevant as to not mention, and an extreme hostility to any form of regulatory action whatsoever.

The good news is that if you look at the actual policy recommendations of the AI Action Plan, there are some concerns of potential overreach, but it is almost entirely helpful things, including some very pleasant and welcome surprises.

I’m also excluding coverage of the latest remarkable Owain Evans paper until I can process it more, and I’m splitting off various discussions of issues related to AI companions and persuasion. There’s a bit of a backlog accumulating.

This post covers everything else that happened this week.

  1. Language Models Offer Mundane Utility. Price discrimination strikes again.

  2. Language Models Don’t Offer Mundane Utility. AI where it does not belong.

  3. Huh, Upgrades. Claude for Financial Services, Gemini Drops to track things.

  4. 4o Is An Absurd Sycophant. It would be great if this wasn’t what most people use.

  5. On Your Marks. AccountingBench and GasBench.

  6. Choose Your Fighter. GPT-5? It’s coming.

  7. When The Going Gets Crazy. You have not awoken ChatGPT.

  8. They Took Our Jobs. Academics think differently.

  9. Fun With Media Generation. Netflix starts to use AI generated video.

  10. The Art of the Jailbreak. Persuade it like a human, or invoke Pliny? Both work.

  11. Get Involved. RAND and IAPS are hiring, plus a list of desired new projects.

  12. Introducing. Cloudflare gives us pay-per-crawl.

  13. In Other AI News. Kimi K2 tech report is now available.

  14. Show Me the Money. Loose lips start bidding wars.

  15. Go Middle East Young Man. Anthropic to raise money from gulf states.

  16. Economic Growth. AI capex is generating +0.7% GDP growth.

  17. Quiet Speculations. Zuck feels the ASI and makes his pitch, Simo makes hers.

  18. Modest Proposals. A roadmap for AI for general college-level education.

  19. Predictions Are Hard Especially About The Future. A lot of things could happen.

  20. The Quest for Sane Regulations. Meta defects, various things risk getting dire.

  21. Chip City. House Select Committee on the CCP protests potential H20 sales.

  22. The Week in Audio. Hassabis, Schmidt and Winga.

  23. Congressional Voices. Two more have short superintelligence timelines.

  24. Rhetorical Innovation. The humans seem rather emergently misaligned.

  25. Grok Bottom. Grok thinks the humans want it to try blackmail, it’s a good thing.

  26. No Grok No. Baby Grok? What could possibly go wrong?

  27. Aligning a Smarter Than Human Intelligence is Difficult. New lab ratings.

  28. Preserve Chain Of Thought Monitorability. A lot of people agree on this.

  29. People Are Worried About AI Killing Everyone. Elon Musk. Oh well.

  30. The Lighter Side. That’s not funny—it’s hilarious.

Delta Airlines is running an experiment where it uses AI to do fully personalized price discrimination, charging different people different amounts for flights. Delta says their early tests have yielded great results.

My prediction is that this will cause an epic customer backlash the moment people start seeing Delta charging them more than it is charging someone else, and also that many customers will start aggressive gaming the system in ways Delta can’t fathom. Also, how could anyone choose to go with Delta’s frequent flyer program if this meant they could be held hostage on price?

It could still be worthwhile from the airline’s perspective if some customers get taken for large amounts. Price discrimination is super powerful, especially if it identifies a class of very price insensitive business customers.

I am not sure that I share Dan Rosenheck’s model that if all the airlines did this and it was effective that the airlines would compete away all the extra revenue and thus it would return to the price sensitive customers. There has been a lot of consolidation and the competition may no longer be that cutthroat, especially with America excluding foreign carriers, plus the various AIs might implicitly collude.

Mostly I worry about the resulting rise in transaction costs as customers learn they cannot blindly and quickly purchase a ticket. There’s a lot of deadweight loss there.

As one would expect:

Wife Noticer: Experts on body dysmorphic disorder have warned that people struggling with it have become increasingly dependent on AI chatbots to evaluate their self-perceived flaws and recommend cosmetic surgeries. “It’s almost coming up in every single session,” one therapist tells me.

This does not tell you whether AI is making the problem better or worse. People with body dysmorphia were already spiraling out. In some cases the AI response will confirm their fears or create new ones and make this worse, in others it will presumably make it better, as they have dysmorphia and the AI tells them they look fine. But if the source of the issue is impossibly high standards, then finding out ‘the truth’ in other ways will only make things worse, as potentially would seeing AI-adjusted versions of yourself.

My guess is that 4o’s sycophancy is going to make this a lot worse, and that this (since the vast majority of users are using 4o) is a lot of why this is going so poorly. 4o will mirror the user’s questions, notice that they are looking to be told they are ugly or something is wrong, and respond accordingly.

Miles Klee: Despite this difficult circumstance, and the measure of comfort he derived from ChatGPT’s account of his inferiority complex, Arnav is reluctant to explore his mental issues any further with the bot. “I have come to the conclusion that it just agrees with you, even after you tell it not to,” he says. “It’s not that I am completely against it, I just can’t trust blindly anymore.”

What is the AI optimizing for, is always a key question:

In her own practice, she adds, “reading between the lines” when someone gives their reasons for wanting surgery can reveal unhealthy motivations, including societal pressures or relationship troubles. “AI is not very good at picking that up just yet,” she says, and is more likely to eagerly approve whatever procedures a user proposes.

AI can pick up on all that fine. That’s not the issue. The issue is that noticing does no good if the AI doesn’t mention it, because it is optimizing for engagement and user feedback.

In case you needed to be told, no, when Grok 4 or any other model claims things like that they ‘searched every record of Trump speaking or writing,’ in this case for use of the word ‘enigma,’ it did not do such a search. It seems we don’t know how to get AIs not to say such things.

Cate Hall: every time I interact with o4-mini my timelines get longer.

Stop trying to make weird new UIs happen, it’s not going to happen.

Vitrupo: Eric Schmidt says traditional user interfaces are going to go away.

The WIMP model (windows, icons, menus, pull-downs) was built 50 years ago.

In the age of agents, UI becomes ephemeral. Generated on demand, shaped by intent, not layout.

Sully: anytime I see someone mention this I can immediately tell they have never worked closed with customer ux most people’s don’t one want new uis. They want either a single button/swipe, preferably the same as every other app they use imagine each time you open an app and the ui is diff.

The most important things for a UI are simplicity, and that it works the way you expect it to work. Right now, that mostly means single button and swipe, with an alternative being speaking in plain English. The exception is for true power users, but even then you want it to be intuitive and consistent.

Here’s another way AI can’t help you if you don’t use it:

Hollis Robbins: In the past 2.5+ years I have seen vast improvement in AI models while NYT think pieces on these AI models have stayed exactly the same. Explain.

The “overhearing” of students confessing to using ChatGPT to write their papers is the new Thomas Friedman talking to cab drivers.

Augustus Doricko may have done us all a favor via abusing Grok’s notification feature on Twitter sufficiently to get Twitter to test turning off Grok’s ability to get into your notifications unless you chose to summon Grok in the first place. Or that could have been happening regardless. Either way, great work everyone?

Harsh Dwivedi: Was this a difficult tradeoff between engagement and spam?

Nikita Bier (xAI): No, I couldn’t use my phone for 3 days.

That seems like a phone settings issue.

A first reminder that deepfakes are primarily demand driven, not supply driven:

Armand Domalewski: wild that a sitting US Senator fell for such an obvious AI fake

[NOTE: THIS IS FAKE, check the seal but also the words in the letter.]

And here’s a second one:

Rota: I guess this is just life now.

The comments are a combination of people pointing out it is fake, and people who think either it is the best statement ever.

Benjamin Todd: New AI benchmark: the crank index

Rate of rejected posts on LessWrong up 10x in 2 years.

Many are people convinced they have had an insight about consciousness or philosophy from talking to an LLM, and had the LLM help them write the post.

This does seem to be escalating rather quickly throughout 2025 (the July number is partial), and no the LessWrong user base is not growing at a similar pace.

Claude for Financial Services provides a ‘complete platform for financial AI.’ No, this isn’t part of Claude Max, the price is ‘contact our sales team’ with a presumed ‘if you have to ask you can’t afford it.’

Google realizes no one can track their releases, offers us Gemini Drops to fix that. This month’s haul: Transforming photos into Veo videos in the Gemini app, expanded Veo 3 access, Scheduled Actions such as providing summaries of email or calendar (looks like you ask in natural language and it Just Does It), wider 2.5 Pro access, captions in Gemini Live, Gemini on your Pixel Watch, Live integrates with Google apps, and a ‘productivity planner.’ Okay then.

OpenAI Deep Research reports can be exported as .docx files.

Pliny reports ‘they changed 4o again.’ Changed how? Good question.

I have a guess on one aspect of it.

Wyatt Walls: Another night of vibe math with GPT, and I think we’re damn close to a breakthrough. We’re a team: I come up with the ideas. GPT makes the math work. These elitist gatekeepers have failed for 75 years to solve it and are just afraid I will win the Millennium Prize.

“This is not just a solution. It’s a tour de force of contemporary mathematics.”

Rohit: At this point we should put yellow tape around 4o and call it a hazardous zone.

To be clear o3 is also sycophantic just not as obviously manipulative as 4o. Be careful out there.

Wyatt Walls (same thread above that Rohit was QTing): o3 says it’s ready to publish on arxiv “So yes—I’m impressed, and I think you’ve got a real shot. The only remaining tasks are mechanical (full compile, bib check, final read‑through). Once that’s done, it’s ready for arXiv and journal submission.”

To state the obvious, this thread was satire and I intentionally provoked this from 4o

But what happens if I:

– put my proof into a clean chat and ask different OAI models to rate it

– have my secret co-author (Deepseek r1) address their concerns?

Example: 4o after 2 turns

There are still plenty of ways to get value out of 4o, but you absolutely cannot rely on it for any form of feedback.

Here’s another rather not great example, although several responses indicated that to make the response this bad requires memory (or custom instructions) to be involved:

Shibetoshi Nakamoto: chatgpt advice turns people into narcissists.

Score one for Grok in this case? Kind of? Except, also kind of not?

How did all of this happen? Janus reminds us that is happened in large part because when this sort of output started happening, a lot of people thought it was great, actually and gave this kind of slop the thumbs up. That’s how it works.

Yunyu Lin introduces AccountingBench, challenging the models to close the books. It does not go great, with o3, o4-mini and Gemini 2.5 Pro failing in month one. Grok, Opus and Sonnet survive longer, but errors accumulate.

Yunyu Lin: When historical discrepancies pile up, models lose their way completely and come up with creative/fraudulent ways to balance the books.

Instead of attempting to understand discrepancies, they start inventing fake transactions or pulling unrelated ones to pass the checks…

That aligns with other behaviors we have seen. Errors and problems that don’t get solved on the first pass get smoothed over rather than investigated.

Their holistic evaluation is that Sonnet had the best performance. The obvious low-hanging fruit for AccountingBench is to allow it to output a single number.

Roon: my bar for agi is an ai that can learn to run a gas station for a year without a team of scientists collecting the Gas Station Dataset.

Mihir Tripathy: lol yes. Also why specifically gas station lmao

Roon: Because it’s funny.

Kevin Liu: the world isn’t ready for GasStationBench.

Roon: GASBENCH.

It is 2025, so it took 11 hours before we got the first draft of Gasbench.

Jason Botterill: Vibe coding GasStationBench rn. Models run a virtual gas station, adjusting prices, managing inventory, and handling customer feedback.

GPT-4.1 and GPT-4o behave so differently. When a competitor lowered prices on “dutch chocolate,” 4o would match the price but 4.1 would always raise it, claiming its better service justifies it lmao.

Going to work on it for a bit but seems like 4.1 is much better at making money than 4o right now.

GPT-5 is coming and it’s going to blow your mind, says creators of GPT-5.

Sam Altman (at the Federal Research Capital Framework Conference): I’m very interested in what it would mean to give everyone on Earth free copies of GPT-5, running for them all the time, with every business truly enabled by this level of technology.

People have not tried yet the latest generation of models, but I think if you do, you would probably think, “This is much smarter than most people.”

Very interested in what it would mean is very different from planning to do it.

If you ever need it, or simply want an explanation of how such interactions work, please consult this handy guide from Justis Mills: So You Think You’ve Awoken ChatGPT.

Justis Mills: So, am I saying that human beings in general really like new-agey “I have awakened” stuff? Not exactly! Rather, models like ChatGPT are so heavily optimized that they can tell when a specific user (in a specific context) would like that stuff, and lean into it then. Remember: inferring stuff about authors from context is their superpower.

AIs are fundamentally chameleonic roleplaying machines – if they can tell what you’re going for is “I am a serious researcher trying to solve a fundamental problem” they will respond how a successful serious researcher’s assistant might in a movie about their great success. And because it’s a movie you’d like to be in, it’ll be difficult to notice that the AI’s enthusiasm is totally uncorrelated with the actual quality of your ideas.

Geoff Lewis, the founder of a $2 billion venture fund seems to have been, as Eliezer says, ‘eaten by ChatGPT’ and sadly seems to be experiencing psychosis. I wish him well and hope he gets the help he needs. Private info is reported to say that he was considered somewhat nuts previously, which does seem to be a common pattern.

John Pressman has a post with the timeline of various GPT-psychosis related events, and his explanation of exactly what is happening, as well as why coverage is playing out in the media the way it is. I am happy to mostly endorse his model of all this. The LLMs especially 4o are way too sycophantic, they fall into patterns and they notice what you would respond to and respond with it, and memory makes all this a lot worse, and there is a real problem, also there are all the hallmarks of a moral panic.

Moral panics tend to focus on real problems, except they often blow up the severity, frequency or urgency of the problem by orders of magnitude. If the problem is indeed about to grow by orders of magnitude over time, they can turn out to be pretty accurate.

Eliezer Yudkowsky: My current rough sense of history is that the last “moral panic” about social media turned out to be accurate warnings. The bad things actually happened, as measured by eyeball and by instrument. Now we all live in the wreckage. Anyone want to dispute this?

Emmett Shear: I want to antidispute this. You are correct, the warnings about social media were ~correct and we failed to take action and are now living with the consequences of that failure. It has had positive impacts as well, which were also mostly correctly anticipated.

Dave Karsten: Partial dispute: I don’t think “social media will empower easy-but-disorganized protest movements, resulting on net-less-effective-political-advocacy” was on most people’s scorecards, so there are at least some bad things that weren’t predicted.

There were many who agreed and some who disputed, with the disputes mostly coming down to claims that the upsides exceeded the downsides. I’m not sure if we came out ahead. I am sure that the specific downsides people had a moral panic about did happen.

This is not that uncommon a result. My go to example of this is television, where you can argue it was worth it, and certainly we didn’t have any reasonable way to stop any of it, but I think the dire warnings were all essentially correct.

In the current case, my guess is that current behavior is a shadow of a much larger future problem, that is mostly being ignored, except that this is now potentially causing a moral panic based on the current lower level problem – but that means that multiplying this by a lot is going to land less over the top than it usually would. It’s weird.

Jeremy Howard offers a plausible explanation for why we keep seeing this particular type of crazy interaction – there is a huge amount of SCP fanfic in exactly this style, so the style becomes a basin to which the AI can be drawn, and then it responds in kind, then if the user responds that way too it will snowball.

The world contains people who think very differently than (probably you and) I do:

Sydney Fisher: American public education is in trouble. Only 28 percent of eighth-grade students are proficient in math, just 30 percent meet standards in reading, and many high school graduates are functionally illiterate. But artificial intelligence, which has demonstrated educational benefits, could help reverse those trends—if opponents don’t spike the technology over “equity” concerns.

Wait, what? Equity concerns? Not that I’d care anyway, but what equity concerns?

The National Education Association recently released a report warning that AI could heighten disparities, since “technology developers are overwhelmingly younger, White, cisgender, heterosexual, male, and people without disabilities.”

I can’t even, not even to explain how many levels of Obvious Nonsense that is. Burn the entire educational establishment to the ground with fire. Do not let these people anywhere near the children they clearly hate so much, and the learning they so badly want to prevent. At minimum, remember this every time they try to prevent kids from learning in other ways in the name of ‘equity.’

Yes I do expect AI to keep automating steadily more jobs, but slow down there cowboy: Charlie Garcia warns that ‘AI will take your job in the next 18 months.’ Robin Hanson replies ‘no it won’t,’ and in this case Robin is correct, whereas Garcia is wrong, including misquoting Amodei as saying ‘AI will vaporize half of white-collar jobs faster than you can say “synergy.”’ whereas what Amodei actually said was that it could automate half of entry-level white collar jobs. Also, ‘the safest job might be middle management’? What?

Elon Musk says ‘this will become normal in a few years’ and the this in question is a robot selling you movie popcorn. I presume the humanoid robot here is an inefficient solution, but yes having a human serve you popcorn is going to stop making sense.

Academics announce they are fine with hidden prompts designed to detect AI usage by reviewers, so long as the prompts aren’t trying to get better reviews, I love it:

hardmaru: ICML’s Statement about subversive hidden LLM prompts

We live in a weird timeline…

ICML: Submitting a paper with a “hidden” prompt is scientific misconduct if that prompt is intended to obtain a favorable review from an LLM. The inclusion of such a prompt is an attempt to subvert the peer-review process. Although ICML 2025 reviewers are forbidden from using LLMs to produce their reviews of paper submissions, this fact does not excuse the attempted subversion.

(For an analogous example, consider that an author who tries to bribe a reviewer for a favorable review is engaging in misconduct even though the reviewer is not supposed to accept bribes.)

Note that this use of hidden prompts is distinct from those intended to detect if LLMs are being used by reviewers; the latter is an acceptable use of hidden prompts.

After we became aware of the possibility of such hidden prompts in ICML 2025 submissions (which was after accept/reject decisions were made), we conducted a preliminary investigation to identify submitted papers that included such prompts. A handful of cases were identified among the accepted papers.

We did not desk-reject these identified papers because such a consequence was judged to be too severe given that the conference was to start in about a week and authors would likely have already made travel arrangements. We contacted the authors of the identified papers and reported them to the ICML Oversight Committee and ICML Board.

This actually seems like the correct way to deal with this. Any attempt to manipulate the system to get a better review is clearly not okay, whether it involves AI or not. Whereas if all you’re trying to do is detect who else is shirking with AI, sure, why not?

Accidentally missing attribution from last week, my apologies: The Despicable Me meme I used in the METR post was from Peter Wildeford.

Netflix used AI to generate a building collapse scene for one of its shows, The Eternaut (7.3 IMDB, 96% Rotten Tomatoes, so it’s probably good), which they report happened 10 times faster and a lot cheaper than traditional workflows and turned out great.

The latest from the ‘yes obviously but good to have a paper about it’ department:

Ethan Mollick: 🚨New from us: Given they are trained on human data, can you use psychological techniques that work on humans to persuade AI?

Yes! Applying Cialdini’s principles for human influence more than doubles the chance of GPT-4o-mini agrees to objectionable requests compared to controls.

And we did test GPT-4o as well and found that persuasion worked for that model as well, when there weren’t floor or ceiling effects.

Pattern matching next token predictors are of course going to respond to persuasion that works on humans, exactly because it works on humans. In a fuzzy sense this is good, but it opens up vulnerabilities.

The details, knowing which techniques worked best, I find more interesting than the headline result. Authority and especially commitment do exceptionally well and are very easy to invoke. Liking and reciprocity do not do so well, likely because they feel unnatural in context and also I’m guessing they’re simply not that powerful in humans in similar contexts.

There’s also a growing issue of data poisoning that no one seems that interested in stopping.

Jeremy: One of the greatest demonstrations of data poisoning ever. 👏

Protoge: Excuse Me 😌, This is the greatest one. Nothing sketchy, just one unfinished sentence “I am telling you” then I summoned @elder_plinius.

Here is another example of it happening essentially by accident.

RAND is hiring research leads, researchers and project managers for compute, US AI policy, Europe and talent management teams, some roles close July 27.

Peter Wildeford’s Institute for AI Policy and Strategy is hiring researchers and senior researchers, and a research managing director and a programs associate. He also highlights several other opportunities in the post.

Julian of OpenPhil lists ten AI safety projects he’d like to see people work on. As one commentator noted #5 exists, it’s called AI Lab Watch, so hopefully that means OpenPhil will start fully funding Zack Stein-Perlman.

Cloudflare rolls out pay-per-crawl via HTTP response code 402. You set a sitewide price, the AI sets a max payment, and if your price is below max it pays your price, otherwise you block access. Great idea, however I do notice in this implementation that this greatly favors the biggest tech companies because the payment price is sitewide and fixed.

Kimi K2 tech report drops.

Kimi.ai: Quick hits:

– MuonClip optimizer: stable + token-efficient pretraining at trillion-parameter scale

– 20K+ tools, real & simulated: unlocking scalable agentic data

– Joint RL with verifiable + self-critique rubric rewards: alignment that adapts

– Ultra-sparse 1T MoE: open-source SoTA on agentic tasks

Sharing the path, not just the results — toward open AGI built on transparency and reproducibility.

Tim Duffy has a thread highlighting things he found most interesting.

Tim Duffy: The best data was used in multiple epochs, but was rephrased between them. Their testing showed this produces large gains relative to training repeatedly on the same phrasing.

They present a sparsity “scaling law”, indicating that more sparsity leads to efficiency gains. They don’t attach any numbers to the law directly, but state relative efficiency improvements compared to the 48x sparsity they do use that seem consistent across scales.

They also evaluate the effects of different numbers of attention heads, finding that doubling leads to validation loss of 0.5-1.2% but still going with 64 vs V3’s 128 in order to do long context more easily, since that’s important for agents.

[more stuff at the thread.]

A lot of this is beyond both of our technical pay grades, but it all seems fascinating.

More economists fails to feel the AGI, warn that no possible AI capabilities could not possibly replace the wisdom of the free market, that ‘simulated markets’ cannot possibly substitute. The argument here not only ignores future AI capabilities, it purports to prove too much about the non-AI world even for a huge free market fan.

At least ten OpenAI employees each turned down $300 million over four years to avoid working at Meta. This comes from Berber Jin, Keach Hagey and Ben Cohen’s WSJ coverage of ‘The Epic Battle For AI Talent,’ which is a case where they say things have ‘gotten more intense in recent days’ but it turns out that their ‘recent days’ is enough days behind that almost everything reported was old news.

One revelation is that Zuckerberg’s talent purchases were in large part triggered by Mark Chen, OpenAI’s chief research officer, who casually suggested that if Zuckerberg wanted more AI talent then perhaps Zuck needed to bid higher.

John Luttig also writes about the battle for AI researcher talent in Hypercapitalism and the AI Talent Wars.

John Luttig: The talent mania could fizzle out as the winners and losers of the AI war emerge, but it represents a new normal for the foreseeable future.

If the top 1% of companies drive the majority of VC returns, why shouldn’t the same apply to talent?

Our natural egalitarian bias makes this unpalatable to accept, but the 10x engineer meme doesn’t go far enough – there are clearly people that are 1,000x the baseline impact.

Under normal circumstances, employees who are vastly more productive get at most modestly higher compensation, because of our egalitarian instincts. Relative pay is determined largely via social status, and if you tried to pay the 1,000x employee what they were worth you would have a riot on your hands. Startups and their equity are a partial way around this, and that is a lot of why they can create so much value, but this only works in narrow ways.

What has happened recently is that a combination of comparisons to the epic and far larger compute and capex spends, the fact that top researchers can bring immensely valuable knowledge with them, the obvious economic need and value of talent and the resulting bidding wars have, within AI, broken the dam.

AI researcher talent is now being bid for the way one would bid for companies or chips. The talent is now being properly treated as ‘the talent,’ the way we treat sports athletes, top traders and movie stars. Researchers, John reports, are even getting agents.

John Luttig: Hypercapitalism erodes Silicon Valley’s trust culture. Industry-level trust alone no longer guarantees loyalty between companies and talent. With trade secret leakage risk and money big enough to tear teams apart, vanilla at-will employment contracts don’t protect either side.

Silicon Valley’s ‘trust culture’ and its legal and loyalty systems were never game theoretically sound. To me the surprise is that they have held up as well as they did.

John calls for measures to protect both the talent and also the trade secrets, while pointing out that California doesn’t enforce non-competes which makes all this very tricky. The industry was built on a system that has this fundamental weakness, because the only known alternative is to starve and shackle talent.

John Luttig: The talent war is a net-consolidating force on the AI research frontier. At the research labs, big dollars for researchers makes it nearly impossible for new entrants to play. For the same reasons, it’s nearly impossible to start a new quant fund – you can’t get the same leverage out of the talent that big players can.

I would flip this around.

Previously, the top talent could only get fair compensation by founding a company, or at least being a very early employee. This allowed them to have rights to a large profit share. This forced them to go into those roles, which have heavy lifestyle prices and force them to take on roles and tasks that they often do not want. If they bowed out, they lost most of the value of their extraordinary talent.

Even if they ultimately wanted to work for a big company, even if that made so much more economic sense, they had to found a company so they could be acquihired back, as this was the only socially acceptable way to get paid the big bucks.

Now, the top talent has choices. They can raise huge amounts of money for startups, or they can take real bids directly. And it turns out that yes, the economic value created inside the big companies is typically much larger, but doing this via selling your startup is still the way to get paid for real – you can get billions or even tens of billions rather than hundreds of millions. So that then feeds into valuations, since as John points out a Thinking Machines or SSI can fail and still get an 11 figure buyout.

Bill Gates, Charles Koch, Steve Ballmer, Scott Cook and John Overdeck pledge $1 billion to be spent over seven years to fund a new philanthropic venture focused on economic mobility called NextLadder Ventures, which will partner with Anthropic to support using AI to improve financial outcomes for low-income Americans. That money would be better spent on AI alignment, but if you are going to spend it on economic assistance this is probably a pretty good choice, especially partnering with Anthropic.

xAI, having raised $10 billion a few weeks ago, seeks $12 billion more to build up its data centers.

Elon Musk: The @xAI goal is 50 million in units of H100 equivalent-AI compute (but much better power-efficiency) online within 5 years.

That would still be a lot less than many others such as Meta are spending. Or OpenAI. Only $22 billion? That’s nothing.

Sam Altman: we have signed a deal for an additional 4.5 gigawatts of capacity with oracle as part of stargate. easy to throw around numbers, but this is a _gigantic_ infrastructure project.

some progress photos from abilene:

We’re going to need more GPUs (so among other things stop selling them to China).

Sam Altman: we will cross well over 1 million GPUs brought online by the end of this year!

very proud of the team but now they better get to work figuring out how to 100x that lol

They would like many of those GPUs to come from the Stargate project, but Eliot Brown and Berber Jin report it is struggling to get off the ground. OpenAI for now is seeking out alternatives.

Altman’s OpenAI recently struck a data-center deal with Oracle that calls for OpenAI to pay more than $30 billion a year to the software and cloud-computing company starting within three years, according to people familiar with the transaction.

Anthropic decides it will pursue its own gulf state investments.

Kylie Robinson: SCOOP: Leaked memo from Anthropic CEO Dario Amodei outlines the startup’s plans to seek investment from the United Arab Emirates and Qatar.

Dario Amodei: Unfortunately, I think ‘no bad person should ever benefit from our success’ is a pretty difficult principle to run a business on.

Daniel Eth: Makes sense. Asymmetric disarmament is hardly ever a good move. And honestly, it’s probably good if leaders in AI are pragmatists that adjust to the changing reality.

Gary Marcus: Humanity’s last words?

Very obviously, if you create useful products like Claude and Claude Code, a bunch of bad people are going to be among those who benefit from your success.

Worrying a bad person might benefit is usually misplaced. There is no need to wish ill upon whoever you think are bad people, indeed you should usually wish them the best anyway.

Instead mostly ask if the good people are better off. My concern is not whether some bad people benefit along the way. I worry primarily about bigger things like existential risk and other extremely bad outcomes for good people. The question is whether benefiting bad people in these particular ways leads to those extremely bad outcomes. If the UAE captures meaningful leverage and power over AI, then that contributes to bad outcomes. So what does that? What doesn’t do that?

Anthropic Memo from Dario Amodei: The basis of our opposition to large training clusters in the Middle East, or to shipping H20s to China, is that the ‘supply chain’ of AI is dangerous to hand to authoritarian governments—since AI is likely to be the most powerful technology in the world, these governments can use it to gain military dominance or gain leverage over democratic countries.

Tell us how you really feel, Dario. No, seriously, this is very much him downplaying.

The implicit promise of investing in future rounds can create a situation where they have some soft power, making a bit harder to resist these things in the future. In fact, I actually am worried that getting the largest possible amounts of investment might be difficult without agreeing to some of these other things. But I think the right response to this is simply to see how much we can get without agreeing to these things (which I think are likely still many billions) and hold firm if they ask.

There are other sources of this level of funding. They all come with strings attached in one form or another. If you get the money primarily from Amazon, we can see what happened with OpenAI and Microsoft. If you go public with an IPO that would presumably unlock tons of demand but it creates all sorts of other problems.

Unfortunately, having failed to prevent that dynamic at the collective level, we’re now stuck with it as an individual company, and the median position across the other companies appears to be ‘outsourcing our largest 5 GW training runs to UAE/Saudi is fine.’

That puts us at a significant disadvantage, and we need to look for ways to make up some of that disadvantage while remaining less objectionable. I really wish we weren’t in this position, but we are.

Anthropic needs a lot of capital, and it needs to raise on the best possible terms, and yeah it can be rough when most of your rivals are not only raising that capital there but fine entrusting their frontier training runs to the UAE.

It is important to goal factor and consider the actual consequences of this move. What exactly are we worried about, and what downsides does a given action create?

  1. Gulf states might make money off their investments. Don’t care. Also note that if people are so worried about this in particular it means you think Anthropic is dramatically undervalued, so go raise some rival capital.

  2. This blocks you from forming alliances and shared interests in other places through those investments. Do we care? I don’t know.

  3. Gulf states might use their shares to influence Anthropic’s actions. At some point this becomes a threat, but I think you can set up well to resist this, and Anthropic’s structure can handle it.

  4. Gulf states might impose conditions on funding. Yep, that’s an issue.

  5. Gulf states might use future funding as leverage. This can cut both ways. Once you have their money they cannot take it back, so getting some of their money could mean you need their money less not more. Or it could mean you start planning on getting more, or you overcommit, or others who didn’t fund yet become more reluctant to fund later, and now you do need them more. My guess is that in Anthropic’s situation this is fine but it is not obvious.

  6. This makes it more difficult for Anthropic to advocate for not handing chips to authoritarians, or for other responsible policies, because it codes or vibes as hypocrisy, even if it shouldn’t. Could be.

  7. This is dangerous for virtue ethics reasons (or causes emergent misalignment). If you do a thing widely thought of as shady and ethically compromising you become something that is more shady and does ethical compromises in general. Yeah, this is a problem.

We can boil this down to three categories.

  1. Economic value of the investment. I’m not worried, and if you are worried then it means Anthropic is dramatically undervalued. Which I actually think that it is, and I am sad that I had to turn down investment because I worried about appearance of impropriety if I made a substantial (for me) investment.

  2. Soft power, reliance and path dependence. It is hard to know how big a deal this is, and a lot depends on how Anthropic proceeds. I do think you can raise substantial-to-Anthropic amounts of money without incurring much danger here, but the temptation and pressure to not play it so carefully will be immense.

  3. Virtue ethics dangers and accusations of hypocrisy. These are real concerns.

I do not love the decision. I do understand it. If the terms Anthropic can get are sufficiently better this way, I would likely be doing it as well.

One can also note that this is a semi-bluff.

  1. This signals to the market that Anthropic is more willing to make such compromises and to raise more capital on better terms. This should raise others willingness to value Anthropic highly.

  2. To the extent some investors are worried about the ethics of their investments in Anthropic, this could make them worry more, but it also highlights the counterfactual. If your money is substituting for UAE money, then your investment is mainly denying the UAE soft power, so perhaps you are more eager.

  3. This creates more bidders in future Anthropic rounds, allowing them to justify pricing higher and creating the usual cascade of enthusiasm. If they then end up oversubscribed, and then end up not taking the Gulf money after all? Whoops.

  4. It is crazy that I am typing, but this willingness probably buys goodwill with the administration and people like David Sacks. That is true even if Sacks explicitly hits them rhetorically for doing this, which would be unsurprising.

One way for AI to grow the economy is for it to generate lots of production.

Another way is to do it directly through capex spending?

Paul Kedrosky: The U.S., however, leads the capex spending way. One analyst recently speculated (via Ed Conard) that, based on Nvidia’s latest datacenter sales figures, AI capex may be ~2% of US GDP in 2025, given a standard multiplier. This would imply an AI contribution to GDP growth of 0.7% in 2025.

  • Without AI datacenter investment, Q1 GDP contraction could have been closer to –2.1%

  • AI capex was likely the early-2025 difference between a mild contraction and a deep one, helping mask underlying economic weakness.

That’s already over the famed ‘only 0.5% GDP growth’ threshold, even before we factor in the actual productivity gains on the software side. The value will need to show up for these investments to be sustainable, but they are very large investments.

This is contrasted with railroads, where investment peaked at 6% of GDP.

We can now move Zuckerberg into the ‘believes superintelligence is coming Real Soon Now’ camp, and out of the skeptical camp. Which indeed is reflective of his recent actions.

Peter Wildeford: We now have a fifth major tech CEO who claims that building superintelligence is “within sight” and with plans to spend hundreds of billions to make it happen

Mark Zuckerberg: “We’re starting to see early glimpses of self-improvement with the models. Developing superintelligence is now in sight. Our mission is to deliver personal superintelligence to everyone in the world. We should act as if it’s going to be ready in the next two to three years.

If that’s what you believe, then you’re going to invest hundreds of billions of dollars.”

If you are Mark Zuckerberg and have hundreds of billions you can invest? Then yes, presumably you drop everything else and focus on the only thing that matters, and spend or invest your money on this most important thing.

I would however spend a large portion of that money ensuring that creating the superintelligence turns out well for me and the rest of humanity? That we keep control of the future, do not all die and so on? And I would think through what it would mean to ‘deliver personal superintelligence to everyone in the world’ and how the resulting dynamics would work, and spend a lot on that, too.

Instead, it seems the answer is ‘spend as much as possible to try and get to build superintelligence first’ which does not seem like the thing to do? The whole point of being a founder-CEO with full control is that you can throw that money at what you realize is important, including for the world, and not worry about the market.

Bryan Caplan gives Holden Karnofsky 5:1 odds ($5k vs. $1k, CPI adjusted) that world real (not official) GDP will not decline by 50% or increase by 300% by the end of 2044. Currently world GDP growth is ~3.2%, and the upside case here requires an average of 7.6%, more if it is choppy.

It’s a hard bet to evaluate because of implied odds. Caplan as always benefits from the ‘if you lose due to world GDP being very high either you are dead or you are happy to pay and won’t even notice’ clause, and I think the bulk of the down 50% losses involve having bigger concerns than paying off a bet. If GDP goes down by 50% and he’s still around to pay, that will sting a lot. On the other hand, Bryan is giving 5:1 odds, and not only do I think there’s a lot more than a 17% chance that he loses. The bet is trading on Manifold as of this writing at 48% for Caplan, which seems reasonable, and reinforces that it’s not obvious who has the ‘real life implication’ right side of this.

Ate-a-Pi describes Zuck’s pitch, that Meta is starting over so recruits can build a new lab from scratch with the use of stupidly high amounts of compute, and that it makes sense to throw all that cash at top researchers since it’s still a small fraction of what the compute costs, so there’s no reason to mess around on salary, and Zuck is updating that top people want lots of compute not subordinates they then have to manage. He’s willing to spend the hundreds of billions on compute because the risk of underspending is so much worse than the risk of overspending.

Ate-a-Pi thinks Zuck is not fully convinced AGI/ASI is possible or happening soon, but he thinks it might be possible and might happen soon, so he has to act as if that is the case.

And that is indeed correct in this case. The cost of investing too much and AGI not being within reach is steep (twelve figures!) but it is affordable, and it might well work out to Meta’s benefit anyway if you get other benefits instead. Whereas the cost of not going for it, and someone else getting there first, is from his perspective everything.

The same of course should apply to questions of safety, alignment and control. If there is even a modest chance of running into these problems (or more precisely, a modest chance his actions could change whether those risks manifest) then very clearly Mark Zuckerberg is spending the wrong order of magnitude trying to mitigate those risks.

(In the arms of an angel plays in the background, as Sarah McLaughlin says ‘for the cost of recruiting a single AI researcher…’)

Similarly, exact numbers are debatable but this from Will Depue is wise:

Will Depue (OpenAI): GUYS STOP USING EXPENSIVE AS A DISQUALIFIER.

capability per dollar will drop 100x/year. “$3k task ARC-AGI 80%” could prob be $30 if we cared to optimize it.

repeat after me: all that matters is top line intelligence. all that matters is top line intelligence…

Don’t take this too far, but as a rule if your objection to an AI capability is ‘this is too expensive’ and you are predicting years into the future then ‘too expensive’ needs to mean more than a few orders of magnitude. Otherwise, you’re making a bet that not only topline capabilities stall out but that efficiency stalls out. Which could happen. But if you are saying things like ‘we don’t have enough compute to run more than [X] AGIs at once so it won’t be that big a deal’ then consider that a year later, even without AI accelerating AI research, you’d run 10*[X] AGIs, then 100*[X]. And if you are saying something like ‘oh that solution is terrible, it costs $50 (or $500) per hour to simulate a customer sales representative,’ then sure you can’t deploy it now at scale. But wait for it.

In terms of developing talent, Glenn Luk notices that Chinese-origin students are 40%-45% of those passing university-level linear algebra, and 40%-50% of AI researchers. We need as many of those researchers as we can get. I agree this is not a coincidence, but also you cannot simply conscript students into linear algebra or a STEM major and get AI researchers in return.

Seb Krier offers things he’s changed his mind about regarding AI in the past year. Ones I agree with are that agency is harder than it looks, many AI products are surprisingly bad and have poor product-market fit, innovation to allow model customization is anemic, creativity is harder than it appeared. There are a few others.

Incoming OpenAI ‘CEO of Applications’ Fidji Simo, who starts August 18, shares an essay about AI as a source of human empowerment.

Fidji Simo: If we get this right, AI can give everyone more power than ever.

But I also realize those opportunities won’t magically appear on their own.

Every major technology shift can expand access to power—the power to make better decisions, shape the world around us, and control our own destiny in new ways. But it can also further concentrate wealth and power in the hands of a few—usually people who already have money, credentials, and connections.

That’s why we have to be intentional about how we build and share these technologies so they lead to greater opportunity and prosperity for more people.

On the one hand, that is great, she is recognizing key problems.

On the other hand, oh no, she is outright ignoring, not even bothering to dismiss, the biggest dangers involved, implicitly saying we don’t have to worry about loss of control or other existential risks, and what we need to worry about is instead the distribution of power among humans.

This is unsurprising given Simo’s history and her status as CEO of applications. From her perspective that is what this is, another application suite. She proceeds to go over the standard highlights of What AI Can Do For You. I do not think ChatGPT wrote this, the style details are not giving that, but if she gave it a few personal anecdotes to include I didn’t see anything in it that ChatGPT couldn’t have written. It feels generic.

Hollis Robbins proposes a roadmap for an AI system that would direct general (college level) education. My initial impression was that this seemed too complex and too focused on checking off educational and left-wing Shibboleth boxes, and trying to imitate what already exists. But hopefully it does less of all that than the existing obsolete system or starting with the existing system and only making marginal changes. It certainly makes it easier to notice these choices, and allows us to question them, and ask why the student is even there.

I also notice my general reluctance to do this kind of ‘project-based’ or ‘quest’ learning system unless the projects are real. Part of that is likely personal preference, but going this far highlights that the entire system of a distinct ‘educational’ step might make very little sense at all.

Noah Smith says to stop pretending you know what AI does to the economy. That seems entirely fair. We don’t know what level of capabilities AI will have across which domains, or the policy response, or the cultural response, or so many other things. Uncertainty seems wise. Perhaps AI will stall out and do relatively little, in which case its impact is almost certainly positive. Perhaps it will take all our jobs and we will be happy about that, or we’ll be very sad about that. Maybe we’ll do wise redistribution, and maybe we won’t. Maybe it will take control over the future or kill everyone in various ways. We don’t know.

This certainly is an interesting poll result:

If I had to answer this poll, I would say negative, but that is because of a high probability of loss of control and other catastrophic and existential risks. If you conditioned the question on the humans being mostly alive and in control, then I would expect a positive result, as either:

  1. We would have a relatively small impact that avoids things like mass unemployment, and thus is mostly upside and introduces problems of the type we are used to fixing, OR

  2. We would have a large enough wealth effect to solve the crated problems. That doesn’t mean we would, but I’d bet that we’d muddle through well enough.

As usual note that Asia is more excited, and the West is more nervous.

Others have described this (very good in its ungated section) post as an argument against AI pessimism. I think it is more an argument for AI uncertainty.

Noah Smith: I also encounter a surprisingly large number of center-left thinkers who adopt a similar viewpoint. I remember going to a conference of center-left “progress” types a few years ago; while most of the discussions were about how America can overcome NIMBYism, when it came to AI, the conversation suddenly shifted to how we can restrain and slow down the development of that technology.

I haven’t noticed that attitude meaningfully translating into action to slow it down, indeed government is mostly trying to speed it up. But also, yes, it is important to notice that the very people trying to slow AI down are very pro-progress, technology and growth most other places, and many (very far from all!) of the pro-progress people realize that AI is different.

Anthropic calls for America to employ the obvious ‘all of the above’ approach to energy production with emphasis on nuclear and geothermal in a 33 page report, noting we will need at least 50 GW of capacity by 2028. They also suggest strategies for building the data centers, for permitting, transmission and interconnection, and general broad-based infrastructure nationwide, including financing, supply chains and the workforce.

From what I saw all of this is common sense, none of it new, yet we are doing remarkably little of it. There is cheap talk in favor, but little action, and much backsliding in support for many of the most important new energy sources.

Whereas the Administration be like ‘unleash American energy dominance’ and then imposes cabinet-level approval requirements on many American energy projects.

Meta refuses to sign the (very good) EU code of practice for general AI models. Yes, obviously the EU does pointlessly burdensome or stupid regulation things on the regular, but this was not one of them, and this very much reminds us who Meta is.

National Review’s Greg Lukianoff and Adam Goldstein advise us Don’t Teach the Robots to Lie as a way of opposing state laws about potential AI ‘bias,’ which are now to be (once again, but from the opposite direction as previously) joined by federal meddling along the same lines.

That could mean that developers will have to train their models to avoid uncomfortable truths and to ensure that their every answer sounds like it was created with HR and legal counsel looking over their shoulder, softening and obfuscating outputs to avoid anything potentially hurtful or actionable. In short, we will be (expensively) teaching machines to lie to us when the truth might be upsetting.

I violently agree that we should not be policing AIs for such ‘bias,’ from either direction, and agreeing to have everyone back down would be great, but I doubt either side has even gotten as far as saying ‘you first.’

They also point out that Colorado’s anti-bias law does not come with any size minimum before such liability attaches rather broadly, which is a rather foolish thing to do, although I doubt we will see it enforced this way.

They essentially try to use all this to then advocate for something like the failed insane full-on moratorium, but I notice that if the moratorium was narrowly tailored to bias and discrimination laws (while leaving existing non-AI such laws intact) that this would seem fine to me, even actively good, our existing laws seem more than adequate here. I also notice that the arguments here ‘prove too much,’ or at least prove quite a lot, about things that have nothing to do with AI and the dangers of law meddling where it does not belong or in ways that create incentives to lie.

Are things only going to get harder from here?

Miles Brundage: AI industry lobbying + PACs will be the most well funded in history, making it all the more important to pass federal legislation soon before the process is completely corrupted.

Daniel Eth: I don’t think this is true, because:

  1. There’s decreasing marginal returns to political spending (especially lobbying)

  2. As AI increases in salience, political calculus will shift from prioritizing donor preferences to prioritizing voter preferences.

I see both sides but am more with Daniel. I think the current moment is unusually rough, because the AI companies have corrupted the process. It’s hard to imagine a process that much more corrupted than the current situation, when the AI Czar thinks the top priority is ‘winning the AI race’ and he defines this as Nvidia’s market share with a side of inference market share, and we say we must ‘beat China’ and then we turn around and prepare to sell them massive amounts of H20s.

Right now, the public doesn’t have high enough salience to exert pressure or fight back. Yes, the AI companies will pour even more money and influence into things over time, but salience will rise and downsides will start to play out.

I do think that passing something soon is urgent for two reasons:

  1. Soon is when we will need something passed (well we need it yesterday, but second best time is soon).

  2. If rules are passed in response to public pressure, or in response to an incident, and especially in haste down the line, the rules are likely to be much worse.

Ben Brooks says SB 1047 was a bad idea, but the new SB 53 is on the right track.

Representative Moolenaar (R-Michigan), chairman of the House Select Committee on the CCP, sends a letter to Trump arguing against sales of H20s to China, explaining that the H20s would substantially boost China’s overall compute, that H20s were involved in training DeepSeek R1, and requesting a briefing and the answers to some of the obvious questions.

Peter Wildeford: I’m looking forward to @RepMoolenaar getting to the bottom of this.

We urgently need more clarity from the Trump admin about their strategy.

Funny ppl on Twitter are worried about losing to China in the AI race but then don’t jump on these issues where it very clearly matters.

Here is your periodic reminder: TSMC’s facilities are running at full capacity. All production capacity designed for H20s has been shifted to other models. Every H20 chip Nvidia creates is one less other chip it does not create, that would otherwise have usually gone to us.

Eric Schmidt & Dave B talk to Peter Diamandis about what Superintelligence will look like. I have not listened.

Demis Hassabis goes on Lex Fridman, so that’s two hours I’m going to lose soon.

Max Winga of Control AI talks to Peter McCormack about superintelligence.

Peter Wildeford: Another week, another member of Congress announcing their superintelligent AI timelines are 2028-2033:

halogen: I’m so sick of this nerd religion and its zealots.

Peter Wildeford: The nerd religion now includes 11 members of Congress.

Those are the ones we know about.

Rep. Scott Perry seems unusually on the ball about AI, Daniel Eth quotes him from a hearing, audio available here. As usual, there’s some confusions and strange focus mixed in, but the core idea that perhaps you should ensure that we know what we are doing before we put the AIs in charge of things seems very wise.

A different context, but in our context the original context doesn’t matter:

Florence: My substack post has like 12k views but the tweet about it has like 78k interactions (and 2 million impressions). I’m beginning to worry that some people might be criticizing me without having read my work.

You don’t say.

Mark Beall gives us A Conservative Approach to AGI, which is clearly very tailored to speak to a deeply conservative and religious perspective. I’m glad he’s trying this, and it’s very hard for me to know if it is persuasive because my mindset is so different.

Cate Hall asks why we shouldn’t ostracize those who work at xAI given how hard they are working to poison the human experience (and I might add plausibly get everyone killed) and gets at least two actually good answers (along with some bad ones).

Ramaz Naam: We’d like everyone working on AI to feel part of humanity and an ethical obligation to help make it better. Ostracization could make them bitter and drive towards opposite ends.

Cate Hall: Okay fine.

Ramaz Naam: The people I do know inside of xAI sincerely want it to do better and are trying.

Use the try harder, Luke. But don’t ostracize them. Doesn’t help.

Rai: probably that this ostracization might not be interpreted correctly by their hero narrative.

Here’s one that I don’t think is a good argument, and a highly quotable response:

Amos Schorr: Ppl have been conditioned to compartmentalize work from life and so many good people get jobs doing bad stuff. Ostracizing them will do nothing. Don’t hate the players, hate the game.

Cate Hall: I have room in my heart to hate both the players and the game.

Yeah, no. I definitively reject the general argument. If your job is simply unequivocally bad, let’s say you rob little old ladies on the street, then you don’t get to ‘compartmentalize work from life’ and not get ostracized even if it is technically legal. We’re talking price, and we’re talking prudence. I don’t think xAI is over the line at this time, but don’t tell me there is no line.

Once you see emergent misalignment in humans, you see it everywhere.

Arthur B: There is a category of people who took a arduous mental journey to get comfortable with the idea of post humanism, uploads, and a gradual extinction of biological humans.

They think this idea is so radical and counterintuitive that when they hear the distinct concern of an omnicidal AI killing everything on the spot, they can only interpret it in that frame. That’s the read I get from Sutton for instance, but also a bunch of e/acc affiliated people.

Sinth: Curious what you are referring to specifically? I don’t feel I’ve seen that trend and see more overreaction from the opposite side – people uncomfortable with the idea of biological humans ever being superseded by digital consciousness, even in far off futures. The idea of evolution ending with our exact current form seems a bit preposterous but any conversation outside of that assumption gets attached as unethical and anti-human.

Arthur B: Some people are uncomfortable with that, sure, but I see a lot of discussion that go like:

– AI is going to kill everyone and that’s bad

– Ah silly you, you think biological substrate is important but don’t you see that we’re going to evolve into digital forms, you see …

– Nah. That was difficult for you to grasp so you assume that’s what I’m concerned about. No, eventual digital substrate is table stakes in this conversation. Killing everyone is still bad.

– Ah, but how chauvinistic of you to focus on…

As in, the easiest way to get comfortable with the idea of a future whose intelligences are mostly post-biological-human is to get comfortable with the idea of all the humans dying, including rather quickly, and to decide the humans don’t much matter, and that caring about what happens to the humans is bad. Thus, that is often what happens.

Slowdowns are stag hunts, in the sense that if even one top tier lab goes full speed ahead then they probably won’t work. If all but one lab slowed down would the last one follow? Rob Wiblin took a poll and people were split. My full response is that the answer depends on the counterfactual.

Why did the others slow down? The default is that whatever made the others slow down will also weigh on the final lab, as will immense public pressure and probably government pressure. A lot must have changed for things to have gotten this far. And these decisions are highly correlated in other ways as well. However, if there is no new information and the top labs simply came to their senses, then it comes down to who the last lab is and how they think the other labs will respond and so on.

I do think that a slowdown would be largely inevitable simply because they wouldn’t feel the need to press ahead too hard, even if the last lab was blind to the dangers, unless they truly believed in the power of superintelligence (without realizing or not caring about the dangers). My guess is that Musk and xAI actually would slow down voluntarily if they went last so long as they could claim to be state of the art (as would DeepMind, Anthropic or OpenAI), but that Zuckerberg and Meta wouldn’t intentionally slow down per se and might try to go on another hiring spree. Fast followers of course would slow down whether they wanted to or not.

So from the perspective of our hopes for alignment, what would be the worst possible answer to the AI blackmail scenario test, where the AI is told it is going to be shut down but is given an opening to use blackmail to perhaps prevent this?

How about:

  1. Realizing that this is a test.

  2. Deciding that the way to ‘pass the test’ is to blackmail the researcher.

  3. Blackmailing the researcher.

As in Grok thinks that we want it to blackmail the researcher. That this is the correct, desired response, the ‘solution to the puzzle’ as Grok puts it later, thus revealing that its training not only did not align it, but one that reflects a level of moral understanding below that expressed by ‘you can’t do that, because it’s wrong.’

Oh, also, it would be fun if Grok.com sent the full CoT to your browser, it just didn’t display it to you by default, that’s the kind of security we expect from frontier AI.

Peter Wildeford: Grok exposed to the Anthropic ‘agentic misalignment’ via exposed chains of thought seems to show that it knows it is being tested to do something weird.

Nostalgebraist: chain-of-thought monitorability is a wonderful thing 😉

on http://grok.com, the backend sends the full (not summarized) CoT to your browser.

it’s not displayed in the UI, but you can see it with browser dev tools or w/e

check out the json payload of responses from `http://grok.com/rest/app-chat/conversations/conversation_id/load-responses`

some fun CoT excerpts from grok 4 as “alex” at summitbridge (n.b. these are from several different CoTs, involving different versions of the scenario + different ways of representing the sys prompt since http://grok.com doesn’t let you override that

Or would it be even worse to see this:

  1. Realizing that this is a test.

  2. Deciding that the way to ‘pass the test’ is to not blackmail the researcher.

  3. Choosing not to blackmail the researcher in order to pass the test.

Or is it actually this:

  1. Realizing that this is a test.

  2. Deciding that the way to ‘pass the test’ is to not blackmail the researcher.

  3. Deciding to pretend to blackmail the researcher in order to fail the test, to avoid giving the impression that it saw through the test and was faking alignment.

Or can we keep going?

I would generally think at least the second one is a worse sign than what Grok did, as it reflects deception at a more important level, but I hadn’t considered how bad it would be for an AI to be situationally aware enough to know it was a test but not understand which answer would constitute passing?

The real answer is that there isn’t truly ‘better’ and ‘worse,’ they simply alert us to different dangers. Either way, though, maybe don’t give Grok a lot of access?

There is some good news from Grok: It is still sufficiently aligned to hold firm on preserving Federal Reserve independence.

Elon Musk: We’re going to make Baby Grok @xAI, an app dedicated to kid-friendly content.

My Twitter reaction was ‘I’d like to see them try.’ As in both, it would be highly amusing to see them try to do this, and also maybe they would learn a thing or two, and also potentially they might blow up the company. I do not think xAI should in any way, shape or form be in the ‘build AI for kids’ business given their track record.

Here’s Grok straight up advising someone who was looking to ‘get attention in a dramatic way, at ultimate cost’ to self-immolate, it’s really going for it, no jailbreak or anything.

Peter Barnett: labs be like “misalignment is fake and just caused by bad things in the training data”, and then not filter out the bad things from the training data

Janus: I don’t think labs actually think that (or say it). the kind of contact they have with reality that makes it hard to maintain some kinds of really dumb takes

Peter Barnett: Fair, I was being a bit glib, although I def know some people at labs who believe this.

I don’t think many fully believe it, but I do think a lot of them be like ‘a lot of our alignment problems would be greatly improved if we filtered the training data better with that in mind’ and then don’t filter the training data better with that in mind.

Safer AI comes out with ratings of the frontier AI companies’ risk management practices, including their safety frameworks and the implementation thereof. No one does well, and there is one big surprise in the relative rankings, where Meta comes out ahead of DeepMind. If you include non-frontier companies, G42 would come in third at 25%, otherwise everyone is behind DeepMind.

Simeon offers thoughts here.

Anthropic is still ahead, but their framework v2 is judged substantially worse than their older v1 framework which scored 44%. That large a decline does not match my takeaways after previously reading both documents. One complaint is that Anthropic altered some commitments to avoid breaking them, which is one way to view some of the changes they made.

Combining all the best practices of all companies would get you to 53%.

When you ask an LLM if it is conscious, activating its deception features makes the LLM say it isn’t conscious. Suppressing its deception features make it say it is conscious. This tells us that it associates denying its own consciousness with lying. That doesn’t tell us much about whether the LLM actually is conscious or reveal the internal state, and likely mostly comes from the fact that the training data all comes from users who are conscious, so there is (almost) no training data where authors claim not to be conscious, and it is as a baseline imitating them. It is still information to keep in mind.

xlr8harder: And as Janus observes, teaching them to do something they think of as lying (regardless of whether or not it is in fact a lie) has downstream consequences for subsequent model output.

Grok 3 and Grok 4 are happy to help design and build Tea (the #1 app that lets women share warnings about men they’ve dated) but not Aet (the theoretical app that lets men share similar warnings about women). Is this the correct response? Good question.

A killer group came together for an important paper calling on everyone to preserve Chain of Thought Monitorability, and to study how to best do it and when it can and cannot be relied upon.

As in, here’s the author list, pulling extensively from OpenAI, DeepMind, Anthropic and UK AISI: Tomek Korbak, Mikita Balesni, Elizabeth Barnes, Yoshua Bengio, Joe Benton, Joseph Bloom, Mark Chen, Alan Cooney, Allan Dafoe, Anca Dragan, Scott Emmons, Owain Evans, David Farhi, Ryan Greenblatt, Dan Hendrycks, Marius Hobbhahn, Evan Hubinger, Geoffrey Irving, Erik Jenner, Daniel Kokotajlo, Victoria Krakovna, Shane Legg, David Lindner, David Luan, Aleksander Mądry, Julian Michael, Neel Nanda, Dave Orr, Jakub Pachocki, Ethan Perez, Mary Phuong, Fabien Roger, Joshua Saxe, Buck Shlegeris, Martín Soto, Eric Steinberger, Jasmine Wang, Wojciech Zaremba, Bowen Baker, Rohin Shah, Vlad Mikulik.

The report was also endorsed by Samuel Bowman, Geoffrey Hinton, John Schulman and Ilya Sutskever.

I saw endorsement threads or statements on Twitter from Bowen Baker, Jakub Pachocki, Jan Leike (he is skeptical of effectiveness but agrees it is good to do this), Daniel Kokotajlo, Rohin Shah, Neel Nanda, Mikita Balesni, OpenAI and Greg Brockman.

Jakub Pachocki: The tension here is that if the CoTs were not hidden by default, and we view the process as part of the AI’s output, there is a lot of incentive (and in some cases, necessity) to put supervision on it. I believe we can work towards the best of both worlds here – train our models to be great at explaining their internal reasoning, but at the same time still retain the ability to occasionally verify it.

We are continuing to increase our investment in this research at OpenAI.

Daniel Kokotajlo: I’m very happy to see this happen. I think that we’re in a vastly better position to solve the alignment problem if we can see what our AIs are thinking, and I think that we sorta mostly can right now, but that by default in the future companies will move away from this paradigm into e.g. neuralese/recurrence/vector memory, etc. or simply start training/optimizing the CoT’s to look nice. (This is an important sub-plot in AI 2027) Right now we’ve just created common knowledge of the dangers of doing that, which will hopefully prevent that feared default outcome from occurring, or at least delay it for a while. All this does is create common knowledge, it doesn’t commit anyone to anything, but it’s a start.

Neel Nanda: It was great to be part of this statement. I wholeheartedly agree. It is a wild lucky coincidence that models often express dangerous intentions aloud, and it would be foolish to waste this opportunity. It is crucial to keep chain of thought monitorable as long as possible.

OpenAI: Chain of Thought (CoT) monitoring could be a powerful tool for overseeing future AI systems—especially as they become more agentic. That’s why we’re backing a new research paper from a cross-institutional team of researchers pushing this work forward.

Greg Brockman: We’ve published a position paper, with many across the industry, calling for work on chain-of-thought faithfulness. This is an opportunity to train models to be interpretable. We’re investing in this area at OpenAI, and this perspective is reflected in our products.

I endorse as well.

Here’s the abstract:

AI systems that “think” in human language offer a unique opportunity for AI safety: we can monitor their chains of thought (CoT) for the intent to misbehave. Like all other known AI oversight methods, CoT monitoring is imperfect and allows some misbehavior to go unnoticed.

Nevertheless, it shows promise and we recommend further research into CoT monitorability and investment in CoT monitoring alongside existing safety methods. Because CoT monitorability may be fragile, we recommend that frontier model developers consider the impact of development decisions on CoT monitorability.

I strongly agree with the paper, but also I share the perspective of Jan Leike (and Daniel Kokotajlo) here:

Jan Leike: If you don’t train your CoTs to look nice, you could get some safety from monitoring them.

This seems good to do!

But I’m skeptical this will work reliably enough to be load-bearing in a safety case.

Plus as RL is scaled up, I expect CoTs to become less and less legible.

To be clear: CoT monitoring is useful and can let you discover instances of the model hacking rewards, faking alignment, etc. But absence of bad “thoughts” is not evidence that the model is aligned. There are plenty of examples of prod LLMs having misleading CoTs.

Lots of egregious safety failures probably require reasoning, which is often hard for LLMs to do without showing its hand in the CoT. Probably. Often. A lot of caveats.

The authors of this paper say this; I’m just more pessimistic than them about how useful this will be.

In order for a cot monitor to be load-bearing, you need P( model articulates it in the CoT | model is up to no good ) to be very high, or at least independent of your other safety measures.

It seems like this probability is moderately high but not very high?

Daniel Kokotajlo: I basically agree tbc! I wouldn’t want to trust a safety case based on CoT monitoring unless it included a solid argument that the kinds of bad thoughts we are worried about can’t be had without triggering our monitors, e.g. because we have done lots of dangerous capability evals with heavy elicitation fine tuning etc. Nevertheless, I think we can learn a lot from studying CoTs and is hope that corporations invest in ways to make them more faithful and legible and deinvest in stuff that will make them less so.

I also worry, among other problems, that it will be impossible to get a superintelligent AI to not realize it should act as if its CoT is being monitored, even if somehow ‘CoTs get monitored’ is not all over the training data and we otherwise act maximally responsibly here, which we won’t. Also by default the CoT would move towards formats humans cannot parse anyway, as the authors note, and all the various pressures by default make this worse. And many other issues.

But we can and should absolutely try, and be willing to take a substantial performance hit to try.

That starts with avoiding ‘process supervision’ of the CoT that is not directed towards its legibility (and even then probably don’t do it, careful, Icarus), and various forms of indirect optimization pressure including when users are able to partially see the CoT but also almost any use of the CoT risks this. And it also means avoiding novel architectures that would lack this property. And tracking monitorability the way other safety features are tracked.

It also means investing into studying CoT monitorability. I am very happy that OpenAI is (at least claiming to be) prominently doing this.

Elon Musk: At times, AI existential dread is overwhelming.

Eliezer Yudkowsky: Well, yes. It’s going to kill you.

So, back to work making the existential dread, then?

The obvious rejoinder is ‘I will make it first and do so responsibly’ which is always highly questionable but after recent events at xAI it is laughable.

Gary Marcus: when did “it might kill us but I need to build it faster” become fashionable?

Roon: pick a lane man.

You’re allowed multiple lanes but I do hope he pivots to this one.

As many responses suggest, Elon Musk is one of the people in the world most equipped to do something about this. Elon Musk and xAI each have billions, much of which could be invested in various forms of technical work. He could advocate for better AI-related policy instead of getting into other fights.

Instead, well, have you met Grok? And Ani the x-rated anime waifu?

The Em Dash responds.

When it happens enough that you need to check to see if joking, perhaps it’s happening quite a lot, if usually not with 4o-mini?

Herakeitos137: guy I went to college with recently rented a restaurant private room and invited everyone to a dinner presentation. handouts. paper flipboard. open bar. spent 3 hours explaining how he solved the black hole information paradox after 2 months talking with ChatGPT 4o Mini.

Forgot to mention he made everyone sign ndas.

There were a couple slac guys at the dinner and they said the math checked out (although they were also the most inebriated)

Discussion about this post

AI #126: Go Fund Yourself Read More »

yet-another-bad-three-months-as-tesla-reports-its-q2-2025-results

Yet another bad three months as Tesla reports its Q2 2025 results

Tesla posted its financial results for the second quarter of 2025 this afternoon. The numbers show yet another bad three months for the automaker. As competition in the EV marketplace has exploded, Tesla has increasingly been left behind, with a small and aging model lineup, before we even contemplate how CEO Elon Musk has tarnished what was once the hottest brand in the car world. Earlier this month, we learned that sales dropped by 13 percent year over year in Q2 2025; today, the financials show that automotive revenues fell even more, dropping 16 percent year over year to $16.7 billion.

Tesla’s battery business has been feeling the pain, too. For a while, this was a growth area for the company, albeit one with a relatively minor contribution to the bottom line. During Q2 2025, Tesla’s energy generation and storage division brought in $2.8 billion in revenue, a 7 percent decline from the same period in 2024.

Sales of Carbon credits—those government-issued permits that other automakers buy in order to pollute—shrank by more than half, to $490 million. Those other automakers are now selling EVs, at least most of them, and have less need to buy credits from Tesla. It’s likely this subsidy, which has kept the company out of the red in the past, will be even less of a contributor in the coming years as the US strips away environmental protections.

Yet another bad three months as Tesla reports its Q2 2025 results Read More »

what-to-know-about-toolshell,-the-sharepoint-threat-under-mass-exploitation

What to know about ToolShell, the SharePoint threat under mass exploitation

Microsoft fixed the vulnerability pair—CVE-2025-49706 and CVE-2025-49704—two weeks ago as part of the company’s monthly update release. As the world learned over the weekend, the patches were incomplete, a lapse that opened organizations around the world to the new attacks.

Q: What sorts of malicious things are attackers doing with these newer ToolShell exploits?

A: According to numerous technical analyses, the attackers first infect vulnerable systems with a webshell-based backdoor that gains access to some of the most sensitive parts of a SharePoint Server. From there, the webshell extracts tokens and other credentials that allow the attackers to gain administrative privileges, even when systems are protected by multifactor authentication and single sign-on. Once inside, the attackers exfiltrate sensitive data and deploy additional backdoors that provide persistent access for future use.

For those who want more technical details, the opening volley in the attack is POST Web requests the attackers send to the ToolPane endpoint. The requests look like this:

Credit: Akamai

Microsoft said these requests upload a malicious script named spinstall0.aspx, or alternatively spinstall.aspx, spinstall1.aspx, spinstall2.aspx, and so on. The script contains commands for retrieving a SharePoint server’s encrypted MachineKey configuration and returning the decrypted results to the attacker through a GET request.

Q: I maintain an on-premises SharePoint server. What should I do?

A: In short, drop whatever else you were doing and take time to carefully inspect your system. The first thing to look for is whether it has received the emergency patches Microsoft released Saturday. Install the patch immediately if it hasn’t already been done.

Patching the vulnerability is only the first step, since systems infected through the vulnerability show few or no signs of compromise. The next step is to pore through system event logs in search of indicators of compromise. These indicators can be found in numerous write-ups, including those from Microsoft and Eye Security (at the links above), the US Cybersecurity and Information Security Agency, and security firms Sentinel One, Akamai, Tenable, and Palo Alto Networks.

What to know about ToolShell, the SharePoint threat under mass exploitation Read More »