Author name: Shannon Garcia

bizarre-egg-laying-mammals-once-ruled-australia—then-lost-their-teeth

Bizarre egg-laying mammals once ruled Australia—then lost their teeth

Eggs came first, no chickens involved —

Finds may indicate what the common ancestor of the platypus and echidna looked like.

A small animal with spiky fur and a long snout strides over grey soil.

Enlarge / The echidna, an egg-laying mammal, doesn’t develop teeth.

Outliers among mammals, monotremes lay eggs instead of giving birth to live young. Only two types of monotremes, the platypus and echidna, still exist, but more monotreme species were around about 100 million years ago. Some of them might possibly be even weirder than their descendants.

Monotreme fossils found in refuse from the opal mines of Lightning Ridge, Australia, have now revealed the opalized jawbones of three previously unknown species that lived during the Cenomanian age of the early Cretaceous. Unlike modern monotremes, these species had teeth. They also include a creature that appears to have been a mashup of a platypus and echidna—an “echidnapus.”

Fossil fragments of three known species from the same era were also found, meaning that at least six monotreme species coexisted in what is now Lightning Ridge. According to the researchers who unearthed these new species, the creatures may have once been as common in Australia as marsupials are today.

“[This is] the most diverse monotreme assemblage on record,” they said in a study recently published in Alcheringa: An Australasian Journal of Paleontology.

The Echidnapus emerges

Named Opalios spendens, the “echidnapus” shows similarities to both ornithorhynchoids (the platypus and similar species) and tachyglossids (echidna and similar species). It is thought to have evolved before the common ancestor of either extant monotreme.

The O. splendens holotype had been fossilized in opal like the other Lightning Ridge specimens, but unlike some, it is preserved so well that the internal structure of its bones is visible. Every mammalian fossil from Lightning Ridge has been identified as a monotreme based partly on their peculiarly large dental canals. While the fossil evidence suggests the jaw and snout of O. splendens are narrow and curved, similar to those of an echidna, it simultaneously displays platypus features.

So what relates the echidnapus to a platypus? Despite its jaw being echidna-like at first glance, its dentary, or the part of the jaw that bears the teeth, is similar in size to that of the platypus ancestor Ornithorhynchus anatinus. Other features related more closely to the platypus than the echidna have to do with its ramus, or the part of the jaw that attaches to the skull. It has a short ascending ramus (the rear end) and twisted horizontal ramus (the front end) that are seen in other ornithorhynchoids.

Another platypus-like feature of O. splendens is the flatness of the front of its lower jaw, which is consistent with the flatness of the platypus snout. The size of its jaw also suggests a body size closer to that of a platypus. Though the echidnapus had characteristics of both surviving monotremes, neither of those have the teeth found on this fossil.

My, what teeth you don’t have

Cretaceous monotremes might not have had as many teeth as the echidnapus, but they all had some teeth. The other two new monotreme species that lived among the Lightning Ridge fauna were Dharragarra aurora and Parvopalus clytiei, and the jaw structure of each of these species is either closer to the platypus or the echidna. D. aurora has the slightly twisted jaw and enlarged canal in its mandible that are characteristic of an ornithorhynchoid. It might even be on the branch that gave rise the platypus.

P. clytiei is the second smallest known monotreme (after another extinct species named Teinolophos trusleri). It was more of an echidna type, with a snout that was curved and deep like that of a tachyglossid rather than flat like that of an ornithorhynchoid. It also had teeth, though fewer than the echidnapus. Why did those teeth end up disappearing altogether in modern monotremes?

Monotremes without teeth came onto the scene when the platypus (Ornithorhynchus anatinus) appeared during the Pleistocene, which began 2.6 million years ago. The researchers think competition for food caused the disappearance of teeth in the platypus—the spread of the Australo-New Guinean water rat may have affected which prey platypuses hunted for. Water rats eat mostly fish and shellfish along with some insects, which are also thought to have been part of the diet of ancient ornithorhynchoids. Turning to softer food to avoid competition may explain why the platypus evolved to be toothless.

As for echidnas, tachyglossids are thought to have lost their teeth after they diverged from ornithorhynchoids near the end of the Cretaceous. Echidnas are insectivores, grinding the hard shells of beetles and ants with spines inside their mouths, so have no need for teeth.

Although there is some idea of what happened to their teeth, the fate of the diverse species of Cretaceous monotremes, which were not only toothy but mostly larger than the modern platypus and echidna, remains unknown. The end of the Cretaceous brought a mass extinction triggered by the Chicxulub asteroid. Clearly, some monotremes survived it, but no monotreme fossils from the time have surfaced yet.

“It is unclear whether diverse monotreme fauna survived the end-Cretaceous mass extinction event, and subsequently persisted,” the researchers said in the same study. “Filling this mysterious interval of monotreme diversity and adaptive development should be a primary focus for research in the future.”

Alcheringa: An Australasian Journal of Palaeontology, 2024. DOI: 10.1080/03115518.2024.2348753

Bizarre egg-laying mammals once ruled Australia—then lost their teeth Read More »

report:-new-“apple-intelligence”-ai-features-will-be-opt-in-by-default

Report: New “Apple Intelligence” AI features will be opt-in by default

“apple intelligence,” i see what you did there —

Apple reportedly plans to announce its first big wave of AI features at WWDC.

Report: New “Apple Intelligence” AI features will be opt-in by default

Apple

Apple’s Worldwide Developers Conference kicks off on Monday, and per usual, the company is expected to detail most of the big new features in this year’s updates to iOS, iPadOS, macOS, and all of Apple’s other operating systems.

The general consensus is that Apple plans to use this year’s updates to integrate generative AI into its products for the first time. Bloomberg’s Mark Gurman has a few implementation details that show how Apple’s approach will differ somewhat from Microsoft’s or Google’s.

Gurman says that the “Apple Intelligence” features will include an OpenAI-powered chatbot, but it will otherwise focus on “features with broad appeal” rather than “whiz-bang technology like image and video generation.” These include summaries for webpages, meetings, and missed notifications; a revamped version of Siri that can control apps in a more granular way; Voice Memos transcription; image enhancement features in the Photos app; suggested replies to text messages; automated sorting of emails; and the ability to “create custom emoji characters on the fly that represent phrases or words as they’re being typed.”

Apple also reportedly hopes to differentiate its AI push by implementing it in a more careful, privacy-focused way. The new features will use the Neural Engine available in newer devices for on-device processing where possible (Gurman says that only Apple’s A17 Pro and the M-series chips will be capable of supporting all the local processing features, though all of Apple’s recent chips feature some flavor of Neural Engine). And where Apple does use the cloud for AI processing, the company will apparently promise that user information isn’t being “sold or read” and is not being used to “build user profiles.”

Apple’s new AI features will also be opt-in by default, where Microsoft and Google have generally enabled features like the Copilot chatbot or AI Overviews by default whether users asked for them or not.

Looking beyond AI, we can also expect the typical grab bag of small- to medium-sized features in all of Apple’s software updates. These reportedly include reworked Control Center and Settings apps, emoji responses and RCS messaging support in the Messages app, a standalone password manager app, Calculator for the iPad, and a handful of other things. Gurman doesn’t expect Apple to announce any hardware at the event, though a number of Macs are past due for a M3- or M4-powered refresh.

Apple’s WWDC keynote happens on June 10 at 1 pm Eastern and can be streamed from Apple’s developer website.

Report: New “Apple Intelligence” AI features will be opt-in by default Read More »

microsoft-is-reworking-recall-after-researchers-point-out-its-security-problems

Microsoft is reworking Recall after researchers point out its security problems

recalling recall? —

Windows Hello authentication, additional encryption being added to protect data.

Microsoft's Recall feature is switching to be opt-in by default, and is adding new encryption protections in an effort to safeguard user data.

Enlarge / Microsoft’s Recall feature is switching to be opt-in by default, and is adding new encryption protections in an effort to safeguard user data.

Microsoft

Microsoft’s upcoming Recall feature in Windows 11 has generated a wave of controversy this week following early testing that revealed huge security holes. The initial version of Recall saves screenshots and a large plaintext database tracking everything that users do on their PCs, and in the current version of the feature, it’s trivially easy to steal and view that database and all of those screenshots for any user on a given PC, even if you don’t have administrator access. Recall also does little to nothing to redact sensitive information from its screenshots or that database.

Microsoft has announced that it’s making some substantial changes to Recall ahead of its release on the first wave of Copilot+ PCs later this month.

“Even before making Recall available to customers, we have heard a clear signal that we can make it easier for people to choose to enable Recall on their Copilot+ PC and improve privacy and security safeguards,” wrote Microsoft Windows and Devices Corporate Vice President Pavan Davuluri in a blog post. “With that in mind we are announcing updates that will go into effect before Recall (preview) ships to customers on June 18.”

First and most significantly, the company says that Recall will be opt-in by default, so users will need to decide to turn it on. It may seem like a small change, but many users never touch the defaults on their PCs, and for Recall to be grabbing all of that data by default definitely puts more users at risk of having their data stolen unawares.

The company also says it’s adding additional protections to Recall to make the data harder to access. You’ll need to enable Windows Hello to use Recall, and you’ll need to authenticate via Windows Hello (whether it’s a face-scanning camera, fingerprint sensor, or PIN) each time you want to open the Recall app to view your data.

Both the screenshots and the SQLite database used for Recall searches are being encrypted and will require Windows Hello authentication to be decrypted. Microsoft described Recall data as “encrypted” before, but there was no specific encryption used for any of the screenshots or the database beyond the Bitlocker full-disk encryption that is turned on by default for most PCs when they sign into a Microsoft account.

That last change should address the biggest problem with Recall: that any user signed in to a PC (or any malware that was able to gain access to the filesystem) could easily view and copy another user’s Recall screenshots and database on the same PC. The text database’s size is measured in kilobytes rather than megabytes or gigabytes, so it wouldn’t take much time to swipe if someone managed to access your system.

Microsoft also reiterated some of its assurances about the privacy and security of Recall writ large, saying that all data is processed locally, that it’s never sent to Microsoft, that you’ll know when Recall has been enabled thanks to taskbar and system tray icons, and that you can disable the feature or exclude specific apps or sites from being snapshotted at your discretion.

All of the new additions to Recall are still being actively developed—current testing builds of Windows 11 still use the unsecured version of Recall, and our review units of the new Surface hardware are being delayed by a week or so, presumably so Microsoft can update them.

Microsoft reiterated that Recall is being released as a preview, a label the company also applies to the Copilot chatbot to deflect criticism of some of its early and ongoing missteps. We’ll need to use the updated version of Recall to see whether the new protections work like they’re supposed to, but it’s at least mildly encouraging to see Microsoft take a beat to rework Recall’s security and default settings before releasing it to the public, even though these are protections should have been present in the first place.

Recall is normally only available on Copilot+ PCs, a new branding banner from Microsoft that applies to PCs with sufficiently fast neural processing units (NPUs), at least 16GB of RAM, and at least 256GB of storage. Existing Windows 11 PCs won’t get Recall, though it can currently be enabled forcibly by the third-party AmperageKit script on Arm PCs that are running version 26100.712 of Windows 11 24H2. It’s possible that tools will exist to enable it on other unsupported PCs later on.

The first wave of Copilot+ PCs will use Qualcomm’s Snapdragon X Elite and X Plus processors exclusively. Intel and AMD systems that meet the Copilot+ requirements won’t be available until later this year, and Microsoft hasn’t said when the Copilot+ features will actually be available for these non-Arm PCs.

Microsoft is reworking Recall after researchers point out its security problems Read More »

vmware-customers-may-stay,-but-broadcom-could-face-backlash-“for-years-to-come”

VMware customers may stay, but Broadcom could face backlash “for years to come”

“The emotional shock has started to metabolize” —

300 director-level IT workers making VMware decisions were questioned.

VMware customers may stay, but Broadcom could face backlash “for years to come”

After acquiring VMware, Broadcom swiftly enacted widespread changes that resulted in strong public backlash. A new survey of 300 director-level IT workers at companies that are customers of North American VMware provides insight into the customer reaction to Broadcom’s overhaul.

The survey released Thursday doesn’t provide feedback from every VMware customer, but it’s the first time we’ve seen responses from IT decision-makers working for companies paying for VMware products. It echos concerns expressed at the announcement of some of Broadcom’s more controversial changes to VMware, like the end of perpetual licenses and growing costs.

CloudBolt Software commissioned Wakefield Research, a market research agency, to run the study from May 9 through May 23. The “CloudBolt Industry Insights Reality Report: VMware Acquisition Aftermath” includes responses from workers at 150 companies with fewer than 1,000 workers and 150 companies with more than 1,000 workers. Survey respondents were invited via email and took the survey online, with the report authors writing that results are subject to sampling variation of ±5.7 percentage points at a 95 percent confidence level.

Notably, Amazon Web Services (AWS) commissioned the report in partnership with CloudBolt. AWS’s partnership with VMware hit a road bump last month when Broadcom stopped allowing AWS to resell the VMware Cloud on AWS offering—a move that AWS said “disappointed it.” Kyle Campos, CloudBolt CTPO, told Ars Technica that the full extent to which AWS was involved in this report was helping underwrite the cost of research. But you can see why AWS would have interest in customer dissatisfaction with VMware.

Widespread worry

Every person surveyed said that they expect VMware prices to rise under Broadcom. In a March “User Group Town Hall,” attendees complained about “price rises of 500 and 600 percent,” according to The Register. We heard in February from ServeTheHome that “smaller” cloud service providers were claiming to see costs grow tenfold. In this week’s survey, 73 percent of respondents said they expect VMware prices to more than double. Twelve percent of respondents expect a price hike of 301 to 500 percent. Only 1 percent anticipate price hikes of 501 to 1,000 percent.

“At this juncture post-acquisition, most larger enterprises seem to have a clear understanding of how their next procurement cycle with Broadcom will be impacted from a pricing and packaging standpoint,” the report noted.

Further, 95 percent of survey respondents said they view Broadcom buying VMware as disruptive to their IT strategy, with 46 percent considering it extremely or very disruptive.

Widespread concerns about cost and IT strategy help explain why 99 percent of the 300 respondents said they are concerned about Broadcom owning VMware, with 46 percent being “very concerned” and 30 percent “extremely concerned.”

Broadcom didn’t respond to Ars’ request for comment.

Not jumping ship yet

Despite widespread anxiety over Broadcom’s VMware, most of the respondents said they will likely stay with VMware either partially (43 percent of respondents) or fully (40 percent). A smaller percentage of respondents said they would move more workloads to the public cloud (38 percent) or a different hypervisor (34 percent) or move entirely to the public cloud (33 percent). This is with 69 percent of respondents having at least one contract expiring with VMware within the next 12 months.

Many companies have already migrated easy-to-move workloads to the public cloud, CloudBolt’s Campos said in a statement. For many firms surveyed, what’s left in the data center “is a mixture of workloads requiring significant modernization or compliance bound to the data center,” including infrastructure components that have been in place for decades. Campos noted that many mission-critical workloads remain in the data center, and moving them is “daunting with unclear ROI.”

“The emotional shock has started to metabolize inside of the Broadcom customer base, but it’s metabolized in the form of strong commitment to mitigating the negative impacts of the Broadcom VMware acquisition,” Campos told Ars Technica.

Resistance to ditching VMware reflects how “embedded” VMware is within customer infrastructures, the CloudBolt exec told Ars, adding:

In many cases, the teams responsible for purchasing, implementing, and operating VMware have never even considered an alternative prior to this acquisition; it’s the only operating reality they know and they are used to buying out of this problem.

Top reasons cited for considering abandoning VMware partially or totally were uncertainty about Broadcom’s plans, concerns about support quality under Broadcom, and changes to relationships with channel partners (each named by 36 percent of respondents).

Following closely was the shift to subscription licensing (34 percent), expected price bumps (33 percent), and personal negative experiences with Broadcom (33 percent). Broadcom’s history with big buys like Symantec and CA Technologies also has 32 percent of people surveyed considering leaving VMware.

Although many firms seem to be weighing their options before potentially leaving VMware, Campos warned that Broadcom could see backlash continue “for months and even years to come,” considering the areas of concern cited in the survey and how all VMware offerings are near-equal candidates for eventual nixing.

VMware customers may stay, but Broadcom could face backlash “for years to come” Read More »

outcry-from-big-ai-firms-over-california-ai-“kill-switch”-bill

Outcry from big AI firms over California AI “kill switch” bill

A finger poised over an electrical switch.

Artificial intelligence heavyweights in California are protesting against a state bill that would force technology companies to adhere to a strict safety framework including creating a “kill switch” to turn off their powerful AI models, in a growing battle over regulatory control of the cutting-edge technology.

The California Legislature is considering proposals that would introduce new restrictions on tech companies operating in the state, including the three largest AI start-ups OpenAI, Anthropic, and Cohere as well as large language models run by Big Tech companies such as Meta.

The bill, passed by the state’s Senate last month and set for a vote from its general assembly in August, requires AI groups in California to guarantee to a newly created state body that they will not develop models with “a hazardous capability,” such as creating biological or nuclear weapons or aiding cyber security attacks.

Developers would be required to report on their safety testing and introduce a so-called kill switch to shut down their models, according to the proposed Safe and Secure Innovation for Frontier Artificial Intelligence Systems Act.

But the law has become the focus of a backlash from many in Silicon Valley because of claims it will force AI start-ups to leave the state and prevent platforms such as Meta from operating open source models.

“If someone wanted to come up with regulations to stifle innovation, one could hardly do better,” said Andrew Ng, a renowned computer scientist who led AI projects at Alphabet’s Google and China’s Baidu, and who sits on Amazon’s board. “It creates massive liabilities for science-fiction risks, and so stokes fear in anyone daring to innovate.”

Outcry from big AI firms over California AI “kill switch” bill Read More »

quotes-from-leopold-aschenbrenner’s-situational-awareness-paper

Quotes from Leopold Aschenbrenner’s Situational Awareness Paper

This post is different.

Usually I offer commentary and analysis. I share what others think, then respond.

This is the second time I am importantly not doing that. The work speaks for itself. It offers a different perspective, a window and a worldview. It is self-consistent. This is what a highly intelligent, highly knowledgeable person actually believes after much thought.

So rather than say where I agree and disagree and argue back (and I do both strongly in many places), this is only quotes and graphs from the paper, selected to tell the central story while cutting length by ~80%, so others can more easily absorb it. I recommend asking what are the load bearing assumptions and claims, and what changes to them would alter the key conclusions.

The first time I used this format was years ago, when I offered Quotes from Moral Mazes. I think it is time to use it again.

Then there will be one or more other posts, where I do respond.

(1) Page 1: The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.

Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the willful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.

Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them.

(2) Page 7: AGI by 2027 is strikingly plausible. GPT-2 to GPT-4 took us from ~preschooler to ~smart high-schooler abilities in 4 years. Tracing trendlines in compute (~0.5 orders of magnitude or OOMs/year), algorithmic efficiencies (~0.5 OOMs/year), and “unhobbling” gains (from chatbot to agent), we should expect another preschooler-to-high-schooler-sized qualitative jump by 2027.

(3) Page 8: I make the following claim: it is strikingly plausible that by 2027, models will be able to do the work of an AI researcher/engineer. That doesn’t require believing in sci-fi; it just requires believing in straight lines on a graph.

(4) Page 9: We are racing through the OOMs extremely rapidly, and the numbers indicate we should expect another ~100,000x effective compute scaleup—resulting in another GPT-2-to-GPT-4-sized qualitative jump—over four years.

(5) Page 14: Of course, even GPT-4 is still somewhat uneven; for some tasks it’s much better than smart high-schoolers, while there are other tasks it can’t yet do. That said, I tend to think most of these limitations come down to obvious ways models are still hobbled, as I’ll discuss in-depth later. The raw intelligence is (mostly) there, even if the models are still artificially constrained; it’ll take extra work to unlock models being able to fully apply that raw intelligence across applications.

(6) Page 19: How did this happen? The magic of deep learning is that it just works—and the trendlines have been astonishingly consistent, despite naysayers at every turn.

(7) Page 21: An additional 2 OOMs of compute (a cluster in the $10s of billions) seems very likely to happen by the end of 2027; even a cluster closer to +3 OOMs of compute ($100 billion+) seems plausible (and is rumored to be in the works at Microsoft/OpenAI).

(8) Page 23: In this piece, I’ll separate out two kinds of algorithmic progress. Here, I’ll start by covering “within-paradigm” algorithmic improvements—those that simply result in better base models, and that straightforwardly act as compute efficiencies or compute multipliers. For example, a better algorithm might allow us to achieve the same performance but with 10x less training compute. In turn, that would act as a 10x (1 OOM) increase in effective compute. (Later, I’ll cover “unhobbling,” which you can think of as “paradigm-expanding/application-expanding” algorithmic progress that unlocks capabilities of base models.)

(9) Page 26: Put together, this suggests we should expect something like 1-3 OOMs of algorithmic efficiency gains (compared to GPT-4) by the end of 2027, maybe with a best guess of ~2 OOMs.

(10) Page 27: In addition to insider bullishness, I think there’s a strong intuitive case for why it should be possible to find ways to train models with much better sample efficiency (algorithmic improvements that let them learn more from limited data). Consider how you or I would learn from a really dense math textbook.

(11) Page 29: All of this is to say that data constraints seem to inject large error bars either way into forecasting the coming years of AI progress. There’s a very real chance things stall out (LLMs might still be as big of a deal as the internet, but we wouldn’t get to truly crazy AGI). But I think it’s reasonable to guess that the labs will crack it, and that doing so will not just keep the scaling curves going, but possibly enable huge gains in model capability.

(12) Page 29: As an aside, this also means that we should expect more variance between the different labs in coming years compared to today. Up until recently, the state of the art techniques were published, so everyone was basically doing the same thing. (And new upstarts or open source projects could easily compete with the frontier, since the recipe was published.) Now, key algorithmic ideas are becoming increasingly proprietary.

(13) Page 33: “Unhobbling” is a huge part of what actually enabled these models to become useful—and I’d argue that much of what is holding back many commercial applications today is the need for further “unhobbling” of this sort. Indeed, models today are still incredibly hobbled! For example:

• They don’t have long-term memory

• They can’t use a computer (they still only have very limited tools).

• They still mostly don’t think before they speak. When you ask ChatGPT to write an essay, that’s like expecting a human to write an essay via their initial stream-of-consciousness (People are working on this though).

• They can (mostly) only engage in short back-and-forth dialogues, rather than going away for a day or a week, thinking about a problem, researching different approaches, consulting other humans, and then writing you a longer report or pull request.

• They’re mostly not personalized to you or your application (just a generic chatbot with a short prompt, rather than having all the relevant background on your company and your work).

It seems like it should be possible, for example via very-long-context, to “onboard” models like we would a new human coworker. This alone would be a huge unlock.

(14) Page 35: In essence, there is a large test-time compute overhang.

(15) Page 38: By the end of this, I expect us to get something that looks a lot like a drop-in remote worker.

(16) Page 41: (One neat way to think about this is that the current trend of AI progress is proceeding at roughly 3x the pace of child development. Your 3x-speed-child just graduated high school; it’ll be taking your job before you know it!)

We are on course for AGI by 2027. These AI systems will basically be able to automate basically all cognitive jobs (think: all jobs that could be done remotely).

To be clear—the error bars are large. Progress could stall as we run out of data, if the algorithmic breakthroughs necessary to crash through the data wall prove harder than expected. Maybe unhobbling doesn’t go as far, and we are stuck with merely expert chatbots, rather than expert coworkers. Perhaps the decade-long trendlines break, or scaling deep learning hits a wall for real this time. (Or an algorithmic breakthrough, even simple unhobbling that unleashes the test-time compute overhang, could be a paradigm-shift, accelerating things further and leading to AGI even earlier.)

(17) Page 42: It seems like many are in the game of downward-defining AGI these days, as just as really good chatbot or whatever. What I mean is an AI system that could fully automate my or my friends’ job, that could fully do the work of an AI researcher or engineer.

(18) Page 43: Addendum. Racing through the OOMs: It’s this decade or bust I used to be more skeptical of short timelines to AGI. One reason is that it seemed unreasonable to privilege this decade, concentrating so much AGI-probability-mass on it (it seemed like a classic fallacy to think “oh we’re so special”). I thought we should be uncertain about what it takes to get AGI, which should lead to a much more “smeared-out” probability distribution over when we might get AGI. However, I’ve changed my mind: critically, our uncertainty over what it takes to get AGI should be over OOMs (of effective compute), rather than over years. We’re racing through the OOMs this decade. Even at its bygone heyday, Moore’s law was only 1–1.5 OOMs/decade. I estimate that we will do ~5 OOMs in 4 years, and over ~10 this decade overall.

In essence, we’re in the middle of a huge scaleup reaping one-time gains this decade, and progress through the OOMs will be multiples slower thereafter. If this scaleup doesn’t get us to AGI in the next 5-10 years, it might be a long way out.

(19) Page 44: Hardware gains: AI hardware has been improving much more quickly than Moore’s law. That’s because we’ve been specializing chips for AI workloads. For example, we’ve gone from CPUs to GPUs; adapted chips for Transformers; and we’ve gone down to much lower precision number formats, from fp64/fp32 for traditional supercomputing to fp8 on H100s. These are large gains, but by the end of the decade we’ll likely have totally specialized AI-specific chips, without much further beyond-Moore’s law gains possible.

Algorithmic progress: In the coming decade, AI labs will invest tens of billions in algorithmic R&D, and all the smartest people in the world will be working on this; from tiny efficiencies to new paradigms, we’ll be picking lots of the low-hanging fruit. We probably won’t reach any sort of hard limit (though “unhobblings” are likely finite), but at the very least the pace of improvements should slow down, as the rapid growth (in $ and human capital investments) necessarily slows down (e.g., most of the smart STEM talent will already be working on AI). (That said, this is the most uncertain to predict, and the source of most of the uncertainty on the OOMs in the 2030s on the plot above.)

(20) Page 46 (start of section 2): AI progress won’t stop at human-level. Hundreds of millions of AGIs could automate AI research, compressing a decade of algorithmic progress (5+ OOMs) into 1 year. We would rapidly go from human-level to vastly superhuman AI systems. The power—and the peril—of superintelligence would be dramatic.

(21) Page 48: Once we get AGI, we won’t just have one AGI. I’ll walk through the numbers later, but: given inference GPU fleets by then, we’ll likely be able to run many millions of them (perhaps 100 million human-equivalents, and soon after at 10x+ human speed).

(22) Page 49: We don’t need to automate everything—just AI research.

(23) Page 50: It’s worth emphasizing just how straightforward and hacky some of the biggest machine learning breakthroughs of the last decade have been: “oh, just add some normalization” (LayerNorm/BatchNorm) or “do f(x)+x instead of f(x)” (residual connections)” or “fix an implementation bug” (Kaplan → Chinchilla scaling laws). AI research can be automated. And automating AI research is all it takes to kick off extraordinary feedback loops.

(24) Page 50: Another way of thinking about it is that given inference fleets in 2027, we should be able to generate an entire internet’s worth of tokens, every single day.

(25) Page 51: By taking some inference penalties, we can trade off running fewer copies in exchange for running them at faster serial speed. (For example, we could go from ~5x human speed to ~100x human speed by “only” running 1 million copies of the automated researchers.

More importantly, the first algorithmic innovation the automated AI researchers work on is getting a 10x or 100x speedup.

This could easily dramatically accelerate existing trends of algorithmic progress, compressing a decade of advances into a year.

(26) Page 51: Don’t just imagine 100 million junior software engineer interns here (we’ll get those earlier, in the next couple years!). Real automated AI researchers be very smart—and in addition to their raw quantitative advantage, automated AI researchers will have other enormous advantages over human researchers.

They’ll be able to read every single ML paper ever written, have been able to deeply think about every single previous experiment ever run at the lab, learn in parallel from each of their copies, and rapidly accumulate the equivalent of millennia of experience.

(27) Page 53: It’s strikingly plausible we’d go from AGI to superintelligence very quickly, perhaps in 1 year.

(28) Page 54: the last 10% of the job of an AI researcher might be particularly hard to automate. This could soften takeoff some, though my best guess is that this only delays things by a couple years.

Maybe another 5 OOMs of algorithmic efficiency will be fundamentally impossible? I doubt it.

(29) Page 59: I’ll take a moment here to acknowledge perhaps the most compelling formulation of the counterargument I’ve heard, by my friend James Bradbury: if more ML research effort would so dramatically accelerate progress, why doesn’t the current academic ML research community, numbering at least in the tens of thousands, contribute more to frontier lab progress?

(Currently, it seems like lab-internal teams, of perhaps a thousand in total across labs, shoulder most of the load for frontier algorithmic progress.) His argument is that the reason is that algorithmic progress is compute-bottlenecked: the academics just don’t have enough compute.

Some responses: Quality-adjusted, I think academics are probably more in the thousands not tens of thousands (e.g., looking only at the top universities).

Academics work on the wrong things.

Even when the academics do work on things like LLM pretraining, they simply don’t have access to the state-of-the-art.

Academics are way worse than automated AI researchers: they can’t work at 10x or 100x human speed, they can’t read and internalize every ML paper ever written, they can’t spend a decade checking every line of code, replicate themselves to avoid onboarding-bottlenecks, etc.

(30) Page 61: I think it’s reasonable to be uncertain how this plays out, but it’s unreasonable to be confident it won’t be doable for the models to get around the compute bottleneck just because it’d be hard for humans to do so.

(31) Page 62: Still, in practice, I do expect somewhat of a long tail to get to truly 100% automation even for the job of an AI researcher/engineer; for example, we might first get systems that function almost as an engineer replacement, but still need some amount of human supervision.

In particular, I expect the level of AI capabilities to be somewhat uneven and peaky across domains: it might be a better coder than the best engineers while still having blindspots in some subset of tasks or skills; by the time it’s human-level at whatever its worst at, it’ll already be substantially superhuman at easier domains to train, like coding.

(32) Page 62: But I wouldn’t expect that phase to last more than a few years; given the pace of AI progress, I think it would likely just be a matter of some additional “unhobbling” (removing some obvious limitation of the models that prevented it from doing the last mile) or another generation of models to get all the way.

(33) Page 68: Solve robotics. Superintelligence won’t stay purely cognitive for long. Getting robotics to work well is primarily an ML algorithms problem (rather than a hardware problem), and our automated AI researchers will likely be able to solve it (more below!). Factories would go from human-run, to AIdirected using human physical labor, to soon being fully run by swarms of robots.

Dramatically accelerate scientific and technological progress. Yes, Einstein alone couldn’t develop neuroscience and build a semiconductor industry, but a billion superintelligent automated scientists, engineers, technologists, and robot technicians would make extraordinary advances in many fields in the space of years.

An industrial and economic explosion. Extremely accelerated technological progress, combined with the ability to automate all human labor, could dramatically accelerate economic growth.

(34) Page 70: Provide a decisive and overwhelming military advantage.

Be able to overthrow the US government. Whoever controls superintelligence will quite possibly have enough power to seize control from pre-superintelligence forces. Even without robots, the small civilization of superintelligences would be able to hack any undefended military, election, television, etc. system, cunningly persuade generals and electorates, economically outcompete nation-states, design new synthetic bioweapons and then pay a human in bitcoin to synthesize it, and so on.

(35) Page 72: Robots. A common objection to claims like those here is that, even if AI can do cognitive tasks, robotics is lagging way behind and so will be a brake on any real-world impacts.

I used to be sympathetic to this, but I’ve become convinced robots will not be a barrier. For years people claimed robots were a hardware problem—but robot hardware is well on its way to being solved.

Increasingly, it’s clear that robots are an ML algorithms problem.

(36) Page 75 (start of part 3): The most extraordinary techno-capital acceleration has been set in motion. As AI revenue grows rapidly, many trillions of dollars will go into GPU, datacenter, and power buildout before the end of the decade. The industrial mobilization, including growing US electricity production by 10s of percent, will be intense.

(37) Page 76: Total AI investment could be north of $1T annually by 2027.

By the end of the decade, we are headed to $1T+ individual training clusters, requiring power equivalent to >20% of US electricity production. Trillions of dollars of capex will churn out 100s of millions of GPUs per year overall.

(38) Page 78: (Note that I think it’s pretty likely we’ll only need a ~$100B cluster, or less, for AGI. The $1T cluster might be what we’ll train and run superintelligence on, or what we’ll use for AGI if AGI is harder than expected. In any case, in a post-AGI world, having the most compute will probably still really matter.)

My rough estimate is that 2024 will already feature $100B- $200B of AI investment.

Big tech has been dramatically ramping their capex numbers: Microsoft and Google will likely do $50B+ , AWS and Meta $40B+, in capex this year. Not all of this is AI, but combined their capex will have grown $50B-$100B year-over-year because of the AI boom, and even then they are still cutting back on other capex to shift even more spending to AI.

(39) Page 80: Let’s play this forward. My best guess is overall compute investments will grow more slowly than the 3x/year largest training clusters, let’s say 2x/year.

(40) Page 81: Reports suggest OpenAI was at a $1B revenue run rate in August 2023, and a $2B revenue run rate in February 2024. That’s roughly a doubling every 6 months. If that trend holds, we should see a ~$10B annual run rate by late 2024/early 2025, even without pricing in a massive surge from any next generation model. One estimate puts Microsoft at ~$5B of incremental AI revenue already.

(41) Page 82: ery naively extrapolating out the doubling every 6 months, supposing we hit a $10B revenue run rate in early 2025, suggests this would happen mid-2026.

That may seem like a stretch, but it seems to me to require surprisingly little imagination to reach that milestone. For example, there are around 350 million paid subscribers to Microsoft Office—could you get a third of these to be willing to pay $100/month for an AI add-on?

For an average worker, that’s only a few hours a month of productivity gained; models powerful enough to make that justifiable seem very doable in the next couple years.

It’s hard to understate the ensuing reverberations. This would make AI products the biggest revenue driver for America’s largest corporations, and by far their biggest area of growth.

We probably see our first many-hundred-billion dollar corporate bond sale then.

Historical precedents

$1T/year of total annual AI investment by 2027 seems outrageous. But it’s worth taking a look at other historical reference classes:

• In their peak years of funding, the Manhattan and Apollo programs reached 0.4% of GDP, or ~$100 billion annually today (surprisingly small!). At $1T/year, AI investment would be about 3% of GDP.

• Between 1996–2001, telecoms invested nearly $1 trillion in today’s dollars in building out internet infrastructure. • From 1841 to 1850, private British railway investments totaled a cumulative ~40% of British GDP at the time. A similar fraction of US GDP would be equivalent to ~$11T over a decade.

• Many trillions are being spent on the green transition.

• Rapidly-growing economies often spend a high fraction of their GDP on investment; for example, China has spent more than 40% of its GDP on investment for two decades (equivalent to $11T annually given US GDP).

• In the historically most exigent national security circumstances— wartime—borrowing to finance the national effort has often comprised enormous fractions of GDP. During WWI, the UK and France, and Germany borrowed over 100% of their GDPs while the US borrowed over 20%; during WWII, the UK and Japan borrowed over 100% of their GDPs while the US borrowed over 60% of GDP (equivalent to over $17T today).

$1T/year of total AI investment by 2027 would be dramatic— among the very largest capital buildouts ever—but would not be unprecedented. And a trillion-dollar individual training cluster by the end of the decade seems on the table.

(42) Page 83: Probably the single biggest constraint on the supply-side will be power. Already, at nearer-term scales (1GW/2026 and especially 10GW/2028), power has become the binding constraint: there simply isn’t much spare capacity, and power contracts are usually long-term locked-in. And building, say, a new gigawatt-class nuclear power plant takes a decade.

(43) Page 84: To most, this seems completely out of the question. Some are betting on Middle Eastern autocracies, who have been going around offering boundless power and giant clusters to get their rulers a seat at the AGI-table.

But it’s totally possible to do this in the United States: we have abundant natural gas.

(44) Page 85: We’re going to drive the AGI datacenters to the Middle East, under the thumb of brutal, capricious autocrats. I’d prefer clean energy too—but this is simply too important for US national security. We will need a new level of determination to make this happen. The power constraint can, must, and will be solved.

(45) Page 86: While chips are usually what comes to mind when people think about AI-supply-constraints, they’re likely a smaller constraint than power. Global production of AI chips is still a pretty small percent of TSMC-leading-edge production, likely less than 10%. There’s a lot of room to grow via AI becoming a larger share of TSMC production.

(46) Page 86: Even if raw logic fabs won’t be the constraint, chip-on-waferon-substrate (CoWoS) advanced packaging (connecting chips to memory, also made by TSMC, Intel, and others) and HBM memory (for which demand is enormous) are already key bottlenecks for the current AI GPU scaleup; these are more specialized to AI, unlike the pure logic chips, so there’s less pre-existing capacity.

In the near term, these will be the primary constraint on churning out more GPUs, and these will be the huge constraints as AI scales. Still, these are comparatively “easy” to scale; it’s been incredible watching TSMC literally build “greenfield” fabs (i.e. entirely new facilities from scratch) to massively scale up CoWoS production this year (and Nvidia is even starting to find CoWoS alternatives to work around the shortage).

(47) Page 87: Before the decade is out, many trillions of dollars of compute clusters will have been built. The only question is whether they will be built in America.

While onshoring more of AI chip production to the US would be nice, it’s less critical than having the actual datacenter (on which the AGI lives) in the US. If having chip production abroad is like having uranium deposits abroad, having the AGI datacenter abroad is like having the literal nukes be built and stored abroad.

The clusters can be built in the US, and we have to get our act together to make sure it happens in the US. American national security must come first, before the allure of free-flowing Middle Eastern cash, arcane regulation, or even, yes, admirable climate commitments. We face a real system competition— can the requisite industrial mobilization only be done in “topdown” autocracies? If American business is unshackled, America can build like none other (at least in red states). Being willing to use natural gas, or at the very least a broad-based deregulatory agenda—NEPA exemptions, fixing FERC and transmission permitting at the federal level, overriding utility regulation, using federal authorities to unlock land and rights of way—is a national security priority.

(48) Page 89 (Start of IIIb): The nation’s leading AI labs treat security as an afterthought. Currently, they’re basically handing the key secrets for AGI to the CCP on a silver platter. Securing the AGI secrets and weights against the state-actor threat will be an immense effort, and we’re not on track.

(49) Page 90: On the current course, the leading Chinese AGI labs won’t be in Beijing or Shanghai—they’ll be in San Francisco and London. In a few years, it will be clear that the AGI secrets are the United States’ most important national defense secrets—deserving treatment on par with B-21 bomber or Columbia-class submarine blueprints, let alone the proverbial “nuclear secrets”—but today, we are treating them the way we would random SaaS software. At this rate, we’re basically just handing superintelligence to the CCP.

(50) Page 91: And this won’t just matter years in the future. Sure, who cares if GPT-4 weights are stolen—what really matters in terms of weight security is that we can secure the AGI weights down the line, so we have a few years, you might say. (Though if we’re building AGI in 2027, we really have to get moving!) But the AI labs are developing the algorithmic secrets—the key technical breakthroughs, the blueprints so to speak—for the AGI right now (in particular, the RL/self-play/synthetic data/etc “next paradigm” after LLMs to get past the data wall). AGI-level security for algorithmic secrets is necessary years before AGIlevel security for weights.

Our failure today will be irreversible soon: in the next 12-24 months, we will leak key AGI breakthroughs to the CCP. It will be the national security establishment’s single greatest regret before the decade is out.

(51) Page 93: The threat model

There are two key assets we must protect: model weights (especially as we get close to AGI, but which takes years of preparation and practice to get right) and algorithmic secrets (starting yesterday).

(52) Page 94: Perhaps the single scenario that most keeps me up at night is if China or another adversary is able to steal the automated-AI-researcher-model-weights on the cusp of an intelligence explosion. China could immediately use these to automate AI research themselves (even if they had previously been way behind)—and launch their own intelligence explosion. That’d be all they need to automate AI research, and build superintelligence. Any lead the US had would vanish.

Moreover, this would immediately put us in an existential race; any margin for ensuring superintelligence is safe would disappear. The CCP may well try to race through an intelligence explosion as fast as possible—even months of lead on superintelligence could mean a decisive military advantage—in the process skipping all the safety precautions any responsible US AGI effort would hope to take.

We’re miles away for sufficient security to protect weights today. Google DeepMind (perhaps the AI lab that has the best security of any of them, given Google infrastructure) at least straight-up admits this. Their Frontier Safety Framework outlines security levels 0, 1, 2, 3, and 4 (~1.5 being what you’d need to defend against well-resourced terrorist groups or cybercriminals, 3 being what you’d need to defend against the North Koreas of the world, and 4 being what you’d need to have even a shot of defending against priority efforts by the most capable state actors)

They admit to being at level 0 (only the most banal and basic measures). If we got AGI and superintelligence soon, we’d literally deliver it to terrorist groups and every crazy dictator out there!

Critically, developing the infrastructure for weight security probably takes many years of lead times—if we think AGI in ~3-4 years is a real possibility and we need state-proof weight security then, we need to be launching the crash effort now.

(53) Page 95: Algorithmic secrets

While people are starting to appreciate (though not necessarily implement) the need for weight security, arguably even more important right now—and vastly underrated—is securing algorithmic secrets.

One way to think about this is that stealing the algorithmic secrets will be worth having a 10x or more larger cluster to the PRC,

(54) Page 96: It’s easy to underrate how important an edge algorithmic secrets will be—because up until ~a couple years ago, everything was published.

(55) Page 97: Put simply, I think failing to protect algorithmic secrets is probably the most likely way in which China is able to stay competitive in the AGI race. (I discuss this more later.)

It’s hard to overstate how bad algorithmic secrets security is right now. Between the labs, there are thousands of people with access to the most important secrets; there is basically no background-checking, silo’ing, controls, basic infosec, etc. Things are stored on easily hackable SaaS services. People gabber at parties in SF. Anyone, with all the secrets in their head, could be offered $100M and recruited to a Chinese lab at any point.

(56) Page 98: There’s a lot of low-hanging fruit on security at AI labs. Merely adopting best practices from, say, secretive hedge funds or Google-customer-data-level security, would put us in a much better position with respect to “regular” economic espionage from the CCP. Indeed, there are notable examples of private sector firms doing remarkably well at preserving secrets. Take quantitative trading firms (the Jane Streets of the world) for example.

A number of people have told me that in an hour of conversation they could relay enough information to a competitor such that their firm’s alpha would go to ~zero—similar to how many key AI algorithmic secrets could be relayed in a short conversation—and yet these firms manage to keep these secrets and retain their edge.

(57) Page 99: While the government does not have a perfect track record on security themselves, they’re the only ones who have the infrastructure, know-how, and competencies to protect nationaldefense-level secrets. Basic stuff like the authority to subject employees to intense vetting; threaten imprisonment for leaking secrets; physical security for datacenters; and the vast know-how of places like the NSA and the people behind the security clearances (private companies simply don’t have the expertise on state-actor attacks).

(58) Page 100: Some argue that strict security measures and their associated friction aren’t worth it because they would slow down American AI labs too much. But I think that’s mistaken:

This is a tragedy of the commons problem. For a given lab’s commercial interests, security measures that cause a 10% slowdown might be deleterious in competition with other labs. But the national interest is clearly better served if every lab were willing to accept the additional friction.

Moreover, ramping security now will be the less painful path in terms of research productivity in the long run. Eventually, inevitably, if only on the cusp of superintelligence, in the extraordinary arms race to come, the USG will realize the situation is unbearable and demand a security crackdown.

Others argue that even if our secrets or weights leak, we will still manage to eke out just ahead by being faster in other ways (so we needn’t worry about security measures). That, too, is mistaken, or at least running way too much risk:

As I discuss in a later piece, I think the CCP may well be able to brutely outbuild the US (a 100GW cluster will be much easier for them). More generally, China might not have the same caution slowing it down that the US will (both reasonable and unreasonable caution!). Even if stealing the algorithms or weights “only” puts them on par with the US model-wise, that might be enough for them to win the race to superintelligence.

Moreover, even if the US squeaks out ahead in the end, the difference between a 1-2 year and 1-2 month lead will really matter for navigating the perils of superintelligence. A 1-2 year lead means at least a reasonable margin to get safety right, and to navigate the extremely volatile period around the intelligence explosion and post-superintelligence.

(59) Page 102: There’s a real mental dissonance on security at the leading AI labs. They full-throatedly claim to be building AGI this decade. They emphasize that American leadership on AGI will be decisive for US national security. They are reportedly planning 7T chip buildouts that only make sense if you really believe in AGI. And indeed, when you bring up security, they nod and acknowledge “of course, we’ll all be in a bunker” and smirk.

And yet the reality on security could not be more divorced from that. Whenever it comes time to make hard choices to prioritize security, startup attitudes and commercial interests prevail over the national interest. The national security advisor would have a mental breakdown if he understood the level of security at the nation’s leading AI labs.

(60) Page 105 (start of IIIc): Reliably controlling AI systems much smarter than we are is an unsolved technical problem. And while it is a solvable problem, things could very easily go off the rails during a rapid intelligence explosion. Managing this will be extremely tense; failure could easily be catastrophic.

There is a very real technical problem: our current alignment techniques (methods to ensure we can reliably control, steer, and trust AI systems) won’t scale to superhuman AI systems. What I want to do is explain what I see as the “default” plan for how we’ll muddle through, and why I’m optimistic. While not enough people are on the ball—we should have much more ambitious efforts to solve this problem!—overall, we’ve gotten lucky with how deep learning has shaken out, there’s a lot of empirical low-hanging fruit that will get us part of the way, and we’ll have the advantage of millions of automated AI researchers to get us the rest of the way.

But I also want to tell you why I’m worried. Most of all, ensuring alignment doesn’t go awry will require extreme competence in managing the intelligence explosion. If we do rapidly transition from from AGI to superintelligence, we will face a situation where, in less than a year, we will go from recognizable human-level systems for which descendants of current alignment techniques will mostly work fine, to much more alien, vastly superhuman systems that pose a qualitatively different, fundamentally novel technical alignment problem; at the same time, going from systems where failure is low-stakes to extremely powerful systems where failure could be catastrophic; all while most of the world is probably going kind of crazy. It makes me pretty nervous.

In essence, we face a problem of handing off trust. By the end of the intelligence explosion, we won’t have any hope of understanding what our billion superintelligences are doing (except as they might choose to explain to us, like they might to a child). And we don’t yet have the technical ability to reliably guarantee even basic side constraints for these systems, like “don’t lie” or “follow the law” or “don’t try to exfiltrate your server.”

Reinforcement from human feedback (RLHF) works very well for adding such side constraints for current systems—but RLHF relies on humans being able to understand and supervise AI behavior, which fundamentally won’t scale to superhuman systems.

The superalignment problem

We’ve been able to develop a very successful method for aligning (i.e., steering/controlling) current AI systems (AI systems dumber than us!): Reinforcement Learning from Human Feedback (RLHF).

The core technical problem of superalignment is simple: how do we control AI systems (much) smarter than us?

RLHF will predictably break down as AI systems get smarter, and we will face fundamentally new and qualitatively different technical challenges. Imagine, for example, a superhuman AI system generating a million lines of code in a new programming language it invented. If you asked a human rater in an RLHF procedure, “does this code contain any security backdoors?” they simply wouldn’t know. They wouldn’t be able to rate the output as good or bad, safe or unsafe, and so we wouldn’t be able to reinforce good behaviors and penalize bad behaviors with RLHF.

In the (near) future, even the best human experts spending lots of time won’t be good enough.

(61) Page 110: If we can’t add these side-constraints, it’s not clear what will happen. Maybe we’ll get lucky and things will be benign by default (for example, maybe we can get pretty far without the AI systems having long-horizon goals, or the undesirable behaviors will be minor). But it’s also totally plausible they’ll learn much more serious undesirable behaviors: they’ll learn to lie, they’ll learn to seek power, they’ll learn to behave nicely when humans are looking and pursue more nefarious strategies when we aren’t watching, and so on.

The primary problem is that for whatever you want to instill the model (including ensuring very basic things, like “follow the law”!) we don’t yet know how to do that for the very powerful AI systems we are building very soon.

(62) Page 111: It sounds crazy, but remember when everyone was saying we wouldn’t connect AI to the internet? The same will go for things like “we’ll make sure a human is always in the loop!”—as people say today.

We’ll have summoned a fairly alien intelligence, one much smarter than us, one whose architecture and training process wasn’t even designed by us but some supersmart previous generation of AI systems, one where we can’t even begin to understand what they’re doing, it’ll be running our military, and its goals will have been learned by a naturalselection-esque process.

Unless we solve alignment—unless we figure out how to instill those side-constraints—there’s no particular reason to expect this small civilization of superintelligences will continue obeying human commands in the long run. It seems totally within the realm of possibilities that at some point they’ll simply conspire to cut out the humans, whether suddenly or gradually.

(63) Page 112: What makes this incredibly hair-raising is the possibility of an intelligence explosion: that we might make the transition from roughly human-level systems to vastly superhuman systems extremely rapidly, perhaps in less than a year.

(64) Page 113: The superintelligence we get by the end of it will be vastly superhuman. We’ll be entirely reliant on trusting these systems, and trusting what they’re telling us is going on—since we’ll have no ability of our own to pierce through what exactly they’re doing anymore.

– One example that’s very salient to me: we may well bootstrap our way to human-level or somewhat-superhuman AGI with systems that reason via chains of thoughts, i.e. via English tokens. This is extraordinarily helpful, because it means the models “think out loud” letting us catch malign behavior (e.g., if it’s scheming against us). But surely having AI systems think in tokens is not the most efficient means to do it, surely there’s something much better that does all of this thinking via internal states—and so the model by the end of the intelligence explosion will almost certainly not think out loud, i.e. will have completely uninterpretable reasoning.

Think: “We caught the AI system doing some naughty things in a test, but we adjusted our procedure a little bit to hammer that out. Our automated AI researchers tell us the alignment metrics look good, but we don’t really understand what’s going on and don’t fully trust them, and we don’t have any strong scientific understanding that makes us confident this will continue to hold for another couple OOMs. So, we’ll probably be fine? Also China just stole our weights and they’re launching their own intelligence explosion, they’re right on our heels.”

It just really seems like this could go off the rails. To be honest, it sounds terrifying.

Yes, we will have AI systems to help us. Just like they’ll automate capabilities research, we can use them to automate alignment research. That will be key, as I discuss below. But—can you trust the AI systems? You weren’t sure whether they were aligned in the first place—are they actually being honest with you about their claims about alignment science?

(65) Page 115: The default plan: how we can muddle through

I think we can harvest wins across a number of empirical bets, which I’ll describe below, to align somewhat-superhuman systems. Then, if we’re confident we can trust these systems, we’ll need to use these somewhat-superhuman systems to automate alignment research—alongside the automation of AI research in general, during the intelligence explosion—to figure out how to solve alignment to go the rest of the way.

(66) Page 116: More generally, the more we can develop good science now, the more we’ll be in a position to verify that things aren’t going off the rails during the intelligence explosion. Even having good metrics we can trust for superalignment is surprisingly difficult—but without reliable metrics during the intelligence explosion, we won’t know whether pressing on is safe or not.

Here are some of the main research bets I see for crossing the gap between human-level and somewhat-superhuman systems.

evaluation is easier than generation. We get some of the way “for free,” because it’s easier for us to evaluate outputs (especially for egregious misbehaviors) than it is to generate them ourselves. For example, it takes me months or years of hard work to write a paper, but only a couple hours to tell if a paper someone has written is any good (though perhaps longer to catch fraud). We’ll have teams of expert humans spend a lot of time evaluating every RLHF example, and they’ll be able to “thumbs down” a lot of misbehavior even if the AI system is somewhat smarter than them.

Scalable oversight. We can use AI assistants to help humans supervise other AI systems—the human-AI team being able to extend supervision farther than the human could alone.

(67) Page 117: generalization. Even with scalable oversight, we won’t be able to supervise AI systems on really hard problems, problems beyond human comprehension. However, we can study: how will the AI systems generalize from human supervision on easy problems (that we do understand and can supervise) to behave on the hard problems (that we can’t understand and can no longer supervise)?

For example, perhaps supervising a model to be honest in simple cases generalizes benignly to the model just being honest in general, even in cases where it’s doing extremely complicated things we don’t understand.

There’s a lot of reasons to be optimistic here: part of the magic of deep learning is that it often generalizes in benign ways (for example, RLHF’ing with only labels on English examples also tends to produce good behavior when it’s speaking French or Spanish, even if that wasn’t part of the training). I’m fairly optimistic that there will both be pretty simple methods that help nudge the models’ generalization in our favor, and that we can develop a strong scientific understanding that helps us predict when generalization will work and when it will fail. To a greater extent that for scalable oversight, the hope is that this will help with alignment even in the “qualitatively” superhuman case.

Here’s another way of thinking about this: if a superhuman model is misbehaving, say breaking the law, intuitively the model should already know that it’s breaking the law. Moreover, “is this breaking the law” is probably a pretty natural concept to the model—and it will be salient in the model’s representation space. The question then is: can we “summon” this concept from the model with only weak supervision?

(68) Page 118: interpretability. One intuitively-attractive way we’d hope to verify and trust that our AI systems are aligned is if we could understand what they’re thinking! For example, if we’re worried that AI systems are deceiving us or conspiring against us, access to their internal reasoning should help us detect that

I’m worried fully reverse-engineering superhuman AI systems will just be an intractable problem—similar, to, say “fully reverse engineering the human brain”—and I’d put this work mostly in the “ambitious moonshot for AI safety” rather than “default plan for muddling through” bucket.

(69) Page 119: “Top-down” interpretability. If mechanistic interpretability tries to reverse engineer neural networks “from the bottom up,” other work takes a more targeted, “top-down” approach, trying to locate information in a model without full understanding of how it is processed.

For example, we might try to build an “AI lie detector” by identifying the parts of the neural net that “light up” when an AI system is lying. This can be a lot more tractable (even if it gives less strong guarantees).

I’m increasingly bullish that top-down interpretability techniques will be a powerful tool—i.e., we’ll be able to build something like an “AI lie detector” —and without requiring fundamental breakthroughs in understanding neural nets.

Chain-of-thought interpretability. As mentioned earlier, I think it’s quite plausible that we’ll bootstrap our way to AGI with systems that “think out loud” via chains of thought.

There’s a ton of work to do here, however, if we wanted to rely on this. How do we ensure that the CoT remains legible?

(70) Page 120: adversarial testing and measurements. Along the way, it’s going to be critical to stress test the alignment of our systems at every step—our goal should be to encounter every failure mode in the lab before we encounter it in the wild.

(71) Page 121: But we also don’t have to solve this problem just on our own. If we manage to align somewhat-superhuman systems enough to trust them, we’ll be in an incredible position: we’ll have millions of automated AI researchers, smarter than the best AI researchers, at our disposal. Leveraging these army of automated researchers properly to solve alignment for even-more superhuman systems will be decisive.

Superdefense

“Getting alignment right” should only be the first of many layers of defense during the intelligence explosion. Alignment will be hard; there will be failures along the way. If at all possible, we need to be in a position where alignment can fail—but failure wouldn’t be catastrophic. This could mean:

Security. An airgapped cluster is the first layer of defense against superintelligence attempting to self-exfiltrate and doing damage in the real world. And that’s only the beginning; we’ll need much more extreme security against model self-exfiltration across the board, from hardware encryption to many-key signoff.

Monitoring. If our AI systems are up to something fishy or malevolent—or a rogue employee tries to use them for unauthorized activities—we need to be able to catch it.

Targeted capability limitations. As much as possible, we should try to limit the model’s capabilities in targeted ways that reduce fallout from failure.

Targeted training method restrictions. There are likely some ways of training models that are inherently riskier—more likely to produce severe misalignments—than others. For example, imitation learning seems relatively safe (hard to see how that would lead to models that have dangerous long term internal goals), while we should avoid long-horizon outcome-based RL.

Will these be foolproof? Not at all. True superintelligence is likely able to get around most-any security scheme for example. Still, they buy us a lot more margin for error—and we’re going to need any margin we can get.

(72) Page 125: I think there’s a pretty reasonable shot that “the default plan” to align “somewhat-superhuman” systems will mostly work. Of course, it’s one thing to speak about a “default plan” in the abstract—it’s another if the team responsible for executing that plan is you and your 20 colleagues (much more stressful!)

There’s still an incredibly tiny number of people seriously working on solving this problem, maybe a few dozen serious researchers. Nobody’s on the ball!

The intelligence explosion will be more like running a war than launching a product. We’re not on track for superdefense, for an airgapped cluster or any of that; I’m not sure we would even realize if a model self-exfiltrated. We’re not on track for a sane chain of command to make any of these insanely high-stakes decisions, to insist on the very-high-confidence appropriate for superintelligence, to make the hard decisions to take extra time before launching the next training run to get safety right or dedicate a large majority of compute to alignment research, to recognize danger ahead and avert it rather than crashing right into it. Right now, no lab has demonstrated much of a willingness to make any costly tradeoffs to get safety right (we get lots of safety committees, yes, but those are pretty meaningless). By default, we’ll probably stumble into the intelligence explosion and have gone through a few OOMs before people even realize what we’ve gotten into.

We’re counting way too much on luck here

(73) Page 126 (start of IIId): Superintelligence will give a decisive economic and military advantage. China isn’t at all out of the game yet. In the race to AGI, the free world’s very survival will be at stake. Can we maintain our preeminence over the authoritarian powers? And will we manage to avoid self-destruction along the way?

(74) Page 127: Our generation too easily takes for granted that we live in peace and freedom. And those who herald the age of AGI in SF too often ignore the elephant in the room: superintelligence is a matter of national security, and the United States must win.

The advent of superintelligence will put us in a situation unseen since the advent of the atomic era: those who have it will wield complete dominance over those who don’t.

A lead of a year or two or three on superintelligence could mean as utterly decisive a military advantage as the US coalition had against Iraq in the Gulf War. A complete reshaping of the military balance of power will be on the line.

Of course, we don’t know the limits of science and the many frictions that could slow things down. But no godlike advances are necessary for a decisive military advantage. And a billion superintelligent scientists will be able to do a lot. It seems clear that within a matter of years, pre-superintelligence militaries would become hopelessly outclassed.

To be even clearer: it seems likely the advantage conferred by superintelligence would be decisive enough even to preemptively take out an adversary’s nuclear deterrent.

It would simply be no contest. If there is a rapid intelligence explosion, it’s plausible a lead of mere months could be decisive.

(for example, the Yi-34B architecture seems to have essentially the Llama2 architecture, with merely a few lines of code changed)

That’s all merely a prelude, however. If and when the CCP wakes up to AGI, we should expect extraordinary efforts on the part of the CCP to compete. And I think there’s a pretty clear path for China to be in the game: outbuild the US and steal the algorithms.

1a. Chips: China now seems to have demonstrated the ability to manufacture 7nm chips. While going beyond 7nm will be difficult (requiring EUV), 7nm is enough! For reference, 7nm is what Nvidia A100s used. The indigenous Huawei Ascend 910B, based on the SMIC 7nm platform, seems to only be ~2-3x worse on performance/$ than an equivalent Nvidia chip would be.

1b. Outbuilding the US: The binding constraint on the largest training clusters won’t be chips, but industrial mobilization— perhaps most of all the 100GW of power for the trillion-dollar cluster. But if there’s one thing China can do better than the US it’s building stuff.

(75) Page 134: To date, US tech companies have made a much bigger bet on AI and scaling than any Chinese efforts; consequently, we are well ahead. But counting out China now is a bit like counting out Google in the AI race when ChatGPT came out in late 2022.

(76) Page 134: A dictator who wields the power of superintelligence would command concentrated power unlike any we’ve ever seen. In addition to being able to impose their will on other countries, they could enshrine their rule internally.

To be clear, I don’t just worry about dictators getting superintelligence because “our values are better.” I believe in freedom and democracy, strongly, because I don’t know what the right values are.

Superintelligence will give those who wield it the power to crush opposition, dissent, and lock in their grand plan for humanity. It will be difficult for anyone to resist the terrible temptation to use this power. I hope, dearly, that we can instead rely on the wisdom of the Framers—letting radically different values flourish, and preserving the raucous plurality that has defined the American experiment.

(77) Page 136: Maintaining a healthy lead will be decisive for safety

On the historical view, the greatest existential risk posed by AGI is that it will enable us to develop extraordinary new means of mass death. This time, these means could even proliferate to become accessible to rogue actors or terrorists.

(78) Page 138: Some hope for some sort of international treaty on safety. This seems fanciful to me. The world where both the CCP and USG are AGI-pilled enough to take safety risk seriously is also the world in which both realize that international economic and military predominance is at stake, that being months behind on AGI could mean being permanently left behind.

Perhaps most importantly, a healthy lead gives us room to maneuver: the ability to “cash in” parts of the lead, if necessary, to get safety right, for example by devoting extra work to alignment during the intelligence explosion.

(79) Page 139: Slowly, the USG is starting to move. The export controls on American chips are a huge deal, and were an incredibly prescient move at the time. But we have to get serious across the board.

The US has a lead. We just have to keep it. And we’re screwing that up right now. Most of all, we must rapidly and radically lock down the AI labs.

(80) Page 141 (Start of Part 4): As the race to AGI intensifies, the national security state will get involved. The USG will wake from its slumber, and by 27/28 we’ll get some form of government AGI project. No startup can handle superintelligence. Somewhere in a SCIF, the endgame will be on.

(81) Page 142: I find it an insane proposition that the US government will let a random SF startup develop superintelligence. Imagine if we had developed atomic bombs by letting Uber just improvise.

It is a delusion of those who have unconsciously internalized our brief respite from history that this will not summon more primordial forces. Like many scientists before us, the great minds of San Francisco hope that they can control the destiny of the demon they are birthing. Right now, they still can; for they are among the few with situational awareness, who understand what they are building. But in the next few years, the world will wake up. So too will the national security state. History will make a triumphant return.

In any case, my main claim is not normative, but descriptive. In a few years, The Project will be on.

(82) Page 145: And somewhere along here, we’ll get the first genuinely terrifying demonstrations of AI: perhaps the oft-discussed “helping novices make bioweapons,” or autonomously hacking critical systems, or something else entirely. It will become clear: this technology will be an utterly decisive military technology.

As with Covid, and even the Manhattan Project, the government will be incredibly late and hamfisted.

(83) Page 146: There are many ways this could be operationalized in practice. To be clear, this doesn’t need to look like literal nationalization, with AI lab researchers now employed by the military or whatever (though it might!). Rather, I expect a more suave orchestration. The relationship with the DoD might look like the relationship the DoD has with Boeing or Lockheed Martin.

Perhaps via defense contracting or similar, a joint venture between the major cloud compute providers, AI labs, and the government is established, making it functionally a project of the national security state. Much like the AI labs “voluntarily” made commitments to the White House in 2023, Western labs might more-or-less “voluntarily” agree to merge in the national effort.

(84) Page 147: Simply put, it will become clear that the development of AGI will fall in a category more like nukes than the internet. Yes, of course it’ll be dual-use—but nuclear technology was dual use too.

It seems pretty clear: this should not be under the unilateral command of a random CEO. Indeed, in the private-labs-developing-superintelligence world, it’s quite plausible individual CEOs would have the power to literally coup the US government.

(85) Page 150: Safety

Simply put: there are a lot of ways for us to mess this up— from ensuring we can reliably control and trust the billions of superintelligent agents that will soon be in charge of our economy and military (the superalignment problem), to controlling the risks of misuse of new means of mass destruction.

Some AI labs claim to be committed to safety: acknowledging that what they are building, if gone awry, could cause catastrophe and promising that they will do what is necessary when the time comes. I do not know if we can trust their promise enough to stake the lives of every American on it. More importantly, so far, they have not demonstrated the competence, trustworthiness, or seriousness necessary for what they themselves acknowledge they are building.

At core, they are startups, with all the usual commercial incentives.

(86) Page 151: One answer is regulation. That may be appropriate in worlds in which AI develops more slowly, but I fear that regulation simply won’t be up to the nature of the challenge of the intelligence explosion. What’s necessary will be less like spending a few years doing careful evaluations and pushing some safety standards through a bureaucracy. It’ll be more like fighting a war.

We’ll face an insane year in which the situation is shifting extremely rapidly every week, in which hard calls based on ambiguous data will be life-or-death, in which the solutions—even the problems themselves—won’t be close to fully clear ahead of time but come down to competence in a “fog of war,” which will involve insane tradeoffs like “some of our alignment measurements are looking ambiguous, we don’t really understand what’s going on anymore, it might be fine but there’s some warning signs that the next generation of superintelligence might go awry, should we delay the next training run by 3 months to get more confidence on safety—but oh no, the latest intelligence reports indicate China stole our weights and is racing ahead on their own intelligence explosion, what should we do?”.

I’m not confident that a government project would be competent in dealing with this—but the “superintelligence developed by startups” alternative seems much closer to “praying for the best” than commonly recognized.

(87) Page 153: We’ll need the government project to win the race against the authoritarian powers—and to give us the clear lead and breathing room necessary to navigate the perils of this situation.

We will want to bundle Western efforts: bring together our best scientists, use every GPU we can find, and ensure the trillions of dollars of cluster buildouts happen in the United States. We will need to protect the datacenters against adversary sabotage, or outright attack.

Perhaps, most of all, it will take American leadership to develop— and if necessary, enforce—a nonproliferation regime.

Ultimately, my main claim here is descriptive: whether we like it or not, superintelligence won’t look like an SF startup, and in some way will be primarily in the domain of national security. I’ve brought up The Project a lot to my San Francisco friends in the past year. Perhaps what’s surprised me most is how surprised most people are about the idea. They simply haven’t considered the possibility. But once they consider it, most agree that it seems obvious.

(88) Page 154: Perhaps the most important free variable is simply whether the inevitable government project will be competent. How will it be organized? How can we get this done? How will the checks and balances work, and what does a sane chain of command look like? Scarcely any attention has gone into figuring this out. Almost all other AI lab and AI governance politicking is a sideshow. This is the ballgame.

(89) Conclusion: And so by 27/28, the endgame will be on. By 28/29 the intelligence explosion will be underway; by 2030, we will have summoned superintelligence, in all its power and might.

Whoever they put in charge of The Project is going to have a hell of a task: to build AGI, and to build it fast; to put the American economy on wartime footing to make hundreds of millions of GPUs; to lock it all down, weed out the spies, and fend off all-out attacks by the CCP; to somehow manage a hundred million AGIs furiously automating AI research, making a decade’s leaps in a year, and soon producing AI systems vastly smarter than the smartest humans; to somehow keep things together enough that this doesn’t go off the rails and produce rogue superintelligence that tries to seize control from its human overseers; to use those superintelligences to develop whatever new technologies will be necessary to stabilize the situation and stay ahead of adversaries, rapidly remaking US forces to integrate those; all while navigating what will likely be the tensest international situation ever seen. They better be good, I’ll say that.

For those of us who get the call to come along for the ride, it’ll be . . . stressful. But it will be our duty to serve the free world—and all of humanity. If we make it through and get to look back on those years, it will be the most important thing we ever did. And while whatever secure facility they find probably won’t have the pleasantries of today’s ridiculouslyovercomped-AI-researcher-lifestyle, it won’t be so bad. SF already feels like a peculiar AI-researcher-college-town; probably this won’t be so different. It’ll be the same weirdly-small circle sweating the scaling curves during the day and hanging out over the weekend, kibitzing over AGI and the lab-politics-of-the-day.

Except, well—the stakes will be all too real.

See you in the desert, friends.

Quotes from Leopold Aschenbrenner’s Situational Awareness Paper Read More »

f1-cars-in-2026-will-be-smaller,-safer,-more-nimble,-more-sustainable

F1 cars in 2026 will be smaller, safer, more nimble, more sustainable

A render of a 2026 F1 car

Enlarge / For 2026, F1 cars are going on a little bit of a diet.

Fédération Internationale de l’Automobile

Earlier today, the Fédération Internationale de l’Automobile laid out the direction for Formula 1’s next set of technical regulations, which will go into effect in 2026. It will be the second big shakeup of F1’s technical regs since 2022 and involves sweeping changes to the hybrid powertrain and a fundamental rethink of how some of the aerodynamics work.

“With this set of regulations, the FIA has sought to develop a new generation of cars that are fully in touch with the DNA of Formula 1—cars that are light, supremely fast and agile but which also remains at the cutting edge of technology, and to achieve this we worked towards what we called a ‘nimble car’ concept. At the center of that vision is a redesigned power unit that features a more even split between the power derived from the internal combustion element and electrical power,” said Nikolas Tombazis, the FIA’s single-seater technical director.

Didn’t we just get new rules?

It feels like F1 only just got through its last big rule change with the (re)introduction of ground-effect cars at the start of 2022. Since the early 1980s, F1 cars have generated aerodynamic grip, or downforce, via front and rear wings. But drivers found it increasingly difficult to follow each other closely through corners as the dirty air from the car in front starved the following car’s front wing of air, robbing it of cornering grip in the process.

The 2022 rules changed this, requiring cars to use a sculpted floor that generates downforce via the venturi effect. This reduced the importance of the front wing, and indeed, the cars were able to race closely in 2022. In two years’ time, F1 cars will use less complicated floors with smaller venturis that generate a smaller ground effect, which the FIA says should mean no more having to run “ultra-stiff and low setups” to avoid the problem of porpoising.

Overall downforce is being reduced by 30 percent, but there’s an even greater reduction in drag—55 percent is the target, which is being done in part to accommodate the new hybrid powertrain.

More hybrid power

The V6 internal combustion engine is becoming less powerful, dropping to an output of 536 hp (400 kW), but the electric motor that also drives the rear wheels will now generate 470 hp (350 kW). That leaves the combined power output roughly where it is today, but only when the battery has enough charge. However, cars will be allowed to harvest twice as much energy (8.5 MJ) per lap under braking as now.

And as Ars has covered in the past, the engines will run on drop-in sustainable fuels. The new engine regulations have succeeded in tempting Honda back into the sport, as well as bringing in Ford and Audi, and possibly Cadillac in time.

Since the cars will be less powerful when they’re just running on internal combustion, more than halving the amount of drag they experience means they shouldn’t be too slow along the straights.

When F1 first introduced its original hybrid, called KERS (for kinetic energy recovery system), the electric motor boost was something the driver could use on demand. But that changed when the current powertrain rules came into effect in 2014, and it became up to the car to decide when to deploy energy from the battery to supplement the V6 motor.

In 2026, that changes again. The hybrid system is programmed to use less of the electric motor’s power as speeds pass 180 mph (270 km/h), down to zero at 220 mph (355 km/h), relying just on the V6 by then. But if a car is following within a second, the chasing driver can override that cutoff, allowing the full 470 hp from the electric motor at speeds of up to 209 mph (337 km/h), with up to half a MJ of extra energy.

F1 cars in 2026 will be smaller, safer, more nimble, more sustainable Read More »

apple-will-update-iphones-for-at-least-5-years-in-rare-public-commitment

Apple will update iPhones for at least 5 years in rare public commitment

finally, something in writing —

UK regulation requires companies to say how long they plan to provide support.

Apple will update iPhones for at least 5 years in rare public commitment

Apple

Apple has taken a rare step and publicly committed to a software support timeline for one of its products, as pointed out by MacRumors. A public regulatory filing for the iPhone 15 Pro (PDF) confirms that Apple will support the device with new software updates for at least five years from its “first supply date” of September 22, 2023, which would guarantee support until at least 2028.

Apple published the filing to comply with new Product Security and Telecommunications Infrastructure (PSTI) regulations from the UK that went into effect in late April. As this plain-language explainer from the Center for Cybersecurity Policy and Law summarizes, the PSTI regulations (among other things) don’t mandate a specific support window for manufacturers of Internet-connected devices, but they do require companies to publish a concrete support window and contact information for someone at the company who can be contacted with bug reports.

As publications like Android Authority have pointed out, five years is less than some Android phone makers like Google and Samsung have publicly committed to; both companies have said they’ll support their latest devices for seven years. But in reality, Apple usually hits or exceeds this seven-year timeline for updates—and does so for iPhones released nearly a decade ago and not just its newest products.

2017’s iPhone 8 and iPhone X, for example, are still receiving iOS 16 security updates. 2015’s iPhone 6S and 2016’s iPhone 7 were receiving iOS 15 updates as recently as March 2024, though these appear to have dried up in recent months. Each of these iPhones also received six or seven years’ worth of new major iOS releases, though not every phone that gets an iOS update supports every feature that newer devices get.

So Apple’s five-year pledge is notable less because it’s an improvement on or departure from the norm but more because Apple virtually never commits to software support timelines in writing.

Take those iOS 15 updates—Apple provided them for nearly a year and a half for iPhones and iPads that didn’t meet the requirements for iOS 16 or 17 but then abruptly (apparently) stopped releasing them. There was never a public commitment to continue releasing iOS 15 updates after iOS 16 came out, nor has there been any statement about iOS 15 updates being discontinued; we can only assume based on the fact that multiple iOS 16 and 17 updates have been released since March with no corresponding update for iOS 15.

The situation with the Mac is the same. Apple’s longstanding practice for decades has been to support the current version of macOS plus the two preceding versions, but that policy is not written down anywhere.

Contrast this with Microsoft, which generally commits to 10-year support timelines for new versions of Windows and publishes specific end-of-support dates years in advance; when Microsoft makes changes, it’s usually to extend the availability of updates in some way. Google has been making similar commitments for Chromebooks and officially certified ChromeOS Flex devices. These public timelines may tie a company’s hands, but they also make it easier for individuals, businesses, and schools to plan technology purchases and upgrades, and make it easier to know exactly how much support you can expect for a hand-me-down used or refurbished system.

Though the PSTI regulations only technically apply in the UK, it’s unlikely that Apple would go to the trouble of releasing iOS security updates in some countries without releasing those updates in all of them. But because a five-year support timeline is so much shorter than what Apple normally provides, it probably won’t matter that much. If Apple exceeds its stated support timeline, the PSTI law requires it to publish a new timeline “as soon as is practicable,” but for now, that date is far off.

Apple will update iPhones for at least 5 years in rare public commitment Read More »

the-motorola-edge-2024-comes-to-the-us-for-$550

The Motorola Edge 2024 comes to the US for $550

Fix your update plan —

Motorola’s Pixel 8a fighter is headed to a carrier store near you.

  • The Motorola Edge 2024.

    Motorola

  • Some companies are still making curved screens.

    Motorola

  • The bottom, which just features a sim tray, speaker, and USB-C port.

    Motorola

  • The top.

    Motorola

  • The side.

    Motorola

Motorola’s newest phone is the Motorola Edge 2024. This is a mid-range phone with the new Qualcomm Snapdragon 7s Gen 2. It costs $550 and will be in stores June 20. Every Motorola phone nowadays looks exactly the same, but Motorola assures us this is new.

The Snapdragon 7s Gen 2 is the bottom of Qualcomm’s “7 series” lineup and features four Cortex-A78 cores and four Cortex-A55 cores built on a 4 nm manufacturing process. The phone has a 144 Hz, 6.6-inch 2400×1080 OLED panel with curved sides. It has 8GB of RAM, 256GB of storage, a 5000 mAh battery, 68 W wired charging, and 15 W wireless charging. Cameras include a mid-range 50 MP Sony “LYTIA” 700C, 13 MP wide-angle, and a 32 MP front camera. The phone has NFC, Wi-Fi 6E, an in-screen fingerprint reader, and—a big addition compared to other Motorola devices—an IP68 rating for dust and water resistance.

Just like on the Moto G Stylus, this phone has a “vegan leather” back option that should be softer than the usual plastic, but it’s still plastic. Unlike that phone, there’s no headphone jack or MicroSD slot. A customizable hardware button on the left side of the phone lets you open the Google Assistant or whatever other app you choose.

Motorola doesn’t officially list the update plan on its site, but the 2023 Motorola Edge has a whopping one OS update and three years of security updates. We’ve asked Motorola if that changed and will update this article if we hear back. This is usually a major downside of Motorola devices. And speaking of updates, the $550 price tag makes this Motorola’s alternative to the $500 Google Pixel 8a, which is getting seven years of OS and security updates. The Pixel is a smaller device (6.1 inches) with a smaller battery (4482 mAh) and less storage (128GB), but the Pixel’s better software, faster SoC, and dramatically longer update plan make it easy to choose over the Motorola.

The phone will land at Amazon, Best Buy, and motorola.com on June 20, with “subsequent availability” at carrier stores like “T-Mobile, Metro by T-Mobile, Spectrum, Consumer Cellular, and on Straight Talk, Total By Verizon, and Visible.”

The Motorola Edge 2024 comes to the US for $550 Read More »

countdown-begins-for-third-try-launching-boeing’s-starliner-crew-capsule

Countdown begins for third try launching Boeing’s Starliner crew capsule

Going today? —

Astronauts Butch Wilmore and Suni Williams have been in prelaunch quarantine for six weeks.

Astronauts Suni Williams and Butch Wilmore, wearing their Boeing spacesuits, leave NASA's crew quarters during a launch attempt May 6.

Enlarge / Astronauts Suni Williams and Butch Wilmore, wearing their Boeing spacesuits, leave NASA’s crew quarters during a launch attempt May 6.

Fresh off repairs at the launch pad in Florida, United Launch Alliance engineers restarted the countdown overnight for the third attempt to send an Atlas V rocket and Boeing’s Starliner spacecraft on a test flight to the International Space Station.

NASA astronauts Butch Wilmore and Suni Williams were expected to awake early Wednesday, put on their blue pressure suits, and head to the launch pad at Cape Canaveral Space Force Station to board the Starliner capsule on top of the 172-foot-tall Atlas V rocket.

Once more through the door

Wilmore and Williams have done this twice before in hopes of launching into space on the first crew flight of Boeing’s Starliner spacecraft. A faulty valve on the Atlas V rocket prevented liftoff May 6, then engineers discovered a helium leak on the Starliner capsule itself. After several weeks of troubleshooting, NASA and Boeing officials decided to proceed with another launch attempt Saturday.

Everything seemed to be coming together for Boeing’s long-delayed crew test flight until a computer problem triggered an automatic hold in the countdown less than four minutes before liftoff. Technicians from United Launch Alliance (ULA), the Atlas V rocket’s builder and operator, traced the problem to a failed power distribution source connected to a ground computer responsible for controlling the final phase of the countdown.

The instantaneous launch opportunity Wednesday is set for 10: 52 am EDT (14: 52 UTC), when the launch site at Cape Canaveral passes underneath the space station’s orbital plane. Forecasters predict a 90 percent chance of good weather for launch. You can watch NASA’s live coverage in the video embedded below.

The countdown began late Tuesday night with the power-up of the Atlas V rocket, which was set to be filled with cryogenic liquid hydrogen and liquid oxygen propellants around 5 am EDT (09: 00 UTC). Kerosene fuel was loaded into the Atlas V’s first-stage booster prior to the mission’s first launch attempt in early May.

The two Starliner astronauts departed crew quarters at NASA’s Kennedy Space Center for the 20-minute drive to the launch pad, where they arrived shortly before 8 am EDT (12: 00 UTC) to climb into their seats inside the Starliner capsule. After pressure checks of the astronauts’ suits and Starliner’s crew cabin, ground teams will evacuate the pad about an hour before launch.

Assuming all systems are “go” for launch, the Atlas V will ignite its Russian-made RD-180 main engine and two solid-fueled boosters to vault away from Cape Canaveral and head northeast over the Atlantic Ocean. Wilmore and Williams will be not only the first people to fly in space on Boeing’s Starliner, but also the first astronauts to ride on an Atlas V rocket, which has flown 99 times before with satellites for the US military, NASA, and commercial customers.

The rocket’s Centaur upper stage will deploy Starliner into space around 15 minutes after liftoff. A critical burn by Starliner’s engines will happen around 31 minutes into the flight to finish the task of placing it into low-Earth orbit, setting it up for an automated docking at the International Space Station at 12: 15 pm EDT (16: 15 UTC) Thursday.

The two-person crew will stay on the station for at least a week, although a mission extension is likely if the mission is going well. Officials may decide to extend the mission to complete more tests or to wait for optimal weather conditions at Starliner’s primary and backup landing sites in New Mexico and Arizona. When weather conditions look favorable, Starliner will undock from the space station and head for landing under parachutes.

The crew test flight is a prerequisite to Boeing’s crew capsule becoming operational for NASA, which awarded multibillion-dollar commercial crew contracts to Boeing and SpaceX in 2014. SpaceX’s Crew Dragon started flying astronauts in 2020, while Boeing’s project has been stricken by years of delays.

Wilmore and Williams, both former US Navy test pilots, will take over manual control of Starliner at several points during the test flight. They will evaluate the spacecraft’s flying characteristics and accommodations for future flights, which will carry four astronauts at a time rather than two.

“The expectation from the media should not be perfection,” Wilmore told Ars earlier this year. “This is a test flight. Flying and operating in space is hard. It’s really hard, and we’re going to find some stuff. That’s expected. It’s the first flight where we are integrating the full capabilities of this spacecraft.”

Countdown begins for third try launching Boeing’s Starliner crew capsule Read More »

new-trailer-for-alien:-romulus-just-wants-to-give-us-a-big,-warm-face-hug

New trailer for Alien: Romulus just wants to give us a big, warm face-hug

No one can hear you scream —

Beware abandoned space stations “haunted” by xenomorphs.

Director Fede Alvarez promises to bring the sci-fi franchise back to its horror roots with Alien: Romulus.

We got our first look at Alien: Romulus, the ninth installment in the sci-fi franchise, in March with a brief teaser. That footage showed promise that horror director Fede Alvarez (Don’t Breathe, Evil Dead) could fulfill his intention to bring this standalone film back to the franchise’s stripped-down space horror roots. Now we have the full trailer, and we’re pretty confident he’s kept that promise. It looks as gory, intense, and delightfully terrifying as the seminal first two films in the franchise.

(Spoilers for Alien and Aliens below.)

As previously reported, Alien: Romulus is set between the events of Alien and Aliens (and is not related to FX/Hulu’s Alien prequel series slated to premiere next year). That is, after Ellen Ripley, the sole survivor of the Nostromo, destroyed the killer xenomorph and launched herself into space in the ship’s lifeboat—along with the ginger cat, Jonesy—and before she woke up after 57 years in hypersleep and battled more xenomorphs while protecting the young orphan, Newt (Carrie Henn). Per the short-and-sweet official premise: “While scavenging the deep ends of a derelict space station, a group of young space colonizers come face to face with the most terrifying life form in the universe.”

Cailee Spaeny (Priscilla, Pacific Rim: Uprising) stars as Rain Carradine, Isabela Merced (The Last of Us) plays Kay, and David Jonsson (Murder Is Easy) plays Andy. Archie Renaux (Shadow and Bone) plays Tyler, Spike Fearn (Aftersun) plays Bjorn, and Aileen Wu plays Navarro. But we aren’t likely to see iconic badass Ellen Ripley (immortalized by Sigourney Weaver) in the film. At this point in the timeline, she’s in the middle of her 57-year stasis with Jonesy as her escape shuttle travels through space toward her fateful encounter with a xenomorph queen.

The teaser offered little more than panicked calls for help (“Get it away from me!”), piercing screams, and a shot of a gore-spattered wall, along with a few frenetic shots of panicked crew members fleeing the alien xenomorph that is no doubt delighted to have fresh hosts in which to hatch its deadly offspring. There was also some special footage screened at CinemaCon in April featuring the expected face-huggers and chest-bursters.

The new trailer opens with ominous heavy footsteps (which punctuate the footage throughout) as Tyler asks Rain if this is really where she wants to spend the rest of her life. It’s unclear which place “this” refers to, but Rain definitely wants to escape, and Tyler has found what he claims is their “only ticket out of here”: becoming space colonizers, one presumes.

Cue the spooky “haunted house in space” vibes as Rain, Tyler, and their fellow colonists explore the aforementioned derelict space station—and get far more than they bargained for, including being attacked by face huggers. We also get a shot of Navarro’s horror as a chest-burster hammers against her rib cage. Kudos to whoever edited this trailer to remove the sound for the final 30 seconds, right after lettering spells out the classic tagline (“In space, no one can hear you scream”). It makes Rain’s final quiet line (“Are you sure you wanna do this?”) and the sudden burst of screaming at the end that much more effective.

Alien: Romulus hits theaters on August 16, 2024.

20th Century Studios

Listing image by 20th Century Studios

New trailer for Alien: Romulus just wants to give us a big, warm face-hug Read More »

the-2024-subaru-solterra-is-nimble-but-sorely-lacks-range,-personality

The 2024 Subaru Solterra is nimble but sorely lacks range, personality

how about an electric Baja —

Slow charging and inefficient driving, Solterra is no electric WRX or Forester.

A Subaru Solterra drives on a dirt road

Enlarge / With just 222 miles of range, you can’t venture far off-grid in the Subaru Solterra.

Subaru

Over the years, Subaru has generated a cult following in the US, making its name with all-wheel drive powertrains and a go-anywhere attitude. Cars like the rally-bred WRXes and STIs did a lot of work here, but lately, Subaru has seemed to go in the opposite direction, phasing out fun drives like the STI lineup in favor of volume-movers like the Ascent and bloated versions of existing models such as the Subaru Wilderness editions.

Its first electric vehicle is perhaps even less in character. The $44,995 Solterra is the result of an ongoing partnership with Toyota and was developed together with the bZ4X. Unlike the Toyota, there’s no single-motor option for the Solterra. It’s all-wheel-drive only, with a pair of identical 107 hp (80 kW) permanent magnet electric motors, one for each axle. That means you can do some, but not all, of the off-road things you’d expect to do with a Subaru.

Looks are deceiving

At first glance, the Solterra looks like the edgy, tech-leaning offspring of a Crosstrek and an Impreza wagon. The 8.3 inches of ground clearance is slightly less than the Outback or Forester, while the Solterra comes in at 184.6 inches (4,689 mm) in length, placing it squarely in the middle of the brand’s stable. It’s a rather compact SUV, even more so when you try to get comfortable in the cockpit. My short frame was cramped, and anyone taller than me won’t feel welcome on long drives.

The large multifunction steering wheel can obscure the small instrument display in front of the driver.

Enlarge / The large multifunction steering wheel can obscure the small instrument display in front of the driver.

Subaru

In what seems to be the norm with Subaru these days, the interior is full of plastic and cloth. Even on this top-line Touring trim test car, which comes in at just under $55,000, there’s a very cheap-looking dash with a plethora of rigid lines. Controls are close by, but the overall layout is borderline infuriating, with slow response times through the central infotainment system and a driver alert system that beeped and shrieked every 20 seconds for one reason or another. There were so many driver warnings and advisories popping up that I eventually tuned them out, which is probably not the intended effect.

Range Non-Rover

There’s about five miles (8 km) of charging difference between the 228-mile (367 km) Premium trim level and the Limited and Touring trims, which have an EPA range of 222 miles (357 km) on a single charge of the 72.8 kWh lithium-ion battery. In my 10 days with the car, the only time I eclipsed 200 miles (321 km) was leaving my driveway with the range reading 201. After about 10 minutes, it slumped back under 200 miles. In fairly normal city and highway conditions, I realized around 180 miles of range (290 km). When the weather called for air conditioning, I lost another 5–7 miles (8–11 km).

  • The Solterra is 184.6 inches (4,689 mm) long, 73.2 inches (1,859 mm) wide, 65 inches (1,651 mm) tall, with a 112.2-inch (2,850 mm) wheelbase. It has a curb weight of between 4,365 and 4,505 lbs (1,980–2,043 kg) depending on trim level.

    Subaru

  • The Toyota-developed infotainment system can be laggy.

    Subaru

  • The back seat has 35.5 inches (902 mm) of rear legroom.

    Subaru

  • There’s 27.7 cubic feet (783 L) of cargo volume with the rear seats in use and the cover in place.

    Subaru

  • Wireless device charging, as well as wireless Apple CarPlay and Android Auto, are available in the Limited and Touring trims.

    Subaru

Charging is slow, however. A stop to recharge from about 20 to 80 percent state of charge took the better part of 45 minutes. At launch, the Solterra was rated at an even longer 56 minutes to DC fast-charge to 80 percent, but for model year 2024, Subaru says that in ideal conditions, this should now be as quick as 35 minutes.

Charging at home was an overnight endeavor—nine hours on a level 2 charger. The Solterra currently features a CCS1 charge port, but in 2025, the company will adopt the J3400 standard, with adapters made available to existing customers so they can charge at Tesla Supercharger sites.

The 2024 Subaru Solterra is nimble but sorely lacks range, personality Read More »