microsoft

having-recovery-and/or-ssd-problems-after-recent-windows-updates?-you’re-not-alone.

Having recovery and/or SSD problems after recent Windows updates? You’re not alone.

The other issue some users have been experiencing is potentially more serious, but also harder to track down. Tom’s Hardware has a summary of the problem: At some point after installing update KB5063878 on Windows 11 24H2, some users began noticing issues with large file transfers on some SSDs. When installing a large update for Cyberpunk 2077, a large game that requires dozens of gigabytes of storage, Windows abruptly stopped seeing the SSD that the game was installed on.

The issues are apparently more pronounced on disks that are more than 60 percent full, when transferring at least 50GB of data. Most of the SSDs were visible again after a system reboot, though one—a 2TB Western Digital SA510 drive—didn’t come back after a reboot.

These issues could be specific to this user’s configuration, and the culprit may not be the Windows update. Microsoft has yet to add the SSD problem to its list of known issues with Windows, but the company confirmed to Ars that it was studying the complaints.

“We’re aware of these reports and are investigating with our partners,” a Microsoft spokesperson told Ars.

SSD controller manufacturer Phison told Tom’s Hardware that it was also looking into the problem.

Having recovery and/or SSD problems after recent Windows updates? You’re not alone. Read More »

sony-makes-the-“difficult-decision”-to-raise-playstation-5-prices-in-the-us

Sony makes the “difficult decision” to raise PlayStation 5 prices in the US

Sony will join Microsoft and Nintendo in raising US prices across its entire game console lineup, the company announced today. Pricing for all current versions of the PlayStation 5 console will increase by $50 starting tomorrow.

The price of the PS5 Digital Edition will increase from $450 to $500; the standard PS5 will increase from $500 to $550; and the PS5 Pro will increase from $700 to $750. If you’ve been on the fence about buying any of these, retailers like Target and Best Buy are still using the old prices as of this writing—for other console price hikes, retailers have sometimes bumped the prices up before the date announced by the manufacturer.

“Similar to many global businesses, we continue to navigate a challenging economic environment,” wrote Sony Global Marketing VP Isabelle Tomatis. “As a result, we’ve made the difficult decision to increase the recommended retail price for PlayStation 5 consoles in the U.S. starting on August 21.”

Sony says it’s not increasing prices for games or accessories and that this round of price increases only affects consoles sold in the US.

Sony was the last of the big three console makers to raise prices this year. Microsoft raised the prices for the Xbox Series S and X consoles in March. And Nintendo has gone through two rounds of price increases—one for Switch and Switch 2 accessories in April and another for more accessories and Switch 1 consoles earlier this month.

Sony makes the “difficult decision” to raise PlayStation 5 prices in the US Read More »

microsoft-and-asus’-answers-to-steamos-and-the-steam-deck-launch-on-october-16

Microsoft and Asus’ answers to SteamOS and the Steam Deck launch on October 16

Asus and Microsoft will be launching their ROG Xbox Ally series of handheld gaming PCs starting October 16, according to an Asus announcement that went out today.

An Xbox-branded extension of Asus’ existing ROG Ally handheld line, the basic ROG Xbox Ally and more powerful ROG Xbox Ally X, both run a version of Windows 11 Home that’s been redesigned with a controller-first Xbox-style user interface. The idea is to preserve the wide game compatibility of Windows—and the wide compatibility with multiple storefronts, including Microsoft’s own, Valve’s Steam, the Epic Games Store, and more—while turning off all of the extra Windows desktop stuff and saving system resources. (This also means that, despite the Xbox branding, these handhelds play Windows PC games and not the Xbox versions.)

Microsoft and Asus initially announced the handhelds in June. Microsoft still isn’t sharing pricing information for either console, so it’s hard to say how their specs and features will stack up against the Steam Deck (starting at $399 for the LCD version, $549 for OLED), Nintendo’s Switch 2 ($450), or past Asus handhelds like the ROG Ally X ($800).

Both consoles share a 7-inch, 1080p IPS display with a 120 Hz refresh rate, Wi-Fi 6E, and Bluetooth 5.4 support, but their internals are quite a bit different. The lower-end Xbox Ally uses an AMD Ryzen Z2 A chip with a 4-core Zen 2-based CPU, an eight-core RDNA2-based GPU, 512GB of storage, and 16GB of LPDDR5X-6400—specs nearly identical to Valve’s 3-year-old Steam Deck. The Xbox Ally X includes a more interesting Ryzen AI Z2 Extreme with an 8-core Zen 5 CPU, a 16-core RDNA3.5 GPU, 1TB of storage, 24GB of LPDDR5X-8000, and a built-in neural processing unit (NPU).

The beefier hardware comes with a bigger battery—80 WHr in the Ally X, compared to 60 WHr in the regular Ally—and that also makes the Ally X around a tenth of a pound (or 45 grams) heavier than the Ally.

Microsoft and Asus’ answers to SteamOS and the Steam Deck launch on October 16 Read More »

github-will-be-folded-into-microsoft-proper-as-ceo-steps-down

GitHub will be folded into Microsoft proper as CEO steps down

Putting GitHub more directly under its AI umbrella makes some degree of sense for Microsoft, given how hard it has pushed tools like GitHub Copilot, an AI-assisted coding tool. Microsoft has continually iterated on GitHub Copilot since introducing it in late 2021, adding support for multiple language models and “agents” that attempt to accomplish plain-language requests in the background as you work on other things.

However, there have been problems, too. Copilot inadvertently exposed the private code repositories of a few major companies earlier this year. And a recent Stack Overflow survey showed that trust in AI-assisted coding tools’ accuracy may be declining even as usage has increased, citing the extra troubleshooting and debugging work caused by “solutions that are almost right, but not quite.”

It’s unclear whether Dohmke’s departure and the elimination of the CEO position will change much in terms of the way GitHub operates or the products it creates and maintains. As GitHub’s CEO, Dohmke was already reporting to Julia Liuson, president of the company’s developer division, and Liuson reported to Core AI group leader Jay Parikh. The CoreAI group itself is only a few months old—it was announced by Microsoft CEO Satya Nadella in January, and “build[ing] out GitHub Copilot” was already one of the group’s responsibilities.

“Ultimately, we must remember that our internal organizational boundaries are meaningless to both our customers and to our competitors,” wrote Nadella when he announced the formation of the CoreAI group.

GitHub will be folded into Microsoft proper as CEO steps down Read More »

ai-in-wyoming-may-soon-use-more-electricity-than-state’s-human-residents

AI in Wyoming may soon use more electricity than state’s human residents

Wyoming’s data center boom

Cheyenne is no stranger to data centers, having attracted facilities from Microsoft and Meta since 2012 due to its cool climate and energy access. However, the new project pushes the state into uncharted territory. While Wyoming is the nation’s third-biggest net energy supplier, producing 12 times more total energy than it consumes (dominated by fossil fuels), its electricity supply is finite.

While Tallgrass and Crusoe have announced the partnership, they haven’t revealed who will ultimately use all this computing power—leading to speculation about potential tenants.

A potential connection to OpenAI’s Stargate AI infrastructure project, announced in January, remains a subject of speculation. When asked by The Associated Press if the Cheyenne project was part of this effort, Crusoe spokesperson Andrew Schmitt was noncommittal. “We are not at a stage that we are ready to announce our tenant there,” Schmitt said. “I can’t confirm or deny that it’s going to be one of the Stargate.”

OpenAI recently activated the first phase of a Crusoe-built data center complex in Abilene, Texas, in partnership with Oracle. Chris Lehane, OpenAI’s chief global affairs officer, told The Associated Press last week that the Texas facility generates “roughly and depending how you count, about a gigawatt of energy” and represents “the largest data center—we think of it as a campus—in the world.”

OpenAI has committed to developing an additional 4.5 gigawatts of data center capacity through an agreement with Oracle. “We’re now in a position where we have, in a really concrete way, identified over five gigawatts of energy that we’re going to be able to build around,” Lehane told the AP. The company has not disclosed locations for these expansions, and Wyoming was not among the 16 states where OpenAI said it was searching for data center sites earlier this year.

AI in Wyoming may soon use more electricity than state’s human residents Read More »

microsoft-is-revamping-windows-11’s-task-manager-so-its-numbers-make-more-sense

Microsoft is revamping Windows 11’s Task Manager so its numbers make more sense

Copilot+ features, and annoying “features”

Microsoft continues to roll out AI features, particularly to PCs that meet the qualifications for the company’s Copilot+ features. These betas enable “agent-powered search” for Intel and AMD Copilot+ PCs, which continue to get most of these features a few weeks or months later than Qualcomm Snapdragon+ PCs. This agent is Microsoft’s latest attempt to improve the dense, labyrinthine Settings app by enabling natural-language search that knows how to respond to queries like “my mouse pointer is too small” or “how to control my PC by voice” (Microsoft’s examples). Like other Copilot+ features, this relies on your PC’s neural processing unit (NPU) to perform all processing locally on-device. Microsoft has also added a tutorial for the “Click to Do” feature that suggests different actions you can perform based on images, text, and other content on your screen.

Finally, Microsoft is tweaking the so-called “Second Chance Out of Box Experience” window (also called “SCOOBE,” pronounced “scooby”), the setup screen that you’ll periodically see on a Windows 11 PC even if you’ve already been using it for months or years. This screen attempts to enroll your PC in Windows Backup, to switch your default browser to Microsoft Edge and its default search engine to Bing, and to import favorites and history into Edge from whatever browser you might have been trying to use before.

If you, like me, experience the SCOOBE screen primarily as a nuisance rather than something “helpful,” it is possible to make it go away. Per our guide to de-cluttering Windows 11, open Settings, go to System, then to Notifications, scroll down, expand the “additional settings” drop-down, and uncheck all three boxes here to get rid of the SCOOBE screen and other irritating reminders.

Most of these features are being released simultaneously to the Dev and Beta channels of the Windows Insider program (from least- to most-stable, the four channels are Canary, Dev, Beta, and Release Preview). Features in the Beta channel are usually not far from being released into the public versions of Windows, so non-Insiders can probably expect most of these things to appear on their PCs in the next few weeks. Microsoft is also gearing up to release the Windows 11 25H2 update, this year’s big annual update, which will enable a handful of features that the company is already quietly rolling out to PCs running version 24H2.

Microsoft is revamping Windows 11’s Task Manager so its numbers make more sense Read More »

microsoft-to-stop-using-china-based-teams-to-support-department-of-defense

Microsoft to stop using China-based teams to support Department of Defense

Last week, Microsoft announced that it would no longer use China-based engineering teams to support the Defense Department’s cloud computing systems, following ProPublica’s investigation of the practice, which cybersecurity experts said could expose the government to hacking and espionage.

But it turns out the Pentagon was not the only part of the government facing such a threat. For years, Microsoft has also used its global workforce, including China-based personnel, to maintain the cloud systems of other federal departments, including parts of Justice, Treasury and Commerce, ProPublica has found.

This work has taken place in what’s known as the Government Community Cloud, which is intended for information that is not classified but is nonetheless sensitive. The Federal Risk and Authorization Management Program, the US government’s cloud accreditation organization, has approved GCC to handle “moderate” impact information “where the loss of confidentiality, integrity, and availability would result in serious adverse effect on an agency’s operations, assets, or individuals.”

The Justice Department’s Antitrust Division has used GCC to support its criminal and civil investigation and litigation functions, according to a 2022 report. Parts of the Environmental Protection Agency and the Department of Education have also used GCC.

Microsoft says its foreign engineers working in GCC have been overseen by US-based personnel known as “digital escorts,” similar to the system it had in place at the Defense Department.

Nevertheless, cybersecurity experts told ProPublica that foreign support for GCC presents an opportunity for spying and sabotage. “There’s a misconception that, if government data isn’t classified, no harm can come of its distribution,” said Rex Booth, a former federal cybersecurity official who now is chief information security officer of the tech company SailPoint.

“With so much data stored in cloud services—and the power of AI to analyze it quickly—even unclassified data can reveal insights that could harm US interests,” he said.

Microsoft to stop using China-based teams to support Department of Defense Read More »

openai-and-partners-are-building-a-massive-ai-data-center-in-texas

OpenAI and partners are building a massive AI data center in Texas

Stargate moves forward despite early skepticism

When OpenAI announced Stargate in January, critics questioned whether the company could deliver on its ambitious $500 billion funding promise. Trump ally and frequent Altman foe Elon Musk wrote on X that “They don’t actually have the money,” claiming that “SoftBank has well under $10B secured.”

Tech writer and frequent OpenAI critic Ed Zitron raised concerns about OpenAI’s financial position, noting the company’s $5 billion in losses in 2024. “This company loses $5bn+ a year! So what, they raise $19bn for Stargate, then what, another $10bn just to be able to survive?” Zitron wrote on Bluesky at the time.

Six months later, OpenAI’s Abilene data center has moved from construction to partial operation. Oracle began delivering Nvidia GB200 racks to the facility last month, and OpenAI reports it has started running early training and inference workloads to support what it calls “next-generation frontier research.”

Despite the White House announcement with President Trump in January, the Stargate concept dates back to March 2024, when Microsoft and OpenAI partnered on a $100 billion supercomputer as part of a five-phase plan. Over time, the plan evolved into its current form as a partnership with Oracle, SoftBank, and CoreWeave.

“Stargate is an ambitious undertaking designed to meet the historic opportunity in front of us,” writes OpenAI in the press release announcing the latest deal. “That opportunity is now coming to life through strong support from partners, governments, and investors worldwide—including important leadership from the White House, which has recognized the critical role AI infrastructure will play in driving innovation, economic growth, and national competitiveness.”

OpenAI and partners are building a massive AI data center in Texas Read More »

office-problems-on-windows-10?-microsoft’s-response-will-soon-be-“upgrade-to-11.”

Office problems on Windows 10? Microsoft’s response will soon be “upgrade to 11.”

Microsoft’s advertised end-of-support date for Windows 10 is October 14, 2025. But in reality, the company will gradually wind down support for the enduring popular operating system over the next three years. Microsoft would really like you to upgrade to Windows 11, especially if it also means upgrading to a new PC, but it also doesn’t want to leave hundreds of millions of home and business PCs totally unprotected.

Those competing goals have led to lots of announcements and re-announcements and clarifications about updates for both Windows 10 itself and the Office/Microsoft 365 productivity apps that many Windows users run on their PCs.

Today’s addition to the pile comes via The Verge, which noticed an update to a support document that outlined when Windows 10 PCs would stop receiving new features for the continuously updated Microsoft 365 apps. Most home users will stop getting new features in August 2026, while business users running the Enterprise versions can expect to stop seeing new features in either October 2026 or January 2027, depending on the product they’re using.

Microsoft had previously committed to supporting its Office apps through October 2028—both the Microsoft 365 versions and perpetually licensed versions like Office 2021 and Office 2024 that don’t get continuous feature updates. That timeline isn’t changing, but it will apparently only cover security and bug-fixing updates rather than updates that add new features.

And while the apps will still be getting updates, Microsoft’s support document makes it clear that users won’t always be able to get fixes for bugs that are unique to Windows 10. If an Office issue exists solely on Windows 10 but not on Windows 11, the official guidance from Microsoft support is that users should upgrade to Windows 11; any support for Windows 10 will be limited to “troubleshooting assistance only,” and “technical workarounds might be limited or unavailable.”

Office problems on Windows 10? Microsoft’s response will soon be “upgrade to 11.” Read More »

everything-tech-giants-will-hate-about-the-eu’s-new-ai-rules

Everything tech giants will hate about the EU’s new AI rules

The code also details expectations for AI companies to respect paywalls, as well as robots.txt instructions restricting crawling, which could help confront a growing problem of AI crawlers hammering websites. It “encourages” online search giants to embrace a solution that Cloudflare is currently pushing: allowing content creators to protect copyrights by restricting AI crawling without impacting search indexing.

Additionally, companies are asked to disclose total energy consumption for both training and inference, allowing the EU to detect environmental concerns while companies race forward with AI innovation.

More substantially, the code’s safety guidance provides for additional monitoring for other harms. It makes recommendations to detect and avoid “serious incidents” with new AI models, which could include cybersecurity breaches, disruptions of critical infrastructure, “serious harm to a person’s health (mental and/or physical),” or “a death of a person.” It stipulates timelines of between five and 10 days to report serious incidents with the EU’s AI Office. And it requires companies to track all events, provide an “adequate level” of cybersecurity protection, prevent jailbreaking as best they can, and justify “any failures or circumventions of systemic risk mitigations.”

Ars reached out to tech companies for immediate reactions to the new rules. OpenAI, Meta, and Microsoft declined to comment. A Google spokesperson confirmed that the company is reviewing the code, which still must be approved by the European Commission and EU member states amid expected industry pushback.

“Europeans should have access to first-rate, secure AI models when they become available, and an environment that promotes innovation and investment,” Google’s spokesperson said. “We look forward to reviewing the code and sharing our views alongside other model providers and many others.”

These rules are just one part of the AI Act, which will start taking effect in a staggered approach over the next year or more, the NYT reported. Breaching the AI Act could result in AI models being yanked off the market or fines “of as much as 7 percent of a company’s annual sales or 3 percent for the companies developing advanced AI models,” Bloomberg noted.

Everything tech giants will hate about the EU’s new AI rules Read More »

what-is-agi?-nobody-agrees,-and-it’s-tearing-microsoft-and-openai-apart.

What is AGI? Nobody agrees, and it’s tearing Microsoft and OpenAI apart.


Several definitions make measuring “human-level” AI an exercise in moving goalposts.

When is an AI system intelligent enough to be called artificial general intelligence (AGI)? According to one definition reportedly agreed upon by Microsoft and OpenAI, the answer lies in economics: When AI generates $100 billion in profits. This arbitrary profit-based benchmark for AGI perfectly captures the definitional chaos plaguing the AI industry.

In fact, it may be impossible to create a universal definition of AGI, but few people with money on the line will admit it.

Over this past year, several high-profile people in the tech industry have been heralding the seemingly imminent arrival of “AGI” (i.e., within the next two years). But there’s a huge problem: Few people agree on exactly what AGI means. As Google DeepMind wrote in a paper on the topic: If you ask 100 AI experts to define AGI, you’ll get “100 related but different definitions.”

This isn’t just academic navel-gazing. The definition problem has real consequences for how we develop, regulate, and think about AI systems. When companies claim they’re on the verge of AGI, what exactly are they claiming?

I tend to define AGI in a traditional way that hearkens back to the “general” part of its name: An AI model that can widely generalize—applying concepts to novel scenarios—and match the versatile human capability to perform unfamiliar tasks across many domains without needing to be specifically trained for them.

However, this definition immediately runs into thorny questions about what exactly constitutes “human-level” performance. Expert-level humans? Average humans? And across which tasks—should an AGI be able to perform surgery, write poetry, fix a car engine, and prove mathematical theorems, all at the level of human specialists? (Which human can do all that?) More fundamentally, the focus on human parity is itself an assumption; it’s worth asking why mimicking human intelligence is the necessary yardstick at all.

The latest example of this definitional confusion causing trouble comes from the deteriorating relationship between Microsoft and OpenAI. According to The Wall Street Journal, the two companies are now locked in acrimonious negotiations partly because they can’t agree on what AGI even means—despite having baked the term into a contract worth over $13 billion.

A brief history of moving goalposts

The term artificial general intelligence has murky origins. While John McCarthy and colleagues coined the term artificial intelligence at Dartmouth College in 1956, AGI emerged much later. Physicist Mark Gubrud first used the term in 1997, though it was computer scientist Shane Legg and AI researcher Ben Goertzel who independently reintroduced it around 2002, with the modern usage popularized by a 2007 book edited by Goertzel and Cassio Pennachin.

Early AI researchers envisioned systems that could match human capability across all domains. In 1965, AI pioneer Herbert A. Simon predicted that “machines will be capable, within 20 years, of doing any work a man can do.” But as robotics lagged behind computing advances, the definition narrowed. The goalposts shifted, partly as a practical response to this uneven progress, from “do everything a human can do” to “do most economically valuable tasks” to today’s even fuzzier standards.

“An assistant of inventor Captain Richards works on the robot the Captain has invented, which speaks, answers questions, shakes hands, tells the time, and sits down when it’s told to.” – September 1928. Credit: Getty Images

For decades, the Turing Test served as the de facto benchmark for machine intelligence. If a computer could fool a human judge into thinking it was human through text conversation, the test surmised, then it had achieved something like human intelligence. But the Turing Test has shown its age. Modern language models can pass some limited versions of the test not because they “think” like humans, but because they’re exceptionally capable at creating highly plausible human-sounding outputs.

The current landscape of AGI definitions reveals just how fractured the concept has become. OpenAI’s charter defines AGI as “highly autonomous systems that outperform humans at most economically valuable work”—a definition that, like the profit metric, relies on economic progress as a substitute for measuring cognition in a concrete way. Mark Zuckerberg told The Verge that he does not have a “one-sentence, pithy definition” of the concept. OpenAI CEO Sam Altman believes that his company now knows how to build AGI “as we have traditionally understood it.” Meanwhile, former OpenAI Chief Scientist Ilya Sutskever reportedly treated AGI as something almost mystical—according to a 2023 Atlantic report, he would lead employees in chants of “Feel the AGI!” during company meetings, treating the concept more like a spiritual quest than a technical milestone.

Dario Amodei, co-founder and chief executive officer of Anthropic, during the Bloomberg Technology Summit in San Francisco, California, US, on Thursday, May 9, 2024.

Dario Amodei, co-founder and chief executive officer of Anthropic, during the Bloomberg Technology Summit in San Francisco on Thursday, May 9, 2024. Credit: Bloomberg via Getty Images

Dario Amodei, CEO of Anthropic, takes an even more skeptical stance on the terminology itself. In his October 2024 essay “Machines of Loving Grace,” Amodei writes that he finds “AGI to be an imprecise term that has gathered a lot of sci-fi baggage and hype.” Instead, he prefers terms like “powerful AI” or “Expert-Level Science and Engineering,” which he argues better capture the capabilities without the associated hype. When Amodei describes what others might call AGI, he frames it as an AI system “smarter than a Nobel Prize winner across most relevant fields” that can work autonomously on tasks taking hours, days, or weeks to complete—essentially “a country of geniuses in a data center.” His resistance to AGI terminology adds another layer to the definitional chaos: Not only do we not agree on what AGI means, but some leading AI developers reject the term entirely.

Perhaps the most systematic attempt to bring order to this chaos comes from Google DeepMind, which in July 2024 proposed a framework with five levels of AGI performance: emerging, competent, expert, virtuoso, and superhuman. DeepMind researchers argued that no level beyond “emerging AGI” existed at that time. Under their system, today’s most capable LLMs and simulated reasoning models still qualify as “emerging AGI”—equal to or somewhat better than an unskilled human at various tasks.

But this framework has its critics. Heidy Khlaaf, chief AI scientist at the nonprofit AI Now Institute, told TechCrunch that she thinks the concept of AGI is too ill-defined to be “rigorously evaluated scientifically.” In fact, with so many varied definitions at play, one could argue that the term AGI has become technically meaningless.

When philosophy meets contract law

The Microsoft-OpenAI dispute illustrates what happens when philosophical speculation is turned into legal obligations. When the companies signed their partnership agreement, they included a clause stating that when OpenAI achieves AGI, it can limit Microsoft’s access to future technology. According to The Wall Street Journal, OpenAI executives believe they’re close to declaring AGI, while Microsoft CEO Satya Nadella has called the idea of using AGI as a self-proclaimed milestone “nonsensical benchmark hacking” on the Dwarkesh Patel podcast in February.

The reported $100 billion profit threshold we mentioned earlier conflates commercial success with cognitive capability, as if a system’s ability to generate revenue says anything meaningful about whether it can “think,” “reason,” or “understand” the world like a human.

Sam Altman speaks onstage during The New York Times Dealbook Summit 2024 at Jazz at Lincoln Center on December 04, 2024 in New York City.

Sam Altman speaks onstage during The New York Times Dealbook Summit 2024 at Jazz at Lincoln Center on December 4, 2024, in New York City. Credit: Eugene Gologursky via Getty Images

Depending on your definition, we may already have AGI, or it may be physically impossible to achieve. If you define AGI as “AI that performs better than most humans at most tasks,” then current language models potentially meet that bar for certain types of work (which tasks, which humans, what is “better”?), but agreement on whether that is true is far from universal. This says nothing of the even murkier concept of “superintelligence”—another nebulous term for a hypothetical, god-like intellect so far beyond human cognition that, like AGI, defies any solid definition or benchmark.

Given this definitional chaos, researchers have tried to create objective benchmarks to measure progress toward AGI, but these attempts have revealed their own set of problems.

Why benchmarks keep failing us

The search for better AGI benchmarks has produced some interesting alternatives to the Turing Test. The Abstraction and Reasoning Corpus (ARC-AGI), introduced in 2019 by François Chollet, tests whether AI systems can solve novel visual puzzles that require deep and novel analytical reasoning.

“Almost all current AI benchmarks can be solved purely via memorization,” Chollet told Freethink in August 2024. A major problem with AI benchmarks currently stems from data contamination—when test questions end up in training data, models can appear to perform well without truly “understanding” the underlying concepts. Large language models serve as master imitators, mimicking patterns found in training data, but not always originating novel solutions to problems.

But even sophisticated benchmarks like ARC-AGI face a fundamental problem: They’re still trying to reduce intelligence to a score. And while improved benchmarks are essential for measuring empirical progress in a scientific framework, intelligence isn’t a single thing you can measure like height or weight—it’s a complex constellation of abilities that manifest differently in different contexts. Indeed, we don’t even have a complete functional definition of human intelligence, so defining artificial intelligence by any single benchmark score is likely to capture only a small part of the complete picture.

The survey says: AGI may not be imminent

There is no doubt that the field of AI has seen rapid, tangible progress in numerous fields, including computer vision, protein folding, and translation. Some excitement of progress is justified, but it’s important not to oversell an AI model’s capabilities prematurely.

Despite the hype from some in the industry, many AI researchers remain skeptical that AGI is just around the corner. A March 2025 survey of AI researchers conducted by the Association for the Advancement of Artificial Intelligence (AAAI) found that a majority (76 percent) of researchers who participated in the survey believed that scaling up current approaches is “unlikely” or “very unlikely” to achieve AGI.

However, such expert predictions should be taken with a grain of salt, as researchers have consistently been surprised by the rapid pace of AI capability advancement. A 2024 survey by Grace et al. of 2,778 AI researchers found that experts had dramatically shortened their timelines for AI milestones after being surprised by progress in 2022–2023. The median forecast for when AI could outperform humans in every possible task jumped forward by 13 years, from 2060 in their 2022 survey to 2047 in 2023. This pattern of underestimation was evident across multiple benchmarks, with many researchers’ predictions about AI capabilities being proven wrong within months.

And yet, as the tech landscape shifts, the AI goalposts continue to recede at a constant speed. Recently, as more studies continue to reveal limitations in simulated reasoning models, some experts in the industry have been slowly backing away from claims of imminent AGI. For example, AI podcast host Dwarkesh Patel recently published a blog post arguing that developing AGI still faces major bottlenecks, particularly in continual learning, and predicted we’re still seven years away from AI that can learn on the job as seamlessly as humans.

Why the definition matters

The disconnect we’ve seen above between researcher consensus, firm terminology definitions, and corporate rhetoric has a real impact. When policymakers act as if AGI is imminent based on hype rather than scientific evidence, they risk making decisions that don’t match reality. When companies write contracts around undefined terms, they may create legal time bombs.

The definitional chaos around AGI isn’t just philosophical hand-wringing. Companies use promises of impending AGI to attract investment, talent, and customers. Governments craft policy based on AGI timelines. The public forms potentially unrealistic expectations about AI’s impact on jobs and society based on these fuzzy concepts.

Without clear definitions, we can’t have meaningful conversations about AI misapplications, regulation, or development priorities. We end up talking past each other, with optimists and pessimists using the same words to mean fundamentally different things.

In the face of this kind of challenge, some may be tempted to give up on formal definitions entirely, falling back on an “I’ll know it when I see it” approach for AGI—echoing Supreme Court Justice Potter Stewart’s famous quote about obscenity. This subjective standard might feel useful, but it’s useless for contracts, regulation, or scientific progress.

Perhaps it’s time to move beyond the term AGI. Instead of chasing an ill-defined goal that keeps receding into the future, we could focus on specific capabilities: Can this system learn new tasks without extensive retraining? Can it explain its outputs? Can it produce safe outputs that don’t harm or mislead people? These questions tell us more about AI progress than any amount of AGI speculation. The most useful way forward may be to think of progress in AI as a multidimensional spectrum without a specific threshold of achievement. But charting that spectrum will demand new benchmarks that don’t yet exist—and a firm, empirical definition of “intelligence” that remains elusive.

Photo of Benj Edwards

Benj Edwards is Ars Technica’s Senior AI Reporter and founder of the site’s dedicated AI beat in 2022. He’s also a tech historian with almost two decades of experience. In his free time, he writes and records music, collects vintage computers, and enjoys nature. He lives in Raleigh, NC.

What is AGI? Nobody agrees, and it’s tearing Microsoft and OpenAI apart. Read More »

microsoft-is-trying-to-get-antivirus-software-away-from-the-windows-kernel

Microsoft is trying to get antivirus software away from the Windows kernel

Working with third-party companies to define these standards and address those companies’ concerns seems to be Microsoft’s way of trying to avoid that kind of controversy this time around.

“We will continue to collaborate deeply with our MVI partners throughout the private preview,” wrote Weston.

Death comes for the Blue Screen

Microsoft is changing the “b” in BSoD, but that’s less interesting than the under-the-hood changes. Credit: Microsoft

Microsoft’s post outlines a handful of other security-related Windows tweaks, including some that take alternate routes to preventing more Crowdstrike-esque outages.

Multiple changes are coming for the “unexpected restart screen,” the less-derogatory official name for what many Windows users know colloquially as the “blue screen of death.” For starters, the screen will now be black instead of blue, a change that Microsoft briefly attempted to make in the early days of Windows 11 but subsequently rolled back.

The unexpected restart screen has been “simplified” in a way that “improves readability and aligns better with Windows 11 design principles, while preserving the technical information on the screen for when it is needed.”

But the more meaningful change is under the hood, in the form of a new feature called “quick machine recovery” (QMR).

If a Windows PC has multiple unexpected restarts or gets into a boot loop—as happened to many systems affected by the Crowdstrike bug—the PC will try to boot into Windows RE, a stripped-down recovery environment that offers a handful of diagnostic options and can be used to enter Safe Mode or open the PC’s UEFI firmware. QMR will allow Microsoft to “broadly deploy targeted remediations to affected devices via Windows RE,” making it possible for some problems to be fixed even if the PCs can’t be booted into standard Windows, “quickly getting users to a productive state without requiring complex manual intervention from IT.”

QMR will be enabled by default on Windows 11 Home, while the Pro and Enterprise versions will be configurable by IT administrators. The QMR functionality and the black version of the blue screen of death will both be added to Windows 11 24H2 later this summer. Microsoft plans to add additional customization options for QMR “later this year.”

Microsoft is trying to get antivirus software away from the Windows kernel Read More »