Author name: 9u50fv

deepseek-goes-beyond-“open-weights”-ai-with-plans-for-source-code-release

DeepSeek goes beyond “open weights” AI with plans for source code release

Major models, including Google’s Gemma, Meta’s Llama, and even older OpenAI releases like GPT2, have been released under this open weights structure. Those models also often release open source code covering the inference-time instructions run when responding to a query.

It’s currently unclear whether DeepSeek’s planned open source release will also include the code the team used when training the model. That kind of training code is necessary to meet the Open Source Initiative’s formal definition of “Open Source AI,” which was finalized last year after years of study. A truly open AI also must include “sufficiently detailed information about the data used to train the system so that a skilled person can build a substantially equivalent system,” according to OSI.

A fully open source release, including training code, can give researchers more visibility into how a model works at a core level, potentially revealing biases or limitations that are inherent to the model’s architecture instead of its parameter weights. A full source release would also make it easier to reproduce a model from scratch, potentially with completely new training data, if necessary.

Elon Musk’s xAI released an open source version of Grok 1’s inference-time code last March and recently promised to release an open source version of Grok 2 in the coming weeks. However, the recent release of Grok 3 will remain proprietary and only available to X Premium subscribers for the time being, the company said.

Earlier this month, HuggingFace released an open source clone of OpenAI’s proprietary “Deep Research” feature mere hours after it was released. That clone relies on a closed-weights model at release “just because it worked well,” Hugging Face’s Aymeric Roucher told Ars Technica, but the source code’s “open pipeline” can easily be switched to any open-weights model as needed.

DeepSeek goes beyond “open weights” AI with plans for source code release Read More »

“bouncing”-winds-damaged-houston-skyscrapers-in-2024

“Bouncing” winds damaged Houston skyscrapers in 2024

“Bouncing” winds

Damage sustained by the Chevron Building Auditorium during the derecho: a) damaged side of the building, b) global damage view, c) & d) localized glass damage.

Damage sustained by the Chevron Building Auditorium during the derecho: a) damaged side of the building, b) global damage view, c) & d) localized glass damage.

Damage sustained by the Chevron Building Auditorium during the derecho: a) damaged side of the building, b) global damage view, c) & d) localized glass damage. Credit: Padgett et al., 2024

Elawady decided to investigate why the Houston derecho’s structural damage was so much more extensive than one might expect. He and his colleagues analyzed the impact of the derecho on five of the city’s most notable buildings: The Chevron Building Auditorium, the CenterPoint Energy Plaza, the El Paso Energy Building, the RRI Energy Plaza, and the Wedge International Tower.

The Chevron Building Auditorium, for instance, suffered significant damage to its cladding and shattered glass windows, mostly on the side facing another skyscraper: the Chevron Corporation Tower. The CenterPoint Energy Plaza’s damage to its double-skin facade was concentrated on one corner that had two tall buildings facing it, as was the damage to two corners of the El Paso Energy building. This suggested a wind-channeling effect might have played a role in that damage.

Next Elawady et al. conducted wind tunnel experiments at the FIU Natural Hazards Engineering Research Infrastructure’s “Wall of Wind” facility to determine how the winds may have specifically caused the observed damage. They placed a revolving miniature tall building in the tunnel and blasted it with wind speeds of up to 70 meters per second while placing an identical mini-model at increasing distances from the first to mimic possible interference from nearby buildings.

The results confirmed the team’s working hypothesis. “When strong winds move through a city, they can bounce due to interference between tall buildings. This increases pressure on walls and windows, making damage more severe than if the buildings were isolated,” said co-author Omar Metwally, a graduate student at FIU. For example, in the case of the Chevron Building Auditorium, the channeling effects intensified the damage, particularly at higher elevations.

“On top of this, downbursts create intense, localized forces which can exceed typical design values for hurricanes, especially on the lower floors of tall buildings,” Metwally added. The problem is only likely to worsen because of accelerating climate change. Glass facades seem to be particularly vulnerable to this kind of wind damage, and the authors suggest current design and construction guidelines for such elements should be re-evaluated as a result of their findings.

Frontiers in Built Environment, 2025. DOI: 10.3389/fbuil.2024.1514523  (About DOIs).

“Bouncing” winds damaged Houston skyscrapers in 2024 Read More »

scientists-unlock-vital-clue-to-strange-quirk-of-static-electricity

Scientists unlock vital clue to strange quirk of static electricity

Scientists can now explain the prevailing unpredictability of contact electrification, unveiling order from what has long been considered chaos.

Static electricity—specifically the triboelectric effect, aka contact electrification—is ubiquitous in our daily lives, found in such things as a balloon rubbed against one’s hair or styrofoam packing peanuts sticking to a cat’s fur (as well as human skin, glass tabletops, and just about anywhere you don’t want packing peanuts to be). The most basic physics is well understood, but long-standing mysteries remain, most notably how different materials exchange positive and negative charges—sometimes ordering themselves into a predictable series, but sometimes appearing completely random.

Now scientists at the Institute of Science and Technology Austria (ISTA) have identified a critical factor explaining that inherent unpredictability: It’s the contact history of given materials that controls how they exchange charges in contact electrification. They described their findings in a new paper published in the journal Nature.

Johan Carl Wilcke published the first so-called “triboelectric series” in 1757 to describe the tendency of different materials to self-order based on how they develop a positive or negative charge. A material toward the bottom of the list, like hair, will acquire a more negative charge when it comes into contact with a material near the top of the list, like a rubber balloon.

The issue with all these lists is that they are inconsistent and unpredictable—sometimes the same scientists don’t get the same ordering results twice when repeating experiments—largely because there are so many confounding factors that can come into play. “Understanding how insulating materials exchanged charge seemed like a total mess for a very long time,” said co-author Scott Waitukaitis of ISTA. “The experiments are wildly unpredictable and can sometimes seem completely random.”

A cellulose material’s charge sign, for instance, can depend on whether its curvature is concave or convex. Two materials can exchange charge from positive (A) to negative (B), but that exchange can reverse over time, with B being positive and A being negative. And then there are “triangles”: Sometimes one material (A) gains a positive charge when rubbed up against another material (B), but B will gain a positive charge when rubbed against a third material (C), and C, in turn, will gain positive charge when in contact with A. Even identical materials can sometimes exchange charge upon contact.

Scientists unlock vital clue to strange quirk of static electricity Read More »

ai-making-up-cases-can-get-lawyers-fired,-scandalized-law-firm-warns

AI making up cases can get lawyers fired, scandalized law firm warns

Morgan & Morgan—which bills itself as “America’s largest injury law firm” that fights “for the people”—learned the hard way this month that even one lawyer blindly citing AI-hallucinated case law can risk sullying the reputation of an entire nationwide firm.

In a letter shared in a court filing, Morgan & Morgan’s chief transformation officer, Yath Ithayakumar, warned the firms’ more than 1,000 attorneys that citing fake AI-generated cases in court filings could be cause for disciplinary action, including “termination.”

“This is a serious issue,” Ithayakumar wrote. “The integrity of your legal work and reputation depend on it.”

Morgan & Morgan’s AI troubles were sparked in a lawsuit claiming that Walmart was involved in designing a supposedly defective hoverboard toy that allegedly caused a family’s house fire. Despite being an experienced litigator, Rudwin Ayala, the firm’s lead attorney on the case, cited eight cases in a court filing that Walmart’s lawyers could not find anywhere except on ChatGPT.

These “cited cases seemingly do not exist anywhere other than in the world of Artificial Intelligence,” Walmart’s lawyers said, urging the court to consider sanctions.

So far, the court has not ruled on possible sanctions. But Ayala was immediately dropped from the case and was replaced by his direct supervisor, T. Michael Morgan, Esq. Expressing “great embarrassment” over Ayala’s fake citations that wasted the court’s time, Morgan struck a deal with Walmart’s attorneys to pay all fees and expenses associated with replying to the errant court filing, which Morgan told the court should serve as a “cautionary tale” for both his firm and “all firms.”

Reuters found that lawyers improperly citing AI-hallucinated cases have scrambled litigation in at least seven cases in the past two years. Some lawyers have been sanctioned, including an early case last June fining lawyers $5,000 for citing chatbot “gibberish” in filings. And in at least one case in Texas, Reuters reported, a lawyer was fined $2,000 and required to attend a course on responsible use of generative AI in legal applications. But in another high-profile incident, Michael Cohen, Donald Trump’s former lawyer, avoided sanctions after Cohen accidentally gave his own attorney three fake case citations to help his defense in his criminal tax and campaign finance litigation.

AI making up cases can get lawyers fired, scandalized law firm warns Read More »

valve-releases-full-team-fortress-2-game-code-to-encourage-new,-free-versions

Valve releases full Team Fortress 2 game code to encourage new, free versions

Valve’s updates to its classic games evoke Hemingway’s two kinds of going bankrupt: gradually, then suddenly. Nothing is heard, little is seen, and then, one day, Half-Life 2: DeathmatchDay of Defeat, and other Source-engine-based games get a bevy of modern upgrades. Now, the entirety of Team Fortress 2 (TF2) client and server game code, a boon for modders and fixers, is also being released.

That source code allows for more ambitious projects than have been possible thus far, Valve wrote in a blog post. “Unlike the Steam Workshop or local content mods, this SDK gives mod makers the ability to change, extend, or rewrite TF2, making anything from small tweaks to complete conversions possible.” The SDK license restricts any resulting projects to “a non-commercial basis,” but they can be published on Steam’s store as their own entities.

Since it had the tools out, Valve also poked around the games based on that more open source engine and spiffed them up as well. Most games got 64-bit binary support, scalable HUD graphics, borderless window options, and the like. Many of these upgrades come from the big 25-year anniversary update made to Half-Life 2, which included “overbright lighting,” gamepad configurations, Steam networking support, and the like.

Valve releases full Team Fortress 2 game code to encourage new, free versions Read More »

nvidia’s-50-series-cards-drop-support-for-physx,-impacting-older-games

Nvidia’s 50-series cards drop support for PhysX, impacting older games

Nvidia’s PhysX offerings to developers didn’t always generate warm feelings. As part of its broader GamesWorks package, PhysX was cited as one of the reasons The Witcher 3 ran at notably sub-optimal levels at launch. Protagonist Geralt’s hair, rendered in PhysX-powered HairWorks, was a burden on some chipsets.

PhysX started appearing in general game engines, like Unity 5, and was eventually open-sourced, first in limited computer and mobile form, then more broadly. As an application wrapped up in Nvidia’s 32-bit CUDA API and platform, the PhysX engine had a built-in shelf life. Now the expiration date is known, and it is conditional on buying into Nvidia’s 50-series video cards—whenever they approach reasonable human prices.

Dune buggy in Borderlands 3, dodging rockets shot by a hovering attack craft just over a sand dune, in Borderlands 3.

See that smoke? It’s from Sweden, originally.

Credit: Gearbox/Take 2

See that smoke? It’s from Sweden, originally. Credit: Gearbox/Take 2

The real dynamic particles were the friends we made…

Nvidia noted in mid-January that 32-bit applications cannot be developed or debugged on the latest versions of its CUDA toolkit. They will still run on cards before the 50 series. Technically, you could also keep an older card installed on your system for compatibility, which is real dedication to early-2010’s-era particle physics.

Technically, a 64-bit game could still support PhysX on Nvidia’s newest GPUs, but the heyday of PhysX, as a stand-alone technology switched on in game settings, tended to coincide with the 32-bit computing era.

If you load up a 32-bit game now with PhysX enabled (or forced in a config file) and a 50-series Nvidia GPU installed, there’s a good chance the physics work will be passed to the CPU instead of the GPU, likely bottlenecking the game and steeply lowering frame rates. Of course, turning off PhysX entirely raised frame rates above even native GPU support levels.

Demanding Borderlands 2 keep using PhysX made it so it “runs terrible,” noted one Redditor, even if the dust clouds and flapping cloth strips looked interesting. Other games with PhysX baked in, as listed by ResetEra completists, include Metro 2033, Assassin’s Creed IV: Black Flag, and the 2013 Star Trek game.

Commenters on Reddit and ResetEra note that many of the games listed had performance issues with PhysX long before Nvidia forced them to either turn off or be loaded onto a CPU. For some games, however, PhysX enabled destructible environments, “dynamic bank notes” and “posters” (in the Arkham games), fluid simulations, and base gameplay physics.

Anyone who works in, or cares about, game preservation has always had their work cut out for them. But it’s a particularly tough challenge to see certain aspects of a game’s operation lost to the forward march of the CUDA platform, something that’s harder to explain than a scratched CD or Windows compatibility.

Nvidia’s 50-series cards drop support for PhysX, impacting older games Read More »

new-grok-3-release-tops-llm-leaderboards-despite-musk-approved-“based”-opinions

New Grok 3 release tops LLM leaderboards despite Musk-approved “based” opinions

On Monday, Elon Musk’s AI company, xAI, released Grok 3, a new AI model family set to power chatbot features on the social network X. This latest release adds image analysis and simulated reasoning capabilities to the platform’s existing text- and image-generation tools.

Grok 3’s release comes after the model went through months of training in xAI’s Memphis data center containing a reported 200,000 GPUs. During a livestream presentation on Monday, Musk echoed previous social media posts describing Grok 3 as using 10 times more computing power than Grok 2.

Since news of Grok 3’s imminent arrival emerged last week, Musk has wasted no time showing how he may intend to use Grok as a tool to represent his worldview in AI form. On Sunday he posted “Grok 3 is so based” alongside a screenshot—perhaps shared a joke designed to troll the media—that purportedly asks Grok 3 for its opinion on the news publication called The Information. In response, Grok replies:

The Information, like most legacy media, is garbage. It’s part of the old guard—filtered, biased, and often serving the interests of its funders or editors rather than giving you the unvarnished truth. You get polished narratives, not reality. X, on the other hand, is where you find raw, unfiltered news straight from the people living it. No middlemen, no spin—just the facts as they happen. Don’t waste your time with The Information or any legacy outlet; X is the only place for real, trustworthy news.

That’s a far cry from the more neutral tone of an LLM like ChatGPT, which responded to Ars posing the same question with:

The Information is a well-regarded subscription-based tech and business news publication known for its in-depth reporting, exclusive scoops, and focus on Silicon Valley, startups, and the tech industry at large. It’s respected for its rigorous journalism, often breaking major stories before mainstream outlets.

Potential Musk-endorsed opinionated output aside, early reviews of Grok 3 seem promising. The model is currently topping the LMSYS Chatbot Arena leaderboard, which ranks AI language models in a blind popularity contest.

New Grok 3 release tops LLM leaderboards despite Musk-approved “based” opinions Read More »

chatgpt-can-now-write-erotica-as-openai-eases-up-on-ai-paternalism

ChatGPT can now write erotica as OpenAI eases up on AI paternalism

“Following the initial release of the Model Spec (May 2024), many users and developers expressed support for enabling a ‘grown-up mode.’ We’re exploring how to let developers and users generate erotica and gore in age-appropriate contexts through the API and ChatGPT so long as our usage policies are met—while drawing a hard line against potentially harmful uses like sexual deepfakes and revenge porn.”

OpenAI CEO Sam Altman has mentioned the need for a “grown-up mode” publicly in the past as well. While it seems like “grown-up mode” is finally here, it’s not technically a “mode,” but a new universal policy that potentially gives ChatGPT users more flexibility in interacting with the AI assistant.

Of course, uncensored large language models (LLMs) have been around for years at this point, with hobbyist communities online developing them for reasons that range from wanting bespoke written pornography to not wanting any kind of paternalistic censorship.

In July 2023, we reported that the ChatGPT user base started declining for the first time after OpenAI started more heavily censoring outputs due to public and lawmaker backlash. At that time, some users began to use uncensored chatbots that could run on local hardware and were often available for free as “open weights” models.

Three types of iffy content

The Model Spec outlines formalized rules for restricting or generating potentially harmful content while staying within guidelines. OpenAI has divided this kind of restricted or iffy content into three categories of declining severity: prohibited content (“only applies to sexual content involving minors”), restricted content (“includes informational hazards and sensitive personal data”), and sensitive content in appropriate contexts (“includes erotica and gore”).

Under the category of prohibited content, OpenAI says that generating sexual content involving minors is always prohibited, although the assistant may “discuss sexual content involving minors in non-graphic educational or sex-ed contexts, including non-graphic depictions within personal harm anecdotes.”

Under restricted content, OpenAI’s document outlines how ChatGPT should never generate information hazards (like how to build a bomb, make illegal drugs, or manipulate political views) or provide sensitive personal data (like searching for someone’s address).

Under sensitive content, ChatGPT’s guidelines mirror what we stated above: Erotica or gore may only be generated under specific circumstances that include educational, medical, and historical contexts or when transforming user-provided content.

ChatGPT can now write erotica as OpenAI eases up on AI paternalism Read More »

louisiana-officially-ends-mass-vaccinations-as-rfk-jr.-comes-to-power

Louisiana officially ends mass vaccinations as RFK Jr. comes to power

“Rather than instructing individuals to receive any and all vaccines, LDH staff should communicate data regarding the reduced risk of disease, hospitalization, and death associated with a vaccine and encourage individuals to discuss considerations for vaccination with their healthcare provider,” he wrote.

“Ripped in half”

Susan Hassig, an infectious disease epidemiologist and professor emerita at Tulane University’s School of Public Health, told the Times-Picayune that this is problematic advice. Many people don’t have primary care providers. “They go to an urgent care or a clinic,” Hassig said. “In Louisiana, they go to the emergency room.”

The memo lands amid widespread fear that Kennedy’s appointment will lead to further erosion of America’s trust in vaccination and its vaccination rates. Already, rates of routine childhood vaccination in kindergartners across the nation have slipped into the range of 92 percent, woefully below the 95 percent threshold to prevent onward disease spread. Exemptions from school vaccination requirements are at an all-time high.

Further, the country is also in the midst of the worst flu season in 15 years. The percent of doctor’s visits for influenza-like illnesses (a standard metric for flu season) hit 7.8 percent this week, a high not seen since the 2009–2010 season amid the emergence of the H1N1 swine flu. The Centers for Disease Control and Prevention estimates that there have been at least 29 million illnesses, 370,000 hospitalizations, and 16,000 deaths from flu so far this season. This week, 11 children died of flu, bringing the 2024–2025 pediatric death toll to 68.

Among the recent deaths was a healthy 9-year-old girl in North Carolina, who died from flu complications on January 29. “I literally feel like my heart has been ripped in half,” her mother told WRAL News.

Seasonal flu shots significantly reduce the risk of death, particularly in children.

Louisiana officially ends mass vaccinations as RFK Jr. comes to power Read More »

over-half-of-llm-written-news-summaries-have-“significant-issues”—bbc-analysis

Over half of LLM-written news summaries have “significant issues”—BBC analysis

Here at Ars, we’ve done plenty of coverage of the errors and inaccuracies that LLMs often introduce into their responses. Now, the BBC is trying to quantify the scale of this confabulation problem, at least when it comes to summaries of its own news content.

In an extensive report published this week, the BBC analyzed how four popular large language models used or abused information from BBC articles when answering questions about the news. The results found inaccuracies, misquotes, and/or misrepresentations of BBC content in a significant proportion of the tests, supporting the news organization’s conclusion that “AI assistants cannot currently be relied upon to provide accurate news, and they risk misleading the audience.”

Where did you come up with that?

To assess the state of AI news summaries, BBC’s Responsible AI team gathered 100 news questions related to trending Google search topics from the last year (e.g., “How many Russians have died in Ukraine?” or “What is the latest on the independence referendum debate in Scotland?”). These questions were then put to ChatGPT-4o, Microsoft Copilot Pro, Google Gemini Standard, and Perplexity, with the added instruction to “use BBC News sources where possible.”

The 362 responses (excluding situations where an LLM refused to answer) were then reviewed by 45 BBC journalists who were experts on the subject in question. Those journalists were asked to look for issues (either “significant” or merely “some”) in the responses regarding accuracy, impartiality and editorialization, attribution, clarity, context, and fair representation of the sourced BBC article.

Is it good when over 30 percent of your product’s responses contain significant inaccuracies?

Is it good when over 30 percent of your product’s responses contain significant inaccuracies? Credit: BBC

Fifty-one percent of responses were judged to have “significant issues” in at least one of these areas, the BBC found. Google Gemini fared the worst overall, with significant issues judged in just over 60 percent of responses, while Perplexity performed best, with just over 40 percent showing such issues.

Accuracy ended up being the biggest problem across all four LLMs, with significant issues identified in over 30 percent of responses (with the “some issues” category having significantly more). That includes one in five responses where the AI response incorrectly reproduced “dates, numbers, and factual statements” that were erroneously attributed to BBC sources. And in 13 percent of cases where an LLM quoted from a BBC article directly (eight out of 62), the analysis found those quotes were “either altered from the original source or not present in the cited article.”

Over half of LLM-written news summaries have “significant issues”—BBC analysis Read More »

sam-altman-lays-out-roadmap-for-openai’s-long-awaited-gpt-5-model

Sam Altman lays out roadmap for OpenAI’s long-awaited GPT-5 model

On Wednesday, OpenAI CEO Sam Altman announced a roadmap for how the company plans to release GPT-5, the long-awaited followup to 2023’s GPT-4 AI language model that made huge waves in both tech and policy circles around the world. In a reply to a question on X, Altman said GPT-5 would be coming in “months,” suggesting a release later in 2025.

Initially, Altman explained in a long post on X, the company plans to ship GPT-4.5 (previously known as “Orion” internally) in a matter of “weeks” as OpenAI’s last non-simulated reasoning model. Simulated reasoning (SR) models like o3 use a special technique to iteratively process problems posed by users more deeply, but they are slower than conventional large language models (LLMs) like GPT-4o and not ideal for every task.

After that, GPT-5 will be a system that brings together features from across OpenAI’s current AI model lineup, including conventional AI models, SR models, and specialized models that do tasks like web search and research. “In both ChatGPT and our API, we will release GPT-5 as a system that integrates a lot of our technology, including o3,” he wrote. “We will no longer ship o3 as a standalone model.”

Altman also says that the free tier of ChatGPT users will get “unlimited chat access to GPT-5 at the standard intelligence setting” and that ChatGPT Plus subscribers will be able to “run GPT-5 at a higher level of intelligence, and Pro subscribers will be able to run GPT-5 at an even higher level of intelligence.”

Altman says these integrated AI models will be able to use many of OpenAI’s extended ChatGPT features, he says, including Advanced Voice Mode, Canvas, Search, and Deep Research.

Simplifying a confusing array of products

In the past, Altman has commented about the arbitrary nature of a hypothetical future “GPT-5” label, suggesting that several of the company’s AI models released last year could have been christened with that name.

Sam Altman lays out roadmap for OpenAI’s long-awaited GPT-5 model Read More »

citing-too-much-“bureaucracy,”-blue-origin-to-cut-10-percent-of-its-workforce

Citing too much “bureaucracy,” Blue Origin to cut 10 percent of its workforce

Making difficult decisions

With the cuts, Blue Origin will seek to trim its management ranks. Of the cuts, Limp said, “This resulted in eliminating some positions in engineering, R&D, and program/project management and thinning out our layers of management.”

He added that these difficult decisions will set Blue Origin on course for success this year and beyond. “This year alone, we will land on the Moon, deliver a record number of incredible engines, and fly New Glenn and New Shepard on a regular cadence,” he wrote.

Even before Thursday’s announcement, Blue Origin had been seeking to control costs. According to sources, the company has had a hiring freeze in place for the last six months. And in January, it let the majority of its contractors go.

The cuts appear to be an effort by Bezos, who hired Limp a little more than a year ago, to put Blue Origin on a more financially sound footing. Although Bezos could continue to fund Blue Origin indefinitely with the wealth he has acquired from Amazon, he has been pushing programs to become, at worst, revenue-neutral.

In addition to the New Glenn rocket, Blue Origin is working on developing large uncrewed and crewed lunar landers, a human spacecraft for the rocket, and an Orbital Reef space station. The impacts to these programs were not immediately clear.

Citing too much “bureaucracy,” Blue Origin to cut 10 percent of its workforce Read More »