AI

deepmind-ai-rivals-the-world’s-smartest-high-schoolers-at-geometry

DeepMind AI rivals the world’s smartest high schoolers at geometry

Demis Hassabis, CEO of DeepMind Technologies and developer of AlphaGO, attends the AI Safety Summit at Bletchley Park on November 2, 2023 in Bletchley, England.

Enlarge / Demis Hassabis, CEO of DeepMind Technologies and developer of AlphaGO, attends the AI Safety Summit at Bletchley Park on November 2, 2023 in Bletchley, England.

A system developed by Google’s DeepMind has set a new record for AI performance on geometry problems. DeepMind’s AlphaGeometry managed to solve 25 of the 30 geometry problems drawn from the International Mathematical Olympiad between 2000 and 2022.

That puts the software ahead of the vast majority of young mathematicians and just shy of IMO gold medalists. DeepMind estimates that the average gold medalist would have solved 26 out of 30 problems. Many view the IMO as the world’s most prestigious math competition for high school students.

“Because language models excel at identifying general patterns and relationships in data, they can quickly predict potentially useful constructs, but often lack the ability to reason rigorously or explain their decisions,” DeepMind writes. To overcome this difficulty, DeepMind paired a language model with a more traditional symbolic deduction engine that performs algebraic and geometric reasoning.

The research was led by Trieu Trinh, a computer scientist who recently earned his PhD from New York University. He was a resident at DeepMind between 2021 and 2023.

Evan Chen, a former Olympiad gold medalist who evaluated some of AlphaGeometry’s output, praised it as “impressive because it’s both verifiable and clean.” Whereas some earlier software generated complex geometry proofs that were hard for human reviewers to understand, the output of AlphaGeometry is similar to what a human mathematician would write.

AlphaGeometry is part of DeepMind’s larger project to improve the reasoning capabilities of large language models by combining them with traditional search algorithms. DeepMind has published several papers in this area over the last year.

How AlphaGeometry works

Let’s start with a simple example shown in the AlphaGeometry paper, which was published by Nature on Wednesday:

The goal is to prove that if a triangle has two equal sides (AB and AC), then the angles opposite those sides will also be equal. We can do this by creating a new point D at the midpoint of the third side of the triangle (BC). It’s easy to show that all three sides of triangle ABD are the same length as the corresponding sides of triangle ACD. And two triangles with equal sides always have equal angles.

Geometry problems from the IMO are much more complex than this toy problem, but fundamentally, they have the same structure. They all start with a geometric figure and some facts about the figure like “side AB is the same length as side AC.” The goal is to generate a sequence of valid inferences that conclude with a given statement like “angle ABC is equal to angle BCA.”

For many years, we’ve had software that can generate lists of valid conclusions that can be drawn from a set of starting assumptions. Simple geometry problems can be solved by “brute force”: mechanically listing every possible fact that can be inferred from the given assumption, then listing every possible inference from those facts, and so on until you reach the desired conclusion.

But this kind of brute-force search isn’t feasible for an IMO-level geometry problem because the search space is too large. Not only do harder problems require longer proofs, but sophisticated proofs often require the introduction of new elements to the initial figure—as with point D in the above proof. Once you allow for these kinds of “auxiliary points,” the space of possible proofs explodes and brute-force methods become impractical.

DeepMind AI rivals the world’s smartest high schoolers at geometry Read More »

wordpad-out;-80gbps-usb-support-and-other-win-11-features-in-testing-this-month

WordPad out; 80Gbps USB support and other Win 11 features in testing this month

Can’t stop won’t stop —

Microsoft’s next batch of Windows 11 feature updates is taking shape.

Green USB-C cable

Windows 11’s big feature update in September included a long list of minor changes, plus the Copilot AI assistant; that update was followed by Windows 11 23H2 in late October, which reset the operating system’s timeline for technical support and security updates but didn’t add much else in and of itself. But Windows development never stops these days, and this month’s Insider Preview builds have already shown us a few things that could end up in the stable version of the operating system in the next couple of months.

One major addition, which rolled out to Dev Channel builds on January 11 and Beta Channel builds today, is support for 80Gbps USB 4 ports. These speeds are part of the USB4 Version 2.0 spec—named with the USB-IF’s typical flair for clarity and consistency—that was published in 2022. Full 80Gbps speeds are still rare and will be for the foreseeable future, but Microsoft says that they’ll be included the Razer Blade 18 and a handful of other PCs with Intel’s 14th-generation HX-series laptop processors. We’d expect the new speeds to proliferate slowly and mostly in high-end systems over the next few months and years.

Another addition to that January 11 Dev Channel build is a change in how the Copilot generative AI assistant works. Normally, Copilot is launched by the user manually, either by clicking the icon on the taskbar, hitting the Win+C key combo, or (in some new PCs) by using the dedicated Copilot button on the keyboard. In recent Dev Channel builds, the Copilot window will open automatically on certain PCs as soon as you log into Windows, becoming part of your default desktop unless you turn it off in Settings.

The Copilot panel will only open by default on screens that meet minimum size and resolution requirements, things that Windows already detects and takes into account when setting your PC’s default zoom and showing available Snap Layouts, among other things. Microsoft says it’s testing the feature on screens that are 27 inches or larger with 1,920 or more horizontal pixels (for most screens, this means a minimum resolution of 1080p). For PCs without Copilot, including those that haven’t been signed into a Microsoft account, the feature will continue to be absent.

The

Enlarge / The “richer weather experience on the Lock screen,” seen in the bottom-center of this screenshot.

Microsoft

Other additions to the Dev Channel builds this month include easy Snipping Tool editing for Android screenshots from phones that have been paired to your PC, custom user-created voice commands, the ability to share URLs directly to services like WhatsApp and Gmail from the Windows share window, a new Weather widget for the Windows lock screen, and app install notifications from the Microsoft store.

Microsoft hasn’t publicized any of the changes it has made to its Canary channel builds since January 4—this is typical since it changes the fastest, and the tested features are the most likely to be removed or significantly tweaked before being released to the public. Most of the significant additions from that announcement have since made it out to the other channels, but there are a couple of things worth noting. First, there’s a new Energy Saver taskbar icon for desktop PCs without batteries, making it easier to tell when the feature is on without creating confusion. And the venerable WordPad app, originally marked for deletion in September, has also been removed from these builds and can’t be reinstalled.

Microsoft doesn’t publish Windows feature updates on an exact cadence beyond its commitment to deliver one with a new version number once per year in the fall. Last year’s first major batch of Windows 11 additions rolled out at the end of February, so a late winter or early spring launch window for the next batch of features could make sense.

WordPad out; 80Gbps USB support and other Win 11 features in testing this month Read More »

“alexa-is-in-trouble”:-paid-for-alexa-gives-inaccurate-answers-in-early-demos

“Alexa is in trouble”: Paid-for Alexa gives inaccurate answers in early demos

Amazon Echo Show 8 with Alexa

Enlarge / Amazon demoed future generative AI capabilties for Alexa in September.

“If this fails to get revenue, Alexa is in trouble.”

A quote from an anonymous Amazon employee in a Wednesday Business Insider report paints a dire picture. Amazon needs its upcoming subscription version of Alexa to drive revenue in ways that its voice assistant never has before.

Amazon declined Ars’ request for comment on the report. But the opening quote in this article could have been uttered by anyone following voice assistants for the past year-plus. All voice assistants have struggled to drive revenue since people tend to use voice assistants for basic queries, like checking the weather, rather than transactions.

Amazon announced plans to drive usage and interest in Alexa by releasing a generative AI version that it said would one day require a subscription.

This leads to the question: Would you pay to use Alexa? Amazon will be challenged to convince people to change how they use Alexa while suddenly paying a monthly rate to enable that unprecedented behavior.

Workers within Amazon seemingly see this obstacle. Insider, citing an anonymous Amazon employee, reported that “some were questioning the entire premise of charging for Alexa. For example, people who already pay for an existing Amazon service, such as Amazon Music, might not be willing to pay additional money to get access to the newer version of Alexa.”

“There is tension over whether people will pay for Alexa or not,” one of the anonymous Amazon workers reportedly said.

Subscription-based Alexa originally planned for June release

Amazon hasn’t publicly confirmed a release date for generative AI Alexa. But Insider’s report, citing “internal documents and people familiar with the matter,” said Amazon has been planning to release its subscription plan on June 30. However, plans for what Insider said will be called “Alexa Plus” and built on “Remarkable Alexa” technology could be delayed due to numerous development challenges.

According to the report, the Remarkable Alexa tech has been being demoed by 15,000 customers and currently succeeds in being conversational but is “deflecting answers, often giving unnecessarily long or inaccurate responses.”

In September, then-SVP of devices and services at Amazon David Limp demoed Alexa understanding more complex commands, including Alexa not requiring the “Hey Alexa” prompt and being able to understand multiple demands for multiple apps through a single spoken phrase.

Insider reported: “The new Alexa still didn’t meet the quality standards expected for Alexa Plus, these people added, noting the technical challenges and complexity of redesigning Alexa.”

“Legacy constraints”

According to the report, people working on the original Alexa insisted on using what they had already built for the standard voice assistant with the paid-for version, resulting in a bloated technology and “internal politics.”

However, the original Alexa is based on a natural language model with multiple parts doing multiple things, compared to the colossal large language model of generative AI Alexa.

Now, generative AI Alexa is reportedly moving to a new technological stack to avoid the “legacy constraints” of today’s Alexa but potentially delaying things.

“Alexa is in trouble”: Paid-for Alexa gives inaccurate answers in early demos Read More »

samsung’s-$1,300-phone-might-someday-have-fees-for-ai-usage

Samsung’s $1,300 phone might someday have fees for AI usage

Will Samsung even care about AI in 2026? —

Samsung says Galaxy S24 AI features are “free until the end of 2025.”

Samsung’s $1,300 phone might someday have fees for AI usage

Samsung

Samsung’s big Galaxy S24 launch was yesterday, and to hear Samsung tell the story, the big highlight of the event was “Galaxy AI.” Another view is that Galaxy AI is the usual bundle of baked-in Samsung features skinned on top of Android, but with generative AI being the hot new thing, Samsung went with AI-centric branding. Whatever value you want to place on Samsung’s AI features, you might soon have to place an actual monetary value on them: Despite devices like the Galaxy S24 Ultra costing $1,300, Samsung might start charging for some of these AI phone features.

The fine print on Samsung’s Galaxy S24 promotional page features 44 asterisks and footnotes, and tucked away in that pile of caveats is the line “Galaxy AI features will be provided for free until the end of 2025 on supported Samsung Galaxy devices.” That means Samsung reserves the right to charge for Galaxy AI after 2025.

AI features that require server time have an ongoing cost. Google and Amazon figured this out in the last AI generation (if we can call it that) with the Google Assistant and Alexa voice assistants. Amazon’s finances on the whole situation are clearer than Google’s, and Amazon’s 2022 Alexa financials were reportedly a $10 billion loss. Amazon is planning on a subscription model for Alexa in the future. Google’s normal user subscription plan is Google One, and while that mainly gets you more account storage, it also unlocks some Google AI features like “Magic eraser” in Google Photos. ChatGPT has a subscription plan for its best model, ChatGPT 4, too. Samsung apparently wants to join the party.

The Galaxy S24's

Enlarge / The Galaxy S24’s “Live translate” feature in the phone app. You can speak one language, and the phone app will repeat your message in a different language after a delay.

Samsung

This is the company that makes Bixby and the notoriously poorly coded Tizen, though, so it’s hard to imagine Galaxy AI features being worth paying for. The first item on Samsung’s “Galaxy AI” promo page is Google’s “Circle to search,” a feature it can’t charge for and didn’t build. The Galaxy AI features made by Samsung include “Interpreter,” which is a copy of Google Translate’s conversation mode, and Voice Recorder, a voice transcription app that is just a copy of Google Recorder (and apparently not as good). “Chat Assist” is part of the keyboard and can rewrite any inputted text with generative AI, making your input sound more “fun” or “professional.” “Note Assist” is a Samsung Notes feature that can generate AI summaries of your notes. The one interesting feature is “Live Translate,” which does voice translation of a phone call, translating communication via speech-to-text-to-speech. There’s a lot that can go wrong there, though.

Samsung is a hardware company, and presumably, a lot of these use on-device processing instead of bothering a server somewhere, so it’s hard to know if Samsung even has any serious costs to recoup. Like most Samsung Android features, this feels more like throwing a pile of stuff at the wall and hoping something sticks rather than a collection of killer apps. These are essentially all just app features, too, meaning they have to compete with the nearly infinite Play Store app selection, and you could easily download a free competitor.

The first step to charging for something like this is throwing the idea out there, so Samsung is probably listening to how people will react between now and the end of 2025.

Samsung’s $1,300 phone might someday have fees for AI usage Read More »

bing-search-shows-few,-if-any,-signs-of-market-share-increase-from-ai-features

Bing Search shows few, if any, signs of market share increase from AI features

high hopes —

Bing’s US and worldwide market share is about the same as it has been for years.

Bing Search shows few, if any, signs of market share increase from AI features

Microsoft

Not quite one year ago, Microsoft announced a “multi-year, multi-billion dollar investment” in OpenAI, a company that had made waves in 2022 with its ChatGPT chatbot and DALL-E image creator. The next month, Microsoft announced that it was integrating a generative AI chatbot into its Bing search engine and Edge browser, and similar generative AI features were announced for Windows in the apps formerly known as Microsoft Office, Microsoft Teams, and other products.

Adding AI features to Bing was meant to give it an edge over Google, and reports indicated that Google was worried enough about it to accelerate its own internal generative AI efforts. Microsoft announced in March 2023 that Bing surpassed the 100 million monthly active users mark based on interest in Bing Chat and its ilk; by Microsoft’s estimates, each percentage of Google’s search market share that Bing could siphon away was worth as much as $2 billion to Microsoft.

A year later, it looks like Microsoft’s AI efforts may have helped Bing on the margins, but they haven’t meaningfully eroded Google’s search market share, according to Bloomberg. Per Bloomberg’s analysis of data from Sensor Tower, Bing usage had been down around 33 percent year over year just before the AI-powered features were added, but those numbers had rebounded by the middle of 2023.

Microsoft hasn’t given an official update on Bing’s monthly active users in quite a while—we’ve asked the company for an update, and will share it if we get one—though Microsoft Chief Marketing Officer Yusuf Medhi told Bloomberg that “millions and millions of people” were still using the new AI features.

StatCounter data mostly tells a similar story. According to its data, Google’s worldwide market share is currently in the low 90s, and it has been for virtually the entire 15-year period for which StatCounter offers data. Bing’s worldwide market share number over the same period has been remarkably stable; it was about 3.5 percent in the summer of 2009, when what had been known as Live Search was renamed Bing in the first place, and as of December 2023, it was still stuck at around 3.4 percent.

Recent US data is slightly more flattering for Microsoft, where Bing’s usage rose from 6.7 percent in December 2022 to 7.7 percent in December 2023. But that doesn’t necessarily suggest any kind of AI-fueled influx in new Bing search users—usage remained in the mid-to-high 6 percent range through most of 2023 before ticking up right at the end of the year—and Bing’s US usage has floated in that same 6–7 percent zone for most of the last decade.

It even seems like Microsoft is making moves to distance its AI efforts from Bing a bit. What began as “Bing Chat” or “the new Bing” is now known as Windows Copilot—both inside Windows 11 and elsewhere. Earlier this week, the Bing Image Creator became “Image Creator from Designer.” Both products still feature Bing branding prominently—the Copilot screen in Windows 11 still says “with Bing” at the top of it, and the Image Creator tool is still hosted on the Bing.com domain. But if these new AI features aren’t driving Bing’s market share up, then it makes sense for Microsoft to create room for them to stand on their own.

That’s not to say Google’s search dominance is assured. Leipzig University researchers published a study earlier this week (PDF) suggesting Google, Bing, and the Bing-powered DuckDuckGo had seen “an overall downward trend in text quality,” especially for heavily SEO-optimized categories like purchase recommendations and product reviews.

Bing Search shows few, if any, signs of market share increase from AI features Read More »

game-developer-survey:-50%-work-at-a-studio-already-using-generative-ai-tools

Game developer survey: 50% work at a studio already using generative AI tools

Do androids dream of Tetris? —

But 84% of devs are at least somewhat concerned about ethical use of those tools.

The future of game development?

Enlarge / The future of game development?

A new survey of thousands of game development professionals finds a near-majority saying generative AI tools are already in use at their workplace. But a significant minority of developers say their company has no interest in generative AI tools or has outright banned their use.

The Game Developers Conference’s 2024 State of the Industry report, released Thursday, aggregates the thoughts of over 3,000 industry professionals as of last October. While the annual survey (conducted in conjunction with research partner Omdia) has been running for 12 years, this is the first time respondents were asked directly about their use of generative AI tools such as ChatGPT, DALL-E, GitHub Copilot, and Adobe Generative Fill.

Forty-nine percent of the survey’s developer respondents said that generative AI tools are currently being used in their workplace. That near-majority includes 31 percent (of all respondents) that say they use those tools themselves and 18 percent that say their colleagues do.

A majority of game developers said their workplace was at least interested in using generative AI tools.

Enlarge / A majority of game developers said their workplace was at least interested in using generative AI tools.

The survey also found that different studio departments showed different levels of willingness to embrace AI tools. Forty-four percent of employees in business and finance said they were using AI tools, for instance, compared to just 16 percent in visual arts and 13 percent in “narrative/writing.”

Among the 38 percent of respondents who said their company didn’t use AI tools, 15 percent said their company was “interested” in pursuing them, while 23 percent said they had “no interest.” In a separate question, 12 percent of respondents said their company didn’t allow the use of AI tools at all, a number that went up to 21 percent for respondents working at the largest “AAA developers.” An additional 7 percent said the use of some specific AI tools was not allowed, while 30 percent said AI tool use was “optional” at their company.

Worries abound

The wide embrace of AI tools hasn’t seemed to lessen worries about their use among developers, though. A full 42 percent of respondents said they were “very concerned” about the ethics of using generative AI in game development, with an additional 42 percent being “somewhat concerned.” Only 12 percent said they were “not concerned at all” about those usage ethics.

Developer policies on AI use varied greatly, with a plurality saying their company had no official policy.

Enlarge / Developer policies on AI use varied greatly, with a plurality saying their company had no official policy.

Overall, respondents offered a split opinion on whether the use of AI tools would be overall positive (21 percent) or negative (18 percent) for the industry. Most respondents seemed split, with 57 percent saying the impact would be “mixed.”

Developers cited coding assistance, content creation efficiency, and the automation of repetitive tasks as the primary uses for AI tools, according to the report.

“I’d like to see AI tools that help with the current workflows and empower individual artists with their own work,” one anonymous respondent wrote. “What I don’t want to see is a conglomerate of artists being enveloped in an AI that just does 99% of the work a creative is supposed to do.”

Elsewhere in the report, the survey found that only 17 percent of developers were at least somewhat interested in using blockchain technology in their upcoming projects, down significantly from 27 percent in 2022. An overwhelming 77 percent of respondents said they had no interest in blockchain technology, similar to recent years.

The survey also found that 57 percent of respondents thought that workers in the game industry should unionize, up from 53 percent last year. Despite this, only 23 percent said they were either in a union or had discussed unionization at their workplace.

Game developer survey: 50% work at a studio already using generative AI tools Read More »

openai-opens-the-door-for-military-uses-but-maintains-ai-weapons-ban

OpenAI opens the door for military uses but maintains AI weapons ban

Skynet deferred —

Despite new Pentagon collab, OpenAI won’t allow customers to “develop or use weapons” with its tools.

The OpenAI logo over a camoflage background.

On Tuesday, ChatGPT developer OpenAI revealed that it is collaborating with the United States Defense Department on cybersecurity projects and exploring ways to prevent veteran suicide, reports Bloomberg. OpenAI revealed the collaboration during an interview with the news outlet at the World Economic Forum in Davos. The AI company recently modified its policies, allowing for certain military applications of its technology, while maintaining prohibitions against using it to develop weapons.

According to Anna Makanju, OpenAI’s vice president of global affairs, “many people thought that [a previous blanket prohibition on military applications] would prohibit many of these use cases, which people think are very much aligned with what we want to see in the world.” OpenAI removed terms from its service agreement that previously blocked AI use in “military and warfare” situations, but the company still upholds a ban on its technology being used to develop weapons or to cause harm or property damage.

Under the “Universal Policies” section of OpenAI’s Usage Policies document, section 2 says, “Don’t use our service to harm yourself or others.” The prohibition includes using its AI products to “develop or use weapons.” Changes to the terms that removed the “military and warfare” prohibitions appear to have been made by OpenAI on January 10.

The shift in policy appears to align OpenAI more closely with the needs of various governmental departments, including the possibility of preventing veteran suicides. “We’ve been doing work with the Department of Defense on cybersecurity tools for open-source software that secures critical infrastructure,” Makanju said in the interview. “We’ve been exploring whether it can assist with (prevention of) veteran suicide.”

The efforts mark a significant change from OpenAI’s original stance on military partnerships, Bloomberg says. Meanwhile, Microsoft Corp., a large investor in OpenAI, already has an established relationship with the US military through various software contracts.

OpenAI opens the door for military uses but maintains AI weapons ban Read More »

just-10-lines-of-code-can-steal-ai-secrets-from-apple,-amd,-and-qualcomm-gpus

Just 10 lines of code can steal AI secrets from Apple, AMD, and Qualcomm GPUs

massive leakage —

Patching all affected devices, which include some Macs and iPhones, may be tough.

ai brain

MEHAU KULYK/Getty Images

As more companies ramp up development of artificial intelligence systems, they are increasingly turning to graphics processing unit (GPU) chips for the computing power they need to run large language models (LLMs) and to crunch data quickly at massive scale. Between video game processing and AI, demand for GPUs has never been higher, and chipmakers are rushing to bolster supply. In new findings released today, though, researchers are highlighting a vulnerability in multiple brands and models of mainstream GPUs—including Apple, Qualcomm, and AMD chips—that could allow an attacker to steal large quantities of data from a GPU’s memory.

The silicon industry has spent years refining the security of central processing units, or CPUs, so they don’t leak data in memory even when they are built to optimize for speed. However, since GPUs were designed for raw graphics processing power, they haven’t been architected to the same degree with data privacy as a priority. As generative AI and other machine learning applications expand the uses of these chips, though, researchers from New York-based security firm Trail of Bits say that vulnerabilities in GPUs are an increasingly urgent concern.

“There is a broader security concern about these GPUs not being as secure as they should be and leaking a significant amount of data,” Heidy Khlaaf, Trail of Bits’ engineering director for AI and machine learning assurance, tells WIRED. “We’re looking at anywhere from 5 megabytes to 180 megabytes. In the CPU world, even a bit is too much to reveal.”

To exploit the vulnerability, which the researchers call LeftoverLocals, attackers would need to already have established some amount of operating system access on a target’s device. Modern computers and servers are specifically designed to silo data so multiple users can share the same processing resources without being able to access each others’ data. But a LeftoverLocals attack breaks down these walls. Exploiting the vulnerability would allow a hacker to exfiltrate data they shouldn’t be able to access from the local memory of vulnerable GPUs, exposing whatever data happens to be there for the taking, which could include queries and responses generated by LLMs as well as the weights driving the response.

In their proof of concept, as seen in the GIF below, the researchers demonstrate an attack where a target—shown on the left—asks the open source LLM Llama.cpp to provide details about WIRED magazine. Within seconds, the attacker’s device—shown on the right—collects the majority of the response provided by the LLM by carrying out a LeftoverLocals attack on vulnerable GPU memory. The attack program the researchers created uses less than 10 lines of code.

An attacker (right) exploits the LeftoverLocals vulnerability to listen to LLM conversations.

Last summer, the researchers tested 11 chips from seven GPU makers and multiple corresponding programming frameworks. They found the LeftoverLocals vulnerability in GPUs from Apple, AMD, and Qualcomm and launched a far-reaching coordinated disclosure of the vulnerability in September in collaboration with the US-CERT Coordination Center and the Khronos Group, a standards body focused on 3D graphics, machine learning, and virtual and augmented reality.

The researchers did not find evidence that Nvidia, Intel, or Arm GPUs contain the LeftoverLocals vulnerability, but Apple, Qualcomm, and AMD all confirmed to WIRED that they are impacted. This means that well-known chips like the AMD Radeon RX 7900 XT and devices like Apple’s iPhone 12 Pro and M2 MacBook Air are vulnerable. The researchers did not find the flaw in the Imagination GPUs they tested, but others may be vulnerable.

Just 10 lines of code can steal AI secrets from Apple, AMD, and Qualcomm GPUs Read More »

as-2024-election-looms,-openai-says-it-is-taking-steps-to-prevent-ai-abuse

As 2024 election looms, OpenAI says it is taking steps to prevent AI abuse

Don’t Rock the vote —

ChatGPT maker plans transparency for gen AI content and improved access to voting info.

A pixelated photo of Donald Trump.

On Monday, ChatGPT maker OpenAI detailed its plans to prevent the misuse of its AI technologies during the upcoming elections in 2024, promising transparency in AI-generated content and enhancing access to reliable voting information. The AI developer says it is working on an approach that involves policy enforcement, collaboration with partners, and the development of new tools aimed at classifying AI-generated media.

“As we prepare for elections in 2024 across the world’s largest democracies, our approach is to continue our platform safety work by elevating accurate voting information, enforcing measured policies, and improving transparency,” writes OpenAI in its blog post. “Protecting the integrity of elections requires collaboration from every corner of the democratic process, and we want to make sure our technology is not used in a way that could undermine this process.”

Initiatives proposed by OpenAI include preventing abuse by means such as deepfakes or bots imitating candidates, refining usage policies, and launching a reporting system for the public to flag potential abuses. For example, OpenAI’s image generation tool, DALL-E 3, includes built-in filters that reject requests to create images of real people, including politicians. “For years, we’ve been iterating on tools to improve factual accuracy, reduce bias, and decline certain requests,” the company stated.

OpenAI says it regularly updates its Usage Policies for ChatGPT and its API products to prevent misuse, especially in the context of elections. The organization has implemented restrictions on using its technologies for political campaigning and lobbying until it better understands the potential for personalized persuasion. Also, OpenAI prohibits creating chatbots that impersonate real individuals or institutions and disallows the development of applications that could deter people from “participation in democratic processes.” Users can report GPTs that may violate the rules.

OpenAI claims to be proactively engaged in detailed strategies to safeguard its technologies against misuse. According to their statements, this includes red-teaming new systems to anticipate challenges, engaging with users and partners for feedback, and implementing robust safety mitigations. OpenAI asserts that these efforts are integral to its mission of continually refining AI tools for improved accuracy, reduced biases, and responsible handling of sensitive requests

Regarding transparency, OpenAI says it is advancing its efforts in classifying image provenance. The company plans to embed digital credentials, using cryptographic techniques, into images produced by DALL-E 3 as part of its adoption of standards by the Coalition for Content Provenance and Authenticity. Additionally, OpenAI says it is testing a tool designed to identify DALL-E-generated images.

In an effort to connect users with authoritative information, particularly concerning voting procedures, OpenAI says it has partnered with the National Association of Secretaries of State (NASS) in the United States. ChatGPT will direct users to CanIVote.org for verified US voting information.

“We want to make sure that our AI systems are built, deployed, and used safely,” writes OpenAI. “Like any new technology, these tools come with benefits and challenges. They are also unprecedented, and we will keep evolving our approach as we learn more about how our tools are used.”

As 2024 election looms, OpenAI says it is taking steps to prevent AI abuse Read More »

what-do-threads,-mastodon,-and-hospital-records-have-in-common?

What do Threads, Mastodon, and hospital records have in common?

A medical technician looks at a scan on a computer monitor.

It’s taken a while, but social media platforms now know that people prefer their information kept away from corporate eyes and malevolent algorithms. That’s why the newest generation of social media sites like Threads, Mastodon, and Bluesky boast of being part of the “fediverse.” Here, user data is hosted on independent servers rather than one corporate silo. Platforms then use common standards to share information when needed. If one server starts to host too many harmful accounts, other servers can choose to block it.

They’re not the only ones embracing this approach. Medical researchers think a similar strategy could help them train machine learning to spot disease trends in patients. Putting their AI algorithms on special servers within hospitals for “federated learning” could keep privacy standards high while letting researchers unravel new ways to detect and treat diseases.

“The use of AI is just exploding in all facets of life,” said Ronald M. Summers of the National Institutes of Health Clinical Center in Maryland, who uses the method in his radiology research. “There’s a lot of people interested in using federated learning for a variety of different data analysis applications.”

How does it work?

Until now, medical researchers refined their AI algorithms using a few carefully curated databases, usually anonymized medical information from patients taking part in clinical studies.

However, improving these models further means they need a larger dataset with real-world patient information. Researchers could pool data from several hospitals into one database, but that means asking them to hand over sensitive and highly regulated information. Sending patient information outside a hospital’s firewall is a big risk, so getting permission can be a long and legally complicated process. National privacy laws and the EU’s GDPR law set strict rules on sharing a patient’s personal information.

So instead, medical researchers are sending their AI model to hospitals so it can analyze a dataset while staying within the hospital’s firewall.

Typically, doctors first identify eligible patients for a study, select any clinical data they need for training, confirm its accuracy, and then organize it on a local database. The database is then placed onto a server at the hospital that is linked to the federated learning AI software. Once the software receives instructions from the researchers, it can work its AI magic, training itself with the hospital’s local data to find specific disease trends.

Every so often, this trained model is then sent back to a central server, where it joins models from other hospitals. An aggregation method processes these trained models to update the original model. For example, Google’s popular FedAvg aggregation algorithm takes each element of the trained models’ parameters and creates an average. Each average becomes part of the model update, with their input to the aggregate model weighted proportionally to the size of their training dataset.

In other words, how these models change gets aggregated in the central server to create an updated “consensus model.” This consensus model is then sent back to each hospital’s local database to be trained once again. The cycle continues until researchers judge the final consensus model to be accurate enough. (There’s a review of this process available.)

This keeps both sides happy. For hospitals, it helps preserve privacy since information sent back to the central server is anonymous; personal information never crosses the hospital’s firewall. It also means machine/AI learning can reach its full potential by training on real-world data so researchers get less biased results that are more likely to be sensitive to niche diseases.

Over the past few years, there has been a boom in research using this method. For example, in 2021, Summers and others used federated learning to see whether they could predict diabetes from CT scans of abdomens.

“We found that there were signatures of diabetes on the CT scanner [for] the pancreas that preceded the diagnosis of diabetes by as much as seven years,” said Summers. “That got us very excited that we might be able to help patients that are at risk.”

What do Threads, Mastodon, and hospital records have in common? Read More »

famous-xkcd-comic-comes-full-circle-with-ai-bird-identifying-binoculars

Famous xkcd comic comes full circle with AI bird-identifying binoculars

Who watches the bird watchers —

Swarovski AX Visio, billed as first “smart binoculars,” names species and tracks location.

The Swarovski Optik Visio binoculars, with an excerpt of a 2014 xkcd comic strip called

Enlarge / The Swarovski Optik Visio binoculars, with an excerpt of a 2014 xkcd comic strip called “Tasks” in the corner.

xckd / Swarovski

Last week, Austria-based Swarovski Optik introduced the AX Visio 10×32 binoculars, which the company says can identify over 9,000 species of birds and mammals using image recognition technology. The company is calling the product the world’s first “smart binoculars,” and they come with a hefty price tag—$4,799.

“The AX Visio are the world’s first AI-supported binoculars,” the company says in the product’s press release. “At the touch of a button, they assist with the identification of birds and other creatures, allow discoveries to be shared, and offer a wide range of practical extra functions.”

The binoculars, aimed mostly at bird watchers, gain their ability to identify birds from the Merlin Bird ID project, created by Cornell Lab of Ornithology. As confirmed by a hands-on demo conducted by The Verge, the user looks at an animal through the binoculars and presses a button. A red progress circle fills in while the binoculars process the image, then the identified animal name pops up on the built-in binocular HUD screen within about five seconds.

In 2014, a famous xkcd comic strip titled Tasks depicted someone asking a developer to create an app that, when a user takes a photo, will check whether the user is in a national park (deemed easy due to GPS) and check whether the photo is of a bird (to which the developer says, “I’ll need a research team and five years”). The caption below reads, “In CS, it can be hard to explain the difference between the easy and the virtually impossible.”

The xkcd comic titled

The xkcd comic titled “Tasks” from September 24, 2014.

It’s been just over nine years since the comic was published, and while identifying the presence of a bird in a photo was solved some time ago, these binoculars arguably go further by identifying the species of the bird in the photo (it also keeps track of location due to GPS). While apps to identify bird species already exist, this feature is now packed into a handheld pair of binoculars.

According to Swarovski, the development of the AX Visio took approximately five years, involving around 390 “hardware parts.” The binoculars incorporate a neural processing unit (NPU) for object recognition processing. The company claims that the device will have a long product life cycle, with ongoing updates and improvements. The company also mentions “an open programming interface” in its press release, potentially allowing industrious users (or handy hackers) to expand the unit’s features over time.

  • The Swarovski Optik Visio binoculars.

    Swarovski Optik

  • The Swarovski Optik Visio binoculars.

    Swarovski Optik

  • The Swarovski Optik Visio binoculars.

    Swarovski Optik

The binoculars, which feature industrial design from Marc Newson, include built-in digital camera, compass, GPS, and discovery-sharing features that can “immediately show your companion where you have seen an animal.” The Visio unit also wirelessly ties into the “SWAROVSKI OPTIK Outdoor App” that can run on a smartphone. The app manages sharing photos and videos captured through the binoculars. (As an aside, we’ve come a long way from computer-connected gadgets that required pesky serial cables in the late 1990s.)

Swarovski says the AX Visio will be available at select retailers and online starting February 1, 2024. While this tech is at a premium price right now, given the speed of tech progress and market competition, we may see similar image-recognizing features built into much cheaper models in the years ahead.

Famous xkcd comic comes full circle with AI bird-identifying binoculars Read More »

lazy-use-of-ai-leads-to-amazon-products-called-“i-cannot-fulfill-that-request”

Lazy use of AI leads to Amazon products called “I cannot fulfill that request”

FILE NOT FOUND —

The telltale error messages are a sign of AI-generated pablum all over the Internet.

I know naming new products can be hard, but these Amazon sellers made some particularly odd naming choices.

Enlarge / I know naming new products can be hard, but these Amazon sellers made some particularly odd naming choices.

Amazon

Amazon users are at this point used to search results filled with products that are fraudulent, scams, or quite literally garbage. These days, though, they also may have to pick through obviously shady products, with names like “I’m sorry but I cannot fulfill this request it goes against OpenAI use policy.”

As of press time, some version of that telltale OpenAI error message appears in Amazon products ranging from lawn chairs to office furniture to Chinese religious tracts. A few similarly named products that were available as of this morning have been taken down as word of the listings spreads across social media (one such example is Archived here).

ProTip: Don't ask OpenAI to integrate a trademarked brand name when generating a name for your weird length of rubber tubing.

Enlarge / ProTip: Don’t ask OpenAI to integrate a trademarked brand name when generating a name for your weird length of rubber tubing.

Other Amazon product names don’t mention OpenAI specifically but feature apparent AI-related error messages, such as “Sorry but I can’t generate a response to that request” or “Sorry but I can’t provide the information you’re looking for,” (available in a variety of colors). Sometimes, the product names even highlight the specific reason why the apparent AI-generation request failed, noting that OpenAI can’t provide content that “requires using trademarked brand names” or “promotes a specific religious institution” or in one case “encourage unethical behavior.”

The repeated invocation of a

Enlarge / The repeated invocation of a “commitment to providing reliable and trustworthy product descriptions” cited in this description is particularly ironic.

The descriptions for these oddly named products are also riddled with obvious AI error messages like, “Apologies, but I am unable to provide the information you’re seeking.” One product description for a set of tables and chairs (which has since been taken down) hilariously noted: “Our [product] can be used for a variety of tasks, such [task 1], [task 2], and [task 3]].” Another set of product descriptions, seemingly for tattoo ink guns, repeatedly apologizes that it can’t provide more information because: “We prioritize accuracy and reliability by only offering verified product details to our customers.”

Spam spam spam spam

Using large language models to help generate product names or descriptions isn’t against Amazon policy. On the contrary, in September Amazon launched its own generative AI tool to help sellers “create more thorough and captivating product descriptions, titles, and listing details.” And we could only find a small handful of Amazon products slipping through with the telltale error messages in their names or descriptions as of press time.

Still, these error-message-filled listings highlight the lack of care or even basic editing many Amazon scammers are exercising when putting their spammy product listings on the Amazon marketplace. For every seller that can be easily caught accidentally posting an OpenAI error, there are likely countless others using the technology to create product names and descriptions that only seem like they were written by a human that has actual experience with the product in question.

A set of clearly real people conversing on Twitter / X.

Enlarge / A set of clearly real people conversing on Twitter / X.

Amazon isn’t the only online platform where these AI bots are outing themselves, either. A quick search for “goes against OpenAI policy” or “as an AI language model” can find a whole lot of artificial posts on Twitter / X or Threads or LinkedIn, for example. Security engineer Dan Feldman noted a similar problem on Amazon back in April, though searching with the phrase “as an AI language model” doesn’t seem to generate any obviously AI-generated search results these days.

As fun as it is to call out these obvious mishaps for AI-generated content mills, a flood of harder-to-detect AI content is threatening to overwhelm everyone from art communities to sci-fi magazines to Amazon’s own ebook marketplace. Pretty much any platform that accepts user submissions that involve text or visual art now has to worry about being flooded with wave after wave of AI-generated work trying to crowd out the human community they were created for. It’s a problem that’s likely to get worse before it gets better.

Listing image by Getty Images | Leon Neal

Lazy use of AI leads to Amazon products called “I cannot fulfill that request” Read More »