Tech

after-borking-my-pixel-4a-battery,-google-borks-me,-too

After borking my Pixel 4a battery, Google borks me, too


The devil is in the details.

The Pixel 4a. It’s finally here! Credit: Google

It is an immutable law of nature that when you receive a corporate email with a subject line like “Changes coming to your Pixel 4a,” the changes won’t be the sort you like. Indeed, a more honest subject line would usually be: “You’re about to get hosed.”

So I wasn’t surprised, as I read further into this January missive from Google, that an “upcoming software update for your Pixel 4a” would “affect the overall performance and stability of its battery.”

How would my battery be affected? Negatively, of course. “This update will reduce your battery’s runtime and charging performance,” the email said. “To address this, we’re providing some options to consider. “

Our benevolent Google overlords were about to nerf my phone battery—presumably in the interests of “not having it erupt in flames,” though this was never actually made clear—but they recognized the problem, and they were about to provide compensation. This is exactly how these kinds of situations should be handled.

Google offered three options: $50 cash money, a $100 credit to Google’s online store, or a free battery replacement. It seemed fair enough. Yes, not having my phone for a week or two while I shipped it roundtrip to Google could be annoying, but at least the company was directly mitigating the harm it was about to inflict. Indeed, users might actually end up in better shape than before, given the brand-new battery.

So I was feeling relatively sunny toward the giant monopolist when I decided to spring for the 50 simoleons. My thinking was that 1) I didn’t want to lose my phone for a couple of weeks, 2) the update might not be that bad, in which case I’d be ahead by 50 bucks, and 3) I could always put the money towards a battery replacement if assumption No. 2 turned out to be mistaken.

The naïveté of youth!

I selected my $50 “appeasement” through an online form, and two days later, I received an email from Bharath on the Google Support Team.

Bharath wanted me to know that I was eligible for the money and it would soon be in my hands… once I performed a small, almost trivial task: giving some company I had never heard of my name, address, phone number, Social Security number, date of birth, and bank account details.

About that $50…

Google was not, in fact, just “sending” me $50. I had expected, since the problem involved their phones and their update, that the solution would require little or nothing from me. A check or prepaid credit card would arrive in the mail, perhaps, or a drone might deliver a crisp new bill from the sky. I didn’t know and didn’t care, so long as it wasn’t my problem.

But it was my problem. To get the cash, I had to create an account with something called “Payoneer.” This is apparently a reputable payments company, but I had never heard of it, and much about its operations is unclear. For instance, I was given three different ways to sign up depending on whether I 1) “already have a Payoneer account from Google,” 2) “don’t have an account,” or 3) “do have a Payoneer account that was not provided nor activated through Google.”

Say what now?

And though Google promised “no transaction fees,” Payoneer appears to charge an “annual account fee” of $29.95… but only to accounts that receive less than $2,000 through Payoneer in any consecutive 12-month period.

Does this fee apply to me if I sign up through the Google offer? I was directed to Payoneer support with any questions, but the company’s FAQ on the annual account fee doesn’t say.

If the fee does apply to me, do I need to sign up for a Payoneer account, give them all of my most personal financial information, wait the “10 to 18 business days” that Google says it will take to get my money, and then return to Payoneer so that I can cancel my account before racking up some $30 charge a year from now? And I’m supposed to do all this just to get…. fifty bucks? One time?

It was far simpler for me to get a recent hundred-dollar rebate on a washing machine… and they didn’t need my SSN or bank account information.

(Reddit users also report that, if you use the wrong web browser to cancel your Payoneer account, you’re hit with an error that says: “This end point requires that the body of all requests be formatted as JSON.”)

Like Lando Calrissian, I realized that this deal was getting worse all the time.

I planned to write Bharath back to switch my “appeasement,” but then I noticed the fine print: No changes are possible after making a selection.

So—no money for me. On the scale of life’s crises, losing $50 is a minor one, and I resolved to move on, facing the world with a cheerful heart and a clear mind, undistracted by the many small annoyances our high-tech overlords continually strew upon the path.

Then the software update arrived.

A decimation situation

When Google said that the new Pixel 4a update would “reduce your battery’s runtime and charging performance,” it was not kidding. Indeed, the update basically destroyed the battery.

Though my phone was three years old, until January of this year, the battery still held up for all-day usage. The screen was nice, the (smallish) phone size was good, and the device remained plenty fast at all the basic tasks: texting, emails, web browsing, snapping photos. I’m trying to reduce both my consumerism and my e-waste, so I was planning to keep the device for at least another year. And even then, it would make a decent hand-me-down device for my younger kids.

After the update, however, the phone burned through a full battery charge in less than two hours. I could pull up a simple podcast app, start playing an episode, and watch the battery percentage decrement every 45 seconds or so. Using the phone was nearly impossible unless one was near a charging cable at all times.

To recap: My phone was shot, I had to jump through several hoops to get my money, and I couldn’t change my “appeasement” once I realized that it wouldn’t work for me.

Within the space of three days, I went from 1) being mildly annoyed at the prospect of having my phone messed with remotely to 2) accepting that Google was (probably) doing it for my own safety and was committed to making things right to 3) berating Google for ruining my device and then using a hostile, data collecting “appeasement” program to act like it cared. This was probably not the impression Google hoped to leave in people’s minds when issuing the Pixel 4a update.

Pixel 4a, disassembled, with two fingers holding its battery above the front half.

Removing the Pixel 4a’s battery can be painful, but not as painful as catching fire. Credit: iFixit

Cheap can be quite expensive

The update itself does not appear to be part of some plan to spy on us or to extract revenue but rather to keep people safe. The company tried to remedy the pain with options that, on the surface, felt reasonable, especially given the fact that batteries are well-known as consumable objects that degrade over time. And I’ve had three solid years of service with the 4a, which wasn’t especially expensive to begin with.

That said, I do blame Google in general for the situation. The inflexibility of the approach, the options that aren’t tailored for ease of use in specific countries, the outsourced tech support—these are all hallmarks of today’s global tech behemoths.

It is more efficient, from an algorithmic, employ-as-few-humans-as-possible perspective, to operate “at scale” by choosing global technical solutions over better local options, by choosing outsourced email support, by trying to avoid fraud (and employee time) through preventing program changes, by asking the users to jump through your hoops, by gobbling up ultra-sensitive information because it makes things easier on your end.

While this makes a certain kind of sense, it’s not fun to receive this kind of “efficiency.” When everything goes smoothly, it’s fine—but whenever there’s a problem, or questions arise, these kinds of “efficient, scalable” approaches usually just mean “you’re about to get screwed.”

In the end, Google is willing to pay me $50, but that money comes with its own cost. I’m not willing to pay with my time nor with the risk of my financial information, and I will increasingly turn to companies that offer a better experience, that care more about data privacy, that build with higher-quality components, and that take good care of customers.

No company is perfect, of course, and this approach costs a bit more, which butts up against my powerful urge to get a great deal on everything. I have to keep relearning the old lesson— as I am once again with this Pixel 4a fiasco—that cheap gear is not always the best value in the long run.

Photo of Nate Anderson

After borking my Pixel 4a battery, Google borks me, too Read More »

“myterms”-wants-to-become-the-new-way-we-dictate-our-privacy-on-the-web

“MyTerms” wants to become the new way we dictate our privacy on the web

Searls and his group are putting up the standards and letting the browsers, extension-makers, website managers, mobile platforms, and other pieces of the tech stack craft the tools. So long as the human is the first party to a contract, the digital thing is the second, a “disinterested non-profit” provides the roster of agreements, and both sides keep records of what they agreed to, the function can take whatever shape the Internet decides.

Terms offered, not requests submitted

Searls’ and his group’s standard is a plea for a sensible alternative to the modern reality of accessing web information. It asks us to stop pretending that we’re all reading agreements stuffed full with opaque language, agreeing to thousands upon thousands of words’ worth of terms every day and willfully offering up information about us. And, of course, it makes people ask if it is due to become another version of Do Not Track.

Do Not Track was a request, while MyTerms is inherently a demand. Websites and services could, of course, simply refuse to show or provide content and data if a MyTerms agent is present, or they could ask or demand that people set the least restrictive terms.

There is nothing inherently wrong with setting up a user-first privacy scheme and pushing for sites and software to do the right thing and abide by it. People may choose to stick to search engines and sites that agree to MyTerms. Time will tell if MyTerms can gain the kind of leverage Searls is aiming for.

“MyTerms” wants to become the new way we dictate our privacy on the web Read More »

can-we-make-ai-less-power-hungry?-these-researchers-are-working-on-it.

Can we make AI less power-hungry? These researchers are working on it.


As demand surges, figuring out the performance of proprietary models is half the battle.

Credit: Igor Borisenko/Getty Images

Credit: Igor Borisenko/Getty Images

At the beginning of November 2024, the US Federal Energy Regulatory Commission (FERC) rejected Amazon’s request to buy an additional 180 megawatts of power directly from the Susquehanna nuclear power plant for a data center located nearby. The rejection was due to the argument that buying power directly instead of getting it through the grid like everyone else works against the interests of other users.

Demand for power in the US has been flat for nearly 20 years. “But now we’re seeing load forecasts shooting up. Depending on [what] numbers you want to accept, they’re either skyrocketing or they’re just rapidly increasing,” said Mark Christie, a FERC commissioner.

Part of the surge in demand comes from data centers, and their increasing thirst for power comes in part from running increasingly sophisticated AI models. As with all world-shaping developments, what set this trend into motion was vision—quite literally.

The AlexNet moment

Back in 2012, Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton, AI researchers at the University of Toronto, were busy working on a convolution neural network (CNN) for the ImageNet LSRVC, an image-recognition contest. The contest’s rules were fairly simple: A team had to build an AI system that could categorize images sourced from a database comprising over a million labeled pictures.

The task was extremely challenging at the time, so the team figured they needed a really big neural net—way bigger than anything other research teams had attempted. AlexNet, named after the lead researcher, had multiple layers, with over 60 million parameters and 650 thousand neurons. The problem with a behemoth like that was how to train it.

What the team had in their lab were a few Nvidia GTX 580s, each with 3GB of memory. As the researchers wrote in their paper, AlexNet was simply too big to fit on any single GPU they had. So they figured out how to split AlexNet’s training phase between two GPUs working in parallel—half of the neurons ran on one GPU, and the other half ran on the other GPU.

AlexNet won the 2012 competition by a landslide, but the team accomplished something way more profound. The size of AI models was once and for all decoupled from what was possible to do on a single CPU or GPU. The genie was out of the bottle.

(The AlexNet source code was recently made available through the Computer History Museum.)

The balancing act

After AlexNet, using multiple GPUs to train AI became a no-brainer. Increasingly powerful AIs used tens of GPUs, then hundreds, thousands, and more. But it took some time before this trend started making its presence felt on the grid. According to an Electric Power Research Institute (EPRI) report, the power consumption of data centers was relatively flat between 2010 and 2020. That doesn’t mean the demand for data center services was flat, but the improvements in data centers’ energy efficiency were sufficient to offset the fact we were using them more.

Two key drivers of that efficiency were the increasing adoption of GPU-based computing and improvements in the energy efficiency of those GPUs. “That was really core to why Nvidia was born. We paired CPUs with accelerators to drive the efficiency onward,” said Dion Harris, head of Data Center Product Marketing at Nvidia. In the 2010–2020 period, Nvidia data center chips became roughly 15 times more efficient, which was enough to keep data center power consumption steady.

All that changed with the rise of enormous large language transformer models, starting with ChatGPT in 2022. “There was a very big jump when transformers became mainstream,” said Mosharaf Chowdhury, a professor at the University of Michigan. (Chowdhury is also at the ML Energy Initiative, a research group focusing on making AI more energy-efficient.)

Nvidia has kept up its efficiency improvements, with a ten-fold boost between 2020 and today. The company also kept improving chips that were already deployed. “A lot of where this efficiency comes from was software optimization. Only last year, we improved the overall performance of Hopper by about 5x,” Harris said. Despite these efficiency gains, based on Lawrence Berkely National Laboratory estimates, the US saw data center power consumption shoot up from around 76 TWh in 2018 to 176 TWh in 2023.

The AI lifecycle

LLMs work with tens of billions of neurons approaching a number rivaling—and perhaps even surpassing—those in the human brain. The GPT 4 is estimated to work with around 100 billion neurons distributed over 100 layers and over 100 trillion parameters that define the strength of connections among the neurons. These parameters are set during training, when the AI is fed huge amounts of data and learns by adjusting these values. That’s followed by the inference phase, where it gets busy processing queries coming in every day.

The training phase is a gargantuan computational effort—Open AI supposedly used over 25,000 Nvidia Ampere 100 GPUs running on all cylinders for 100 days. The estimated power consumption is 50 GW-hours, which is enough to power a medium-sized town for a year. According to numbers released by Google, training accounts for 40 percent of the total AI model power consumption over its lifecycle. The remaining 60 percent is inference, where power consumption figures are less spectacular but add up over time.

Trimming AI models down

The increasing power consumption has pushed the computer science community to think about how to keep memory and computing requirements down without sacrificing performance too much. “One way to go about it is reducing the amount of computation,” said Jae-Won Chung, a researcher at the University of Michigan and a member of the ML Energy Initiative.

One of the first things researchers tried was a technique called pruning, which aimed to reduce the number of parameters. Yann LeCun, now the chief AI scientist at Meta, proposed this approach back in 1989, terming it (somewhat menacingly) “the optimal brain damage.” You take a trained model and remove some of its parameters, usually targeting the ones with a value of zero, which add nothing to the overall performance. “You take a large model and distill it into a smaller model trying to preserve the quality,” Chung explained.

You can also make those remaining parameters leaner with a trick called quantization. Parameters in neural nets are usually represented as a single-precision floating point number, occupying 32 bits of computer memory. “But you can change the format of parameters to a smaller one that reduces the amount of needed memory and makes the computation faster,” Chung said.

Shrinking an individual parameter has a minor effect, but when there are billions of them, it adds up. It’s also possible to do quantization-aware training, which performs quantization at the training stage. According to Nvidia, which implemented quantization training in its AI model optimization toolkit, this should cut the memory requirements by 29 to 51 percent.

Pruning and quantization belong to a category of optimization techniques that rely on tweaking the way AI models work internally—how many parameters they use and how memory-intensive their storage is. These techniques are like tuning an engine in a car to make it go faster and use less fuel. But there’s another category of techniques that focus on the processes computers use to run those AI models instead of the models themselves—akin to speeding a car up by timing the traffic lights better.

Finishing first

Apart from optimizing the AI models themselves, we could also optimize the way data centers run them. Splitting the training phase workload evenly among 25 thousand GPUs introduces inefficiencies. “When you split the model into 100,000 GPUs, you end up slicing and dicing it in multiple dimensions, and it is very difficult to make every piece exactly the same size,” Chung said.

GPUs that have been given significantly larger workloads have increased power consumption that is not necessarily balanced out by those with smaller loads. Chung figured that if GPUs with smaller workloads ran slower, consuming much less power, they would finish roughly at the same time as GPUs processing larger workloads operating at full speed. The trick was to pace each GPU in such a way that the whole cluster would finish at the same time.

To make that happen, Chung built a software tool called Perseus that identified the scope of the workloads assigned to each GPU in a cluster. Perseus takes the estimated time needed to complete the largest workload on a GPU running at full. It then estimates how much computation must be done on each of the remaining GPUs and determines what speed to run them so they finish at the same. “Perseus precisely slows some of the GPUs down, and slowing down means less energy. But the end-to-end speed is the same,” Chung said.

The team tested Perseus by training the publicly available GPT-3, as well as other large language models and a computer vision AI. The results were promising. “Perseus could cut up to 30 percent of energy for the whole thing,” Chung said. He said the team is talking about deploying Perseus at Meta, “but it takes a long time to deploy something at a large company.”

Are all those optimizations to the models and the way data centers run them enough to keep us in the green? It takes roughly a year or two to plan and build a data center, but it can take longer than that to build a power plant. So are we winning this race or losing? It’s a bit hard to say.

Back of the envelope

As the increasing power consumption of data centers became apparent, research groups tried to quantify the problem. A Lawerence Berkley Laboratory team estimated that data centers’ annual energy draw in 2028 would be between 325 and 580 TWh in the US—that’s between 6.7 and 12 percent of the total US electricity consumption. The International Energy Agency thinks it will be around 6 percent by 2026. Goldman Sachs Research says 8 percent by 2030, while EPRI claims between 4.6 and 9.1 percent by 2030.

EPRI also warns that the impact will be even worse because data centers tend to be concentrated at locations investors think are advantageous, like Virginia, which already sends 25 percent of its electricity to data centers. In Ireland, data centers are expected to consume one-third of the electricity produced in the entire country in the near future. And that’s just the beginning.

Running huge AI models like ChatGPT is one of the most power-intensive things that data centers do, but it accounts for roughly 12 percent of their operations, according to Nvidia. That is expected to change if companies like Google start to weave conversational LLMs into their most popular services. The EPRI report estimates that a single Google search today uses around 0.3 watts of power, while a single Chat GPT query bumps that up to 2.9 watts. Based on those values, the report estimates that an AI-powered Google search would require Google to deploy 400,000 new servers that would consume 22.8 TWh per year.

“AI searches take 10x the electricity of a non-AI search,” Christie, the FERC commissioner, said at a FERC-organized conference. When FERC commissioners are using those numbers, you’d think there would be rock-solid science backing them up. But when Ars asked Chowdhury and Chung about their thoughts on these estimates, they exchanged looks… and smiled.

Closed AI problem

Chowdhury and Chung don’t think those numbers are particularly credible. They feel we know nothing about what’s going on inside commercial AI systems like ChatGPT or Gemini, because OpenAI and Google have never released actual power-consumption figures.

“They didn’t publish any real numbers, any academic papers. The only number, 0.3 watts per Google search, appeared in some blog post or other PR-related thingy,” Chodwhury said. We don’t know how this power consumption was measured, on what hardware, or under what conditions, he said. But at least it came directly from Google.

“When you take that 10x Google vs ChatGPT equation or whatever—one part is half-known, the other part is unknown, and then the division is done by some third party that has no relationship with Google nor with Open AI,” Chowdhury said.

Google’s “PR-related thingy” was published back in 2009, while the 2.9-watts-per-ChatGPT-query figure was probably based on a comment about the number of GPUs needed to train GPT-4 made by Jensen Huang, Nvidia’s CEO, in 2024. That means the “10x AI versus non-AI search” claim was actually based on power consumption achieved on entirely different generations of hardware separated by 15 years. “But the number seemed plausible, so people keep repeating it,” Chowdhury said.

All reports we have today were done by third parties that are not affiliated with the companies building big AIs, and yet they arrive at weirdly specific numbers. “They take numbers that are just estimates, then multiply those by a whole lot of other numbers and get back with statements like ‘AI consumes more energy than Britain, or more than Africa, or something like that.’ The truth is they don’t know that,” Chowdhury said.

He argues that better numbers would require benchmarking AI models using a formal testing procedure that could be verified through the peer-review process.

As it turns out, the ML Energy Initiative defined just such a testing procedure and ran the benchmarks on any AI models they could get ahold of. The group then posted the results online on their ML.ENERGY Leaderboard.

AI-efficiency leaderboard

To get good numbers, the first thing the ML Energy Initiative got rid of was the idea of estimating how power-hungry GPU chips are by using their thermal design power (TDP), which is basically their maximum power consumption. Using TDP was a bit like rating a car’s efficiency based on how much fuel it burned running at full speed. That’s not how people usually drive, and that’s not how GPUs work when running AI models. So Chung built ZeusMonitor, an all-in-one solution that measured GPU power consumption on the fly.

For the tests, his team used setups with Nvidia’s A100 and H100 GPUs, the ones most commonly used at data centers today, and measured how much energy they used running various large language models (LLMs), diffusion models that generate pictures or videos based on text input, and many other types of AI systems.

The largest LLM included in the leaderboard was Meta’s Llama 3.1 405B, an open-source chat-based AI with 405 billion parameters. It consumed 3352.92 joules of energy per request running on two H100 GPUs. That’s around 0.93 watt-hours—significantly less than 2.9 watt-hours quoted for ChatGPT queries. These measurements confirmed the improvements in the energy efficiency of hardware. Mixtral 8x22B was the largest LLM the team managed to run on both Ampere and Hopper platforms. Running the model on two Ampere GPUs resulted in 0.32 watt-hours per request, compared to just 0.15 watt-hours on one Hopper GPU.

What remains unknown, however, is the performance of proprietary models like GPT-4, Gemini, or Grok. The ML Energy Initiative team says it’s very hard for the research community to start coming up with solutions to the energy efficiency problems when we don’t even know what exactly we’re facing. We can make estimates, but Chung insists they need to be accompanied by error-bound analysis. We don’t have anything like that today.

The most pressing issue, according to Chung and Chowdhury, is the lack of transparency. “Companies like Google or Open AI have no incentive to talk about power consumption. If anything, releasing actual numbers would harm them,” Chowdhury said. “But people should understand what is actually happening, so maybe we should somehow coax them into releasing some of those numbers.”

Where rubber meets the road

“Energy efficiency in data centers follows the trend similar to Moore’s law—only working at a very large scale, instead of on a single chip,” Nvidia’s Harris said. The power consumption per rack, a unit used in data centers housing between 10 and 14 Nvidia GPUs, is going up, he said, but the performance-per-watt is getting better.

“When you consider all the innovations going on in software optimization, cooling systems, MEP (mechanical, electrical, and plumbing), and GPUs themselves, we have a lot of headroom,” Harris said. He expects this large-scale variant of Moore’s law to keep going for quite some time, even without any radical changes in technology.

There are also more revolutionary technologies looming on the horizon. The idea that drove companies like Nvidia to their current market status was the concept that you could offload certain tasks from the CPU to dedicated, purpose-built hardware. But now, even GPUs will probably use their own accelerators in the future. Neural nets and other parallel computation tasks could be implemented on photonic chips that use light instead of electrons to process information. Photonic computing devices are orders of magnitude more energy-efficient than the GPUs we have today and can run neural networks literally at the speed of light.

Another innovation to look forward to is 2D semiconductors, which enable building incredibly small transistors and stacking them vertically, vastly improving the computation density possible within a given chip area. “We are looking at a lot of these technologies, trying to assess where we can take them,” Harris said. “But where rubber really meets the road is how you deploy them at scale. It’s probably a bit early to say where the future bang for buck will be.”

The problem is when we are making a resource more efficient, we simply end up using it more. “It is a Jevons paradox, known since the beginnings of the industrial age. But will AI energy consumption increase so much that it causes an apocalypse? Chung doesn’t think so. According to Chowdhury, if we run out of energy to power up our progress, we will simply slow down.

“But people have always been very good at finding the way,” Chowdhury added.

Photo of Jacek Krywko

Jacek Krywko is a freelance science and technology writer who covers space exploration, artificial intelligence research, computer science, and all sorts of engineering wizardry.

Can we make AI less power-hungry? These researchers are working on it. Read More »

ceo-of-ai-ad-tech-firm-pledging-“world-free-of-fraud”-sentenced-for-fraud

CEO of AI ad-tech firm pledging “world free of fraud” sentenced for fraud

In May 2024, the website of ad-tech firm Kubient touted that the company was “a perfect blend” of ad veterans and developers, “committed to solving the growing problem of fraud” in digital ads. Like many corporate sites, it also linked old blog posts from its home page, including a May 2022 post on “How to create a world free of fraud: Kubient’s secret sauce.”

These days, Kubient’s website cannot be reached, the team is no more, and CEO Paul Roberts is due to serve one year and one day in prison, having pled guilty Thursday to creating his own small world of fraud. Roberts, according to federal prosecutors, schemed to create $1.3 million in fraudulent revenue statements to bolster Kubient’s initial public offering (IPO) and significantly oversold “KAI,” Kubient’s artificial intelligence tool.

The core of the case is an I-pay-you, you-pay-me gambit that Roberts initiated with an unnamed “Company-1,” according to prosecutors. Kubient and this firm would each bill the other for nearly identical amounts, with Kubient purportedly deploying KAI to find instances of ad fraud in the other company’s ad spend.

Roberts, prosecutors said, “directed Kubient employees to generate fake KAI reports based on made-up metrics and no underlying data at all.” These fake reports helped sell the story to independent auditors and book the synthetic revenue in financial statements, according to Roberts’ indictment.

CEO of AI ad-tech firm pledging “world free of fraud” sentenced for fraud Read More »

sometimes,-it’s-the-little-tech-annoyances-that-sting-the-most

Sometimes, it’s the little tech annoyances that sting the most

Anyone who has suffered the indignity of splinter, a blister, or a paper cut knows that small things can sometimes be hugely annoying. You aren’t going to die from any of these conditions, but it’s still hard to focus when, say, the back of your right foot is rubbing a new blister against the inside of your not-quite-broken-in-yet hiking boots.

I found myself in the computing version of this situation yesterday, when I was trying to work on a new Mac Mini and was brought up short by the fact that my third mouse button (that is, clicking on the scroll wheel) did nothing. This was odd, because I have for many years assigned this button to “Mission Control” on macOS—a feature that tiles every open window on your machine, making it quick and easy to switch apps. When I got the new Mini, I immediately added this to my settings. Boom!

And yet there I was, a couple hours later, clicking the middle mouse button by reflex and getting no result. This seemed quite odd—had I only imagined that I made the settings change? I made the alteration again in System Settings and went back to work.

But after a reboot later that day to install an OS update, I found that my shortcut setting for Mission Control had once again been wiped away. This wasn’t happening with any other settings changes, and it was strangely vexing.

When it happened a third time, I switched into full “research and destroy the problem” mode. One of my Ars colleagues commiserated with me, writing, “This kind of powerful-annoying stuff is just so common. I swear at least once every few months, some shortcut or whatever just stops working, and sometimes, after a week or so, it starts working again. No rhyme, reason, or apparent causality except that computers are just [unprintable expletives].”

But even if computers are [unprintable expletives], their problems have often been encountered and fixed by some other poor soul. I turned to the Internet for help… and immediately stumbled upon an Apple discussion thread called “MacOS mouse shortcuts are reset upon restart or shutdown.” The poster—and most of those replying—said that the odd behavior had only appeared in macOS Sequoia. One reply claimed to have identified the source of the bug and offered a fix:

Sometimes, it’s the little tech annoyances that sting the most Read More »

italy-demands-google-poison-dns-under-strict-piracy-shield-law

Italy demands Google poison DNS under strict Piracy Shield law

Spotted by TorrentFreak, AGCOM Commissioner Massimiliano Capitanio took to LinkedIn to celebrate the ruling, as well as the existence of the Italian Piracy Shield. “The Judge confirmed the value of AGCOM’s investigations, once again giving legitimacy to a system for the protection of copyright that is unique in the world,” said Capitanio.

Capitanio went on to complain that Google has routinely ignored AGCOM’s listing of pirate sites, which are supposed to be blocked in 30 minutes or less under the law. He noted the violation was so clear-cut that the order was issued without giving Google a chance to respond, known as inaudita altera parte in Italian courts.

This decision follows a similar case against Internet backbone firm Cloudflare. In January, the Court of Milan found that Cloudflare’s CDN, DNS server, and WARP VPN were facilitating piracy. The court threatened Cloudflare with fines of up to 10,000 euros per day if it did not begin blocking the sites.

Google could face similar sanctions, but AGCOM has had difficulty getting international tech behemoths to acknowledge their legal obligations in the country. We’ve reached out to Google for comment and will update this report if we hear back.

Italy demands Google poison DNS under strict Piracy Shield law Read More »

apple-loses-$1b-a-year-on-prestigious,-minimally-viewed-apple-tv+:-report

Apple loses $1B a year on prestigious, minimally viewed Apple TV+: report

The Apple TV+ streaming service “is losing more than $1 billion annually,” according to The Information today.

The report also claimed that Apple TV+’s subscriber count reached “around 45 million” in 2024, citing the two anonymous sources.

Ars reached out to Apple for comment on the accuracy of The Information’s report and will update this article if we hear back.

According to one of the sources, Apple TV+ has typically spent over $5 billion annually on content since 2019, when Apple TV+ debuted. Last year, though, Apple CEO Tim Cook reportedly cut the budget by about $500 million. The reported numbers are similar to a July report from Bloomberg that claimed that Apple had spent over $20 billion on Apple TV+’s library. For comparison, Netflix has 301.63 million subscribers and expects to spend $18 billion on content in 2025.

In the year preceding Apple TV+’s debut, Apple services chief Eddy Cue reportedly pushed back on executive requests to be stingier with content spending, “a person with direct knowledge of the matter” told The Information.

But Cook started paying closer attention to Apple TV+’s spending after the 2022 Oscars, where the Apple TV+ original CODA won Best Picture. The award signaled the significance of Apple TV+ as a business.

Per The Information, spending related to Apple TV+ previously included lavish perks for actors and producers. Apple paid “hundreds of thousands of dollars per flight” to transport Apple TV+ actors and producers to promotional events, The Information said, noting that such spending “is common in Hollywood” but “more unusual at Apple.” Apple’s finance department reportedly pushed Apple TV+ executives to find better flight deals sometime around 2023.

In 2024, Cook questioned big-budget Apple TV+ films, like the $200 million Argylle, which he said failed to generate impressive subscriber boosts or viewership, an anonymous “former Apple TV+ employee” shared. Cook reportedly cut about $500 million from the Apple TV+ content budget in 2024.

Apple loses $1B a year on prestigious, minimally viewed Apple TV+: report Read More »

hp-avoids-monetary-damages-over-bricked-printers-in-class-action-settlement

HP avoids monetary damages over bricked printers in class-action settlement

HP also now provides disclaimers on the product pages for most of the printers that it sells, stating that the device “is intended to work only with cartridges that have a new or reused HP chip” and uses Dynamic Security “to block cartridges using a non-HP chip.”

“Periodic firmware updates will maintain the effectiveness of these measures and block cartridges that previously worked. A reused HP chip enables the use of reused, remanufactured, and refilled cartridges,” the disclaimer says, adding a link to a support page about Dynamic Security. The support page notes that “most HP printers can be configured to receive updates either automatically or with a notification that allows you to choose whether to update or not.” However, some HP programs, like Instant Ink, require users to enable automatic firmware updates on HP printers.

All this means that, despite the recently approved settlement, Dynamic Security remains a critical part of most HP printers, and HP will continue to feel entitled to use firmware updates to suddenly block printers made after December 1, 2016, from using non-HP ink and toner. Owners of HP printers made after that date that allow automatic updates and still work with third-party accessories shouldn’t be surprised if that ability is suddenly bricked one day.

Dynamic litigation

While HP isn’t paying a sum to class-action members this time, it has previously agreed to pay millions in relation to bricking printers: In 2022, it agreed to pay $1.35 million to European customers, and in 2020, the Italian Antitrust Authority fined HP for 10 million euros. In 2019, HP said it would pay $1.5 million to settle a similar class-action case in California, and it paid approximately AUD$50 each to Australian customers impacted by Dynamic Security in 2018.

There’s also an open case against HP regarding its ink practices, a class-action complaint filed in the US District Court for the Northern District of Illinois in January 2024. The lawsuit centers on Dynamic Security firmware updates pushed “in late 2022 and early 2023″ and accuses HP of creating a “monopoly in the aftermarket for replacement cartridges” [PDF]. The plaintiffs seek an order declaring that HP broke the law, an injunction against Dynamic Security, and monetary and punitive damages.

Another lawsuit, filed in mid-2022 about some HP all-in-one printers failing to scan or fax without ink, was dismissed.

HP’s printer arm has other pressing matters to address, though. Earlier this month, a firmware update broke specific HP printer models, preventing them from printing, even when using HP-brand ink. HP told Ars last week that it’s “actively working on a solution.”

HP avoids monetary damages over bricked printers in class-action settlement Read More »

apple-and-google-in-the-hot-seat-as-european-regulators-ignore-trump-warnings

Apple and Google in the hot seat as European regulators ignore Trump warnings

The European Commission is not backing down from efforts to rein in Big Tech. In a series of press releases today, the European Union’s executive arm has announced actions against both Apple and Google. Regulators have announced that Apple will be required to open up support for non-Apple accessories on the iPhone, but it may be too late for Google to make changes. The commission says the search giant has violated the Digital Markets Act, which could lead to a hefty fine.

Since returning to power, Donald Trump has railed against European regulations that target US tech firms. In spite of rising tensions and tough talk, the European Commission seems unfazed and is continuing to follow its more stringent laws, like the Digital Markets Act (DMA). This landmark piece of EU legislation aims to make the digital economy more fair. Upon coming into force last year, the act labeled certain large tech companies, including Apple and Google, as “gatekeepers” that are subject to additional scrutiny.

Europe’s more aggressive regulation of Big Tech is why iPhone users on the continent can install apps from third-party app markets while the rest of us are stuck with the Apple App Store. As for Google, the European Commission has paid special attention to search, Android, and Chrome, all of which dominate their respective markets.

Apple’s mobile platform plays second fiddle to Android in Europe, but it’s large enough to make the company subject to the DMA. The EU has now decreed that Apple is not doing enough to support interoperability on its platform. As a result, it will be required to make several notable changes. Apple will have to provide other companies and developers with improved access to iOS for devices like smartwatches, headphones, and TVs. This could include integration with notifications, faster data transfers, and streamlined setup.

The commission is also forcing Apple to release additional technical documentation, communication, and notifications for upcoming features for third parties. The EU believes this change will encourage more companies to build products that integrate with the iPhone, giving everyone more options aside from Apple’s.

Regulators say both sets of measures are the result of a public comment period that began late last year. We’ve asked Apple for comment on this development but have not heard back as of publication time. Apple is required to make these changes, and failing to do so could lead to fines. However, Google is already there.

Apple and Google in the hot seat as European regulators ignore Trump warnings Read More »

plex-ups-its-price-for-first-time-in-a-decade,-changes-remote-streaming-access

Plex ups its price for first time in a decade, changes remote-streaming access

Plex is a bit hard to explain these days. Even if you don’t know its roots as an outgrowth of a Mac port of the Xbox Media Center project, Plex is not your typical “streaming” service, given how most people use it. So as Plex announces its first price increase to its Plex Pass subscription in more than 10 years, it has its work cut out explaining why, what’s included, and what is changing.

Starting April 29, the cost of a Plex Pass rises from $5 to $7 monthly, from $40 to $70 annually, and a lifetime pass now costs $250, previously $120. In a blog post, Plex cites rising costs and its commitment to an independent service that supports “personal media.”

“We are all in on the continued success of Plex Pass and personal media,” the post states. “This price increase will ensure that we can keep investing dedicated resources in developing new features, while supporting and growing your favorites.” The post cites a roadmap that contains an integration with Common Sense Media, a new “bespoke server management app” for managing server users, and “an open and documented API for server integrations,” including custom metadata agents.

Someone in a remote video stream must have a Pass

And then, after that note, Plex hits the big change: Streaming “personal media”—i.e., video files, not audio, photos, or offerings from Plex’s ad-supported movies and TV—from outside your own network will no longer be a free Plex feature, starting April 29. “Fully free” might be the better way to put it, because if a server owner has a Plex Pass subscription, their users can still access their server for free.

But if you’ve been hosting your own Plex server to maintain access to your stuff while you’re away or relying on the kindness of non-Pass-having friends with servers, either you or your server-owning friends will need a Plex Pass subscription by the end of April.

Alternatively, you, as a non-server-running Plex viewer, can get a cheaper Remote Watch Pass for $2 per month or $20 a year. That doesn’t include Plex Pass features like offline downloads, skipping a show intro or credits, or the like, but it does keep you connected to your “personal media” vendors.

Plex ups its price for first time in a decade, changes remote-streaming access Read More »

nvidia-announces-dgx-desktop-“personal-ai-supercomputers”

Nvidia announces DGX desktop “personal AI supercomputers”

During Tuesday’s Nvidia GTX keynote, CEO Jensen Huang unveiled two “personal AI supercomputers” called DGX Spark and DGX Station, both powered by the Grace Blackwell platform. In a way, they are a new type of AI PC architecture specifically built for running neural networks, and five major PC manufacturers will build the supercomputers.

These desktop systems, first previewed as “Project DIGITS” in January, aim to bring AI capabilities to developers, researchers, and data scientists who need to prototype, fine-tune, and run large AI models locally. DGX systems can serve as standalone desktop AI labs or “bridge systems” that allow AI developers to move their models from desktops to DGX Cloud or any AI cloud infrastructure with few code changes.

Huang explained the rationale behind these new products in a news release, saying, “AI has transformed every layer of the computing stack. It stands to reason a new class of computers would emerge—designed for AI-native developers and to run AI-native applications.”

The smaller DGX Spark features the GB10 Grace Blackwell Superchip with Blackwell GPU and fifth-generation Tensor Cores, delivering up to 1,000 trillion operations per second for AI.

Meanwhile, the more powerful DGX Station includes the GB300 Grace Blackwell Ultra Desktop Superchip with 784GB of coherent memory and the ConnectX-8 SuperNIC supporting networking speeds up to 800Gb/s.

The DGX architecture serves as a prototype that other manufacturers can produce. Asus, Dell, HP, and Lenovo will develop and sell both DGX systems, with DGX Spark reservations opening today and DGX Station expected later in 2025. Additional manufacturing partners for the DGX Station include BOXX, Lambda, and Supermicro, with systems expected to be available later this year.

Since the systems will be manufactured by different companies, Nvidia did not mention pricing for the units. However, in January, Nvidia mentioned that the base-level configuration for a DGX Spark-like computer would retail for around $3,000.

Nvidia announces DGX desktop “personal AI supercomputers” Read More »

nvidia-announces-“rubin-ultra”-and-“feynman”-ai-chips-for-2027-and-2028

Nvidia announces “Rubin Ultra” and “Feynman” AI chips for 2027 and 2028

On Tuesday at Nvidia’s GTC 2025 conference in San Jose, California, CEO Jensen Huang revealed several new AI-accelerating GPUs the company plans to release over the coming months and years. He also revealed more specifications about previously announced chips.

The centerpiece announcement was Vera Rubin, first teased at Computex 2024 and now scheduled for release in the second half of 2026. This GPU, named after a famous astronomer, will feature tens of terabytes of memory and comes with a custom Nvidia-designed CPU called Vera.

According to Nvidia, Vera Rubin will deliver significant performance improvements over its predecessor, Grace Blackwell, particularly for AI training and inference.

Specifications for Vera Rubin, presented by Jensen Huang during his GTC 2025 keynote.

Specifications for Vera Rubin, presented by Jensen Huang during his GTC 2025 keynote.

Vera Rubin features two GPUs together on one die that deliver 50 petaflops of FP4 inference performance per chip. When configured in a full NVL144 rack, the system delivers 3.6 exaflops of FP4 inference compute—3.3 times more than Blackwell Ultra’s 1.1 exaflops in a similar rack configuration.

The Vera CPU features 88 custom ARM cores with 176 threads connected to Rubin GPUs via a high-speed 1.8 TB/s NVLink interface.

Huang also announced Rubin Ultra, which will follow in the second half of 2027. Rubin Ultra will use the NVL576 rack configuration and feature individual GPUs with four reticle-sized dies, delivering 100 petaflops of FP4 precision (a 4-bit floating-point format used for representing and processing numbers within AI models) per chip.

At the rack level, Rubin Ultra will provide 15 exaflops of FP4 inference compute and 5 exaflops of FP8 training performance—about four times more powerful than the Rubin NVL144 configuration. Each Rubin Ultra GPU will include 1TB of HBM4e memory, with the complete rack containing 365TB of fast memory.

Nvidia announces “Rubin Ultra” and “Feynman” AI chips for 2027 and 2028 Read More »