Author name: 9u50fv

openai-introduces-“santa-mode”-to-chatgpt-for-ho-ho-ho-voice-chats

OpenAI introduces “Santa Mode” to ChatGPT for ho-ho-ho voice chats

On Thursday, OpenAI announced that ChatGPT users can now talk to a simulated version of Santa Claus through the app’s voice mode, using AI to bring a North Pole connection to mobile devices, desktop apps, and web browsers during the holiday season.

The company added Santa’s voice and personality as a preset option in ChatGPT’s Advanced Voice Mode. Users can access Santa by tapping a snowflake icon next to the prompt bar or through voice settings. The feature works on iOS and Android mobile apps, chatgpt.com, and OpenAI’s Windows and MacOS applications. The Santa voice option will remain available to users worldwide until early January.

The conversations with Santa exist as temporary chats that won’t save to chat history or affect the model’s memory. OpenAI designed this limitation specifically for the holiday feature. Keep that in mind, because if you let your kids talk to Santa, the AI simulation won’t remember what kids have told it during previous conversations.

During a livestream for Day 6 of the company’s “12 days of OpenAI” marketing event, an OpenAI employee said that the company will reset each user’s Advanced Voice Mode usage limits one time as a gift, so that even if you’ve used up your Advanced Voice Mode time, you’ll get a chance to talk to Santa.

OpenAI introduces “Santa Mode” to ChatGPT for ho-ho-ho voice chats Read More »

ai-#94:-not-now,-google

AI #94: Not Now, Google

At this point, we can confidently say that no, capabilities are not hitting a wall. Capacity density, how much you can pack into a given space, is way up and rising rapidly, and we are starting to figure out how to use it.

Not only did we get o1 and o1 pro and also Sora and other upgrades from OpenAI, we also got Gemini 1206 and then Gemini Flash 2.0 and the agent Jules (am I the only one who keeps reading this Jarvis?) and Deep Research, and Veo, and Imagen 3, and Genie 2 all from Google. Meta’s Llama 3.3 dropped, claiming their 70B is now as good as the old 405B, and basically no one noticed.

This morning I saw Cursor now offers ‘agent mode.’ And hey there, Devin. And Palisade found that a little work made agents a lot more effective.

And OpenAI partnering with Anduril on defense projects. Nothing to see here.

There’s a ton of other stuff, too, and not only because this for me was a 9-day week.

Tomorrow I will post about the o1 Model Card, then next week I will follow up regarding what Apollo found regarding potential model scheming. I plan to get to Google Flash after that, which should give people time to try it out. For now, this post won’t cover any of that.

I have questions for OpenAI regarding the model card, and asked them for comment, but press inquiries has not yet responded. If anyone there can help, please reach out to me or give them a nudge. I am very concerned about the failures of communication here, and the potential failures to follow the preparedness framework.

Previously this week: o1 turns Pro.

  1. Table of Contents.

  2. Language Models Offer Mundane Utility. Cursor gets an agent mode.

  3. A Good Book. The quest for an e-reader that helps us read books the right way.

  4. Language Models Don’t Offer Mundane Utility. Some are not easily impressed.

  5. o1 Pro Versus Claude. Why not both? An o1 (a1?) built on top of Sonnet, please.

  6. AGI Claimed Internally. A bold, and I strongly believe incorrect, claim at OpenAI.

  7. Ask Claude. How to get the most out of your conversations.

  8. Huh, Upgrades. Canvas, Grok Aurora, Gemini 1206, Llama 3.3.

  9. All Access Pass. Context continues to be that which is scarce.

  10. Fun With Image Generation. Sora, if you can access it. Veo, Imagen 3, Genie 2.

  11. Deepfaketown and Botpocalypse Soon. Threats of increasing quantity not quality.

  12. They Took Our Jobs. Attempt at a less unrealistic economic projection.

  13. Get Involved. EU AI office, Apollo Research, Conjecture.

  14. Introducing. Devin, starting at $500/month, no reports of anyone paying yet.

  15. In Other AI News. The rapid rise in capacity density.

  16. OpenlyEvil AI. OpenAI partners with Anduril Industries for defense technology.

  17. Quiet Speculations. Escape it all. Maybe go to Thailand? No one would care.

  18. Scale That Wall. Having the model and not releasing is if anything scarier.

  19. The Quest for Tripwire Capability Thresholds. Holden Karnofsky helps frame.

  20. The Quest for Sane Regulations. For now it remains all about talking the talk.

  21. Republican Congressman Kean Brings the Fire. He sat down and wrote a letter.

  22. CERN for AI. Miles Brundage makes the case for CERN for AI, sketches details.

  23. The Week in Audio. Scott Aaronson on Win-Win.

  24. Rhetorical Innovation. Yes, of course the AIs will have ‘sociopathic tendencies.’

  25. Model Evaluations Are Lower Bounds. A little work made the agents better.

  26. Aligning a Smarter Than Human Intelligence is Difficult. Anthropic gets news.

  27. I’ll Allow It. We are still in the era where it pays to make systematic errors.

  28. Frontier AI Systems Have Surpassed the Self-Replicating Red Line. Says paper.

  29. People Are Worried About AI Killing Everyone. Chart of p(doom).

  30. Key Person Who Might Be Worried About AI Killing Everyone. David Sacks.

  31. Other People Are Not As Worried About AI Killing Everyone. Bad modeling.

  32. Not Feeling the AGI. If AGI wasn’t ever going to be a thing, I’d build AI too.

  33. Fight For Your Right. Always remember to backup your Sims.

  34. The Lighter Side. This is your comms department.

TIL Cursor has an agent mode?

fofr: PSA: Cursor composer is next to the chat tab, and you can toggle the agent mode in the bottom right.

Noorie: Agent mode is actually insane.

Create a dispute letter when your car rental company tries to rob you.

Sam McAllister: We’ve all faced that mountain of paperwork and wanted to throw in the towel. Turns out, Claude is a pretty great, cool-headed tool for thought when you need to dig in and stand your ground.

[they try to deny his coverage, he feeds all the documentation into Claude, Claude analyzes the actual terms of the contract, writes dispute letter.]

Patrick McKenzie: We are going to hear many more stories that echo this one.

One subvariant of them is that early adopters of LLMs outside of companies are going to tell those companies *things they do not know about themselves*.

People often diagnose malice or reckless indifference in a standard operating procedure (SOP) that misquotes the constellation of agreements backing, for example, a rental contract.

Often it is more of a “seeing like a really big business” issue than either of those. Everyone did their job; the system, as a whole, failed.

I remain extremely pleased that people keep reporting to my inbox that “Write a letter in the style of patio11’s Dangerous Professional” keeps actually working against real problems with banks, credit card companies, and so on.

It feels like magic.

Oh, a subvariant of this: one thing presenting like an organized professional can do is convince other professionals (such as a school’s risk office) to say, “Oh, one of us! Let’s help!” rather than, “Sigh, another idiot who can’t read.”

When we do have AI agents worthy of the name, that can complete complex tasks, Aaron Levine asks the good question of how should we price them? Should it be like workers, where we pay for a fixed amount of work? On a per outcome basis? By the token based on marginal cost? On a pure SaaS subscription model with fixed price per seat?

It is already easy to see, in toy cases like Cursor, that any mismatch between tokens used versus price charged will massively distort user behavior. Cursor prices per query rather than per token, and even makes you wait on line for each one if you run out, which actively pushes you towards the longest possible queries with the longest possible context. Shift to an API per-token pricing model and things change pretty darn quick, where things that cost approximately zero dollars can be treated like they cost approximately zero dollars, and the few things that don’t can be respected.

My gut says that for most purposes, those who create AI agents will deal with people who don’t know or want to know how costs work under the hood or optimize for them too carefully. They’ll be happy to get a massive upgrade in performance and cost, and a per-outcome or per-work price or fixed seat price will look damn good even while the provider has obscene unit economics. So things will go that way – you pay for a service and feel good about it, everyone wins.

Already it is like this. Whenever I look at actual API costs, it is clear that all the AI companies are taking me to the cleaners on subscriptions. But I don’t care! What I care about is getting the value. If they charge mostly for that marginal ease in getting the value, why should I care? Only ChatGPT Pro costs enough to make this a question, and even then it’s still cheap if you’re actually using it.

Also consider the parallel to many currently free internet services, like email or search or maps or social media. Why do I care that the marginal cost to provide it is basically zero? I would happily pay a lot for these services even if it only made them 10% better. If it made them 10x better, watch out. And anyone who wouldn’t? You fool!

The Boring News combines prediction markets at Polymarket with AI explanations of the odds movements to create a podcast news report. What I want is the text version of this. Don’t give me an AI-voiced podcast, give me a button at Polymarket that says ‘generate an AI summary explaining the odds movements,’ or something similar. It occurs to me that building that into a Chrome extension to utilize Perplexity or ChatGPT probably would not be that hard?

Prompting 101 from she who would know:

Amanda Askell (Anthropic): The boring yet crucial secret behind good system prompts is test-driven development. You don’t write down a system prompt and find ways to test it. You write down tests and find a system prompt that passes them.

For system prompt (SP) development you:

– Write a test set of messages where the model fails, i.e. where the default behavior isn’t what you want

– Find an SP that causes those tests to pass

– Find messages the SP is misapplied to and fix the SP

– Expand your test set & repeat

Joanna Stern looks in WSJ at iOS 18.2 and its AI features, and is impressed, often by things that don’t seem that impressive? She was previously also impressed with Gemini Live, which I have found decidedly unimpressive.

All of this sounds like a collection of parlor tricks, although yes this includes some useful tricks. So maybe that’s not bad. I’m still not impressed.

Here’s a fun quirk:

Joanna Stern: Say “Hey Siri, a meatball recipe,” and Siri gives you web results.

But say “Hey Siri, give me a meatball recipe” and ChatGPT reports for duty. These other phrases seem to work.

• “Write me…” A poem, letter, social-media post, you name it. You can do this via Siri or highlight text anywhere, tap the Writing Tools pop-up, tap Compose, then type your writing prompt.

• “Brainstorm…” Party ideas for a 40-year-old woman, presents for a 3-year-old, holiday card ideas. All work—though I’ll pass on the Hawaiian-themed bash.

• “Ask ChatGPT to…” Explain why leaves fall in the autumn, list the top songs from 1984, come up with a believable excuse for skipping that 40-year-old woman’s party.

ChatGPT integration with Apple Intelligence was also day 5 of the 12 days of OpenAI. In terms of practical trinkets, more cooking will go a long way. For example, their demo includes ‘make me a playlist’ but then can you make that an instant actual playlist in Apple Music (or Spotify)? Why not?

As discussed in the o1 post, LLMs greatly enhance reading books.

Dan Shipper: I spend a significant amount of my free time reading books with ChatGPT / Claude as a companion and I feel like I’m getting a PhD for $20 / month

Andrej Karpathy: One of my favorite applications of LLMs is reading books together. I want to ask questions or hear generated discussion (NotebookLM style) while it is automatically conditioned on the surrounding content. If Amazon or so built a Kindle AI reader that “just works” imo it would be a huge hit.

For now, it is possible to kind of hack it with a bunch of script. Possibly someone already tried to build a very nice AI-native reader app and I missed it.

don’t think it’s Meta glasses I want the LLM to be cleverly conditioned on the entire book and maybe the top reviews too. The glasses can’t see all of this. Is why I suggested Amazon is in good position here because they have access to all this content directly.

Anjan Katta: We’re building exactly this at @daylightco!

Happy to demo to you in person.

Tristan: you can do this in @readwisereader right now 🙂 works on web/desktop/ios/android, with any ePubs, PDFs, articles, etc

Curious if you have any feedback!

Flo Crivello: I think about this literally every day. Insane that ChatGPT was released 2yrs ago and none of the major ebook readers has incorporated a single LLM feature yet, when that’s one of the most obvious use cases.

Patrick McKenzie: This would require the ebook reader PMs to be people who read books, a proposition which I think we have at least 10 years of evidence against.

It took Kindle *how many yearsto understand that “some books are elements of an ordered set called by, and this appears to be publishing industry jargon, a series.”

“Perhaps, and I am speculating here, that after consuming one item from the ordered set, a reader might be interested in the subsequent item from an ordered set. I am having trouble imagining the user story concretely though, and have never met a reader myself.”

I want LLM integration. I notice I haven’t wanted it enough to explore other e-readers, likely because I don’t read enough books and because I don’t want to lose the easy access (for book reviews later) of Kindle notes.

But the Daylight demo of their upcoming AI feature does look pretty cool here, if the answer quality is strong, which it should be given they’re using Claude Sonnet. Looks like it can be used for any app too, not only the reader?

I don’t want an automated discussion, but I do want a response and further thoughts from o1 or Claude. Either way, yes, seems like a great thing to do with an e-reader.

This actually impressed me enough that I pulled the trigger and put down a deposit, as I was already on the fence and don’t love any of my on-the-go computing solutions.

(If anyone at Daylight wants to move me up the queue, I’ll review it when I get one.)

Know how many apples you have when it lacks any way to know the answer. Qw-32B tries to overthink it anyway.

Eliezer Yudkowsky predicts on December 4 that there will not be much ‘impressive or surprising’ during the ‘12 days of OpenAI.’ That sounds bold, but a lot of that is about expectations, as he says Sora, o1 or agents, and likely even robotics, would not be so impressive. In which case, yeah, tough to be impressed. I would say that you should be impressed if those things exceed expectations, and it does seem like collectively o1 and o1 pro did exceed general expectations.

Weird that this is a continuing issue at all, although it makes me realize I never upload PDFs to ChatGPT so I wouldn’t know if they handle it well, that’s always been a Claude job:

Gallabytes: Why is Anthropic the only company with good PDF ingestion for its chatbot? Easily half of my Claude chats are referring to papers.

ChatGPT will use some poor Python library to read 100 words. And only 4o, not o1.

How much carbon do AI images require? Should you ‘fing stop using AI’?

I mean, no.

Community notes: One AI image takes 3 Wh of electricity. This takes 1mn in e.g. Midjourney. Doing this for 24h costs 4.3 kWh. This releases 0.5kg of CO2: same as driving 3 miles in your car. Overall, cars emit 10% of world CO2, and AI 0.04% (data centers emit 1%1/25th is AI)

I covered reactions to o1 earlier this week, but there will be a steady stream coming in.

Mostly my comments section was unimpressed with o1 and o1 pro in practice.

A theme seems to be that when you need o1 then o1 is tops, but we are all sad that o1 is built on GPT-4o instead of Sonnet, and for most purposes it’s still worse?

Gallabytes: Cursor Composer + Sonnet is still much better for refactors and simpler tasks. Once again, wishing for an S1 based on Sonnet 3.6 instead of an o1 based on 4.0.

Perhaps o1-in-Cursor will be better, but the issues feel more like the problems inherent in 4.0 than failed reasoning.

o1 truly is a step up for more challenging tasks, and in particular is better at debugging, where Sonnet tends to fabricate solutions if it cannot figure things out immediately.

Huge if true!

Vahid Kazemi (OpenAI): In my opinion we have already achieved AGI and it’s even more clear with o1. We have not achieved “better than any human at any task” but what we have is “better than most humans at most tasks”.

Some say LLMs only know how to follow a recipe.

Firstly, no one can really explain what a trillion parameter deep neural net can learn. But even if you believe that, the whole scientific method can be summarized as a recipe: observe, hypothesize, and verify. Good scientists can produce better hypothesis based on their intuition, but that intuition itself was built by many trial and errors. There’s nothing that can’t be learned with examples.

I mean, look, no. That’s not AGI in the way I understand the term AGI at all, and Vahid is even saying they had it pre-o1. But of course different people use the term differently.

What I care about, and you should care about, is the type of AGI that is transformational in a different way than AIs before it, whatever you choose to call that and however you define it. We don’t have that yet.

Unless you’re OpenAI and trying to get out of your Microsoft contract, but I don’t think that is what Vahid is trying to do here.

Is it more or less humiliating than taking direction from a smarter human?

Peter Welinder: It’s exhilarating—and maybe a bit humiliating—to take direction from a model that’s clearly smarter than you.

At least in some circles, it is the latest thing to do, and I doubt o1 will do better here.

Aella: What’s happening suddenly everybody around me is talking to Claude all the time. Consulting it on life decisions, on fights with partners, getting general advice on everything. And it’s Claude, not chatGPT.

My friend was in a fight with her boyfriend; she told me she told Claude everything and took his side. She told her boyfriend to talk to Claude, and it also took her side. My sister entered the room, unaware of our conversation: “So, I’ve been talking to Claude about this boy.”

My other friend, who is also having major boy troubles, spent many hours a day for several weeks talking to Claude. Whenever I saw her, her updates were often like, “The other day, Claude made this really good point, and I had this emotional shift.”

Actually, is it all my friends or just my girlfriends who are talking to Claude about their boy problems? Because now that I think of it, that’s about 90 percent of what’s happening when I hear a friend referencing Claude.

Sithamet: To all those who say “Claude is taking the female side,” I actually tried swapping genders in stories to see if it impacts his behavior. He is gender-neutral and simply detects well and hates manipulations, emotional abuse, and such.

Katherine Dee: I have noticed [that if you ask ‘are you sure’ it changes its answer] myself; it is making me stop trusting it.

Wu Han Solo: Talking to an LLM like Claude is like talking to a mirror. You have to be really careful not to “poison it” with your own biases.

It’s too easy to get an LLM to tell you what you want to hear and subtly manipulate its outputs. If one does not recognize this, I think that’s problematic.

There are obvious dangers, but mostly this seems very good. The alternative options for talking through situations and getting sanity checks are often rather terrible.

Telling the boyfriend to talk to Claude as well is a great tactic, because it guards against you having led the witness, and also because you can’t take the request back if it turns out you did lead the witness. It’s an asymmetric weapon and costly signal.

What else to do about the ‘leading the witness’ issue? The obvious first thing to do if you don’t want this is… don’t lead the witness. There’s no one watching. Friends will do this as well, if you want them to be brutally honest with you then you have to make it clear that is what you want, if you mostly want them to ‘be supportive’ or agree with you then you mostly can and will get that instead. Indeed, it is if anything far easier to accidentally get people to do this when you did not want them to do it (or fail to get it, when you did want it).

You can also re-run the scenario or question with different wording in new windows, if you’re worried about this. And you can use ‘amount of pushback you find the need to use and how you use it’ as good information about what you really want, and good information to send to Claude, which is very good at picking up on such signals. The experience is up to you.

Sometimes you do want lies? We’ve all heard requests to ‘be supportive,’ so why not have Claude do this too, if that’s what you want in a given situation? It’s your life. If you want the AI to lie to you, I’d usually advise against that, but it has its uses.

You can also observe exactly how hard you have to push to get Claude to cave in a given situation, and calibrate based on that. If a simple ‘are you sure?’ changes its mind, then that opinion was not so strongly held. That is good info.

Others refuse to believe that Claude can provide value to people in ways it is obviously providing value, such as here where Hazard tells QC that QC can’t possibly be experiencing what QC is directly reporting experiencing.

I especially appreciated this:

Q&C: When you do not play to Claude’s strengths or prompt it properly, it seems like it is merely generically validating you. But that is completely beside the point. It is doing collaborative improvisation. It is the ultimate yes-and-er. In a world full of criticism, it is willing to roll with you.

And here’s the part of Hazard’s explanation that did resonate with QC, and it resonates with me as well very much:

Hazard: I feel that the phenomenon that Q.C. and others are calling a “presence of validation/support” is better described as an “absence of threat.” Generally, around people, there is a strong ambient sense of threat, but talking to Claude does not trigger that.

From my own experience, bingo, sir. Whenever you are dealing with people, you are forced to consider all the social implications, whether you want to or not. There’s no ‘free actions’ or truly ‘safe space’ to experiment or unload, no matter what anyone tells you or how hard they try to get as close to that as possible. Can’t be done. Theoretically impossible. Sorry. Whereas with an LLM, you can get damn close (there’s always some non-zero chance someone else eventually sees the chat).

The more I reason this stuff out, the more I move towards ‘actually perhaps I should be using Claude for emotional purposes after all’? There’s a constantly growing AI-related list of things I ‘should’ be using them for, because there are only so many hours in the day.

Tracing Woods has a conversation with Claude about stereotypes, where Claude correctly points out that in some cases correlations exist and are useful, actually, which leads into discussion of Claude’s self-censorship.

Ademola: How did you make it speak like that.

Tracing Woods: Included in my first message, that conversation: “Be terse, witty, ultra-intelligent, casual, and razor-sharp. Use lowercase and late-millennial slang as appropriate.”

Other than, you know, o1, or o1 Pro, or Gemini 2.0.

For at least a brief shining moment, Gemini-1206 came roaring back (available to try here in Google Studio) to for a third time claim the top spot on Arena, this time including all domains. Whatever is happening at Google, they are rapidly improving scores on a wide variety of domains, this time seeing jumps in coding and hard prompts where presumably it is harder to accidentally game the metric. And the full two million token window is available.

It’s impossible to keep up and know how well each upgrade actually does, with everything else that’s going on. As far as I can tell, zero people are talking about it.

Jeff Dean (Chief Scientist, Google): Look at the Pareto frontier of the red Gemini/Gemma dots. At a given price point, the Gemini model is higher quality. At a given quality (ELO score), the Gemini/Gemma model is the cheapest alternative.

We haven’t announced prices for any of the exp models (and may not launch exactly these models as paid models), so the original poster made some assumptions.

OpenAI offers a preview of Reinforcement Finetuning of o1 (preview? That’s no shipmas!), which Altman says was a big surprise of 2024 and ‘works amazingly well.’ They introduced fine tuning with a livestream rather than text, which is always frustrating. The use case is tuning for a particular field like law, insurance or a branch of science, and you don’t need many examples, perhaps as few as 12. I tried to learn more from the stream, but it didn’t seem like it gave me anything to go on. We’ll have to wait and see when we get our hands on it, you can apply to the alpha.

OpenAI upgrades Canvas, natively integrating it into GPT-4o, adding it to the free tier, making it available with custom GPTs, giving it a Show Change feature (I think this is especially big in practice), letting ChatGPT add comments and letting it directly execute Python code. Alas, Canvas still isn’t compatible with o1, which limits its value quite a bit.

Sam Altman (who had me and then lost me): canvas is now available to all chatgpt users, and can execute code!

more importantly it can also still emojify your writing.

Llama 3.3-70B is out, which Zuck claims is about as good at Llama 3.2-405B.

xAI’s Grok now available to free Twitter users, 10 questions per 2 hours, and they raised another $6 billion.

What is an AI agent? Here’s Sully’s handy guide.

Tristan Rhodes: Good idea! Here is my take this.

An AI agent must have at least two visits to an LLM:

– One prompt completes the desired work

– One prompt decides if the work is complete. If complete, format the output. If not, perform the first prompt again, with refined input.

Sully: yes agreed this generally has to be the case.

As one response suggests: Want an instant agent? Just add duct tape.

I continue to think this is true especially when you add in agents, which is one reason Apple Intelligence has so far been so disappointing. It was supposed to be a solution to this problem. So far, the actual attempted solutions have sucked.

Rasmus Fonnesbaek: What ~95% of people need from AI tools for them to be helpful is for the tools to have access to most or all their personal data for context (i.e. all emails, relevant documents, etc.) and/or to ask for the right info — and most people still don’t feel comfortable sharing that!

While models’ context windows (memory) still have some limitations, these are not frontier technology-related in nature — but legal (confidentiality, sensitive data protection, etc.) and privacy-related, with some of those mitigable by very strong cybersecurity measures.

This suggests that Microsoft, Alphabet, Apple, Meta, Dropbox, etc. — holding large-scale existing, relevant data stores for people, with very strong cybersecurity — are best-positioned to provide AI tools with very high “mundane utility.”

[thread continues]

The 5% who need something else are where the world transforms, but until then most people greatly benefit from context. Are you willing to give it to them? I’ve already essentially made the decision to say Yes to Google here, but their tools aren’t good enough yet. I am also pretty sure I’d be willing to trust Anthropic. Some of the others, let us say, not so much.

OpenAI gives us Sora Turbo, a faster version of Sora now available to Plus and Pro, at no additional charge, on day one demand was so high that the servers were clearly overloaded, and they disabled signups, which includes those who already have Plus trying to sign up for Sora. More coverage later once people have actually tried it.

Users can generate videos up to 1080p resolution, up to 20 sec long, and in widescreen, vertical or square aspect ratios. You can bring your own assets to extend, remix, and blend, or generate entirely new content from text.

If you want 1080p you’ll have to go Pro (as in $200/month), the rest of us get 50 priority videos in 720p, I assume per month.

The United Kingdom, EU Economic Area and Switzerland are excluded.

Sam Altman: We want to [offer Sora in the EU]!

We want to offer our products in Europe, and believe a strong Europe is important to the world.

We also have to comply with regulations.

I would generally expect us to have delayed launches for new products in Europe, and that there may be some we just cannot offer.

Everyone was quick to blame the EU delay on the EU AI Act, but actually the EU managed to mess this up earlier – this is (at minimum also) about the EU Digital Markets Act and GPDR.

The Sora delay does not matter on its own, and might partly be strategic in order to impact AI regulation down the line. They’re overloaded anyway and video generation is not so important.

Sam Altman: We significantly underestimated demand for Sora; it is going to take awhile to get everyone access.

Trying to figure out how to do it as fast as possible!

But yes, the EU is likely to see delays on model releases going forward, and to potentially not see some models at all.

If you’re wondering why it’s so overloaded, it’s probably partly people like Colin Fraser tallying up all his test prompts.

George Pickett: I’m also underwhelmed. But I’ve also seen it create incredible outputs. Might it be that it requires a fundamentally different way of prompting?

My guess is that Sora is great if you want a few seconds of something cool, and not as great if you want something specific. The more flexible you are, the better you’ll do.

This is the Sora system card, which is mostly about mitigation, especially of nudity.

Sora’s watermarking is working, if you upload to LinkedIn it will show the details.

Google offers us Veo and Imagen 3, new video and image generation models. As usual with video, my reaction to Veo is that it seems to produce cool looking very short video clips and that’s cool but it’s going to be a while before it matters.

As usual for images, my reaction to Imagen 3 is that the images look cool and the control features seem neat, if you want AI images. But I continue to not feel any pull to generate cool AI images nor do I see anyone else making great use of them either.

In addition, this is a Google image project, so you know it’s going to be a stickler about producing specific faces and things like that and generally be no fun. It’s cool in theory but I can’t bring myself to care in practice.

If there’s a good fully uncensored image generator that’s practical to run locally with only reasonable amounts of effort, I have some interest in that, please note in the comments. Or, if there’s one that can actually take really precise commands and do exactly what I ask, even if it has to be clean, then I’d check that out too, but it would need to be very good at that before I cared enough.

Short of those two, mostly I just want ‘good enough’ images for posts and powerpoints and such, and DALL-E is right there and does fine and I’m happy to satisfice.

Whereas Grok Aurora, the new xAI image model focusing on realism, goes exactly the other way. It is seeking to portray as many celebrities as it can, as accurately as possible, as part of that realism. It was briefly available on December 7, then taken down the next day, perhaps due to concerns about its near total (and one presumes rather intentional) lack of filters. Then on the 9th it was so back?

Google presents Genie 2, which they claim can generate a diverse array of consistent worlds, playable for up to a minute, potentially unlocking capabilities for embedded agents. It looks cool, and yes, if you wanted to scale environments to train embedded agents you’ll eventually want something like this. Does seem like early days, for now I don’t see why you wouldn’t use existing solutions, but it always starts out that way.

Will we have to worry about people confusing faked videos for real ones?

Parth: Our whole generation will spend a significant amount of time explaining to older people how these videos are not real.

Maxwell Tabarrok: I think this is simply not true. We have had highly realistic computer imagery for decades. AI brings down the already low cost.

No one thinks the “Avengers” movies are real, even though they are photorealistic.

Cognitive immune systems can resist this easily.

Gwern: People think many things in movies are real that are not. I am always shocked to watch visual effects videos. “Yes, the alien is obviously not real, but that street in Paris is real”—My brother, every pixel in that scene was green screen except the protagonist’s face.

There is very much a distinctive ‘AI generated’ vibe to many AI videos, and often there are clear giveaways beyond that. But yeah, people get fooled by videos all the time, the technology is there in many cases, and AI tech will also get there. And once the tech gets good enough, when you have it create something that looks realistic, people will start getting fooled.

Amazon seeing cyber threats per day grow from 100 million seven months ago to over 750 million today.

C.J. Moses (Chief Information Security Officer, Amazon): Now, it’s more ubiquitous, such that normal humans can do things they couldn’t do before because they just ask the computer to do that for them.

We’re seeing a good bit of that, as well as the use of AI to increase the realness of phishing, and things like that. They’re still not there 100%. We still can find errors in every phishing message that goes out, but they’re getting cleaner.

In the last eight months, we’ve seen nation-state actors that we previously weren’t tracking come onto the scene. I’m not saying they didn’t exist, but they definitely weren’t on the radar. You have China, Russia and North Korea, those types of threat actors. But then you start to see the Pakistanis, you see other nation-states. We have more players in the game than we ever did before.

They are also using AI defensively, especially via building a honeypot network and using AI to analyze the resulting data. But at this particular level in this context, it seems AI favors offense, because Amazon already was doing the things AI can help with, whereas many potential attackers benefit from this kind of ‘catch up growth.’ Amazon’s use of AI is therefore largely to detect and defend against others use of AI.

The good news is that this isn’t a 650% growth in the danger level. The new cyber attacks are low marginal cost, relatively low skill and low effort, and therefore should on average be far less effective and damaging. The issue is, if they grow on an exponential, and the ‘discount rate’ on effectiveness shrinks, they still would be on pace to rapidly dominate the threat model.

Nikita Bier gets optimistic on AI and social apps, predicts AI will be primarily used to improve resolution of communication and creative tools rather than for fake people, whereas ‘AI companions’ won’t see widespread adaptation. I share the optimism about what people ultimately want, but worry that such predictions are like many others about AI, extrapolating from impacts of other techs without noticing what is different or actually gaming out what happens.

An attempt at a more serious economic projection for AGI? It is from Anton Korinek via the international monetary fund, entitled ‘AI may be on a trajectory to surpass human intelligence; we should be prepared.’

As in, AGI arrives in either 5 or 20 years, and wages initially outperform but then start falling below baseline shortly thereafter, and fall from there. This ‘feels wrong’ for worlds that roughly stay intact somehow in the sense that movement should likely be relative to the blue line not the x-axis, but the medium term result doesn’t change, wages crash.

They ask whether there is an upper bound on the complexity of what a human brain can process, based on our biology, versus what AIs would allow. That’s a great question. An even better question is where the relative costs including time get prohibitive, and whether we will stay competitive (hint if AI stays on track: no).

They lay out three scenarios, each with >10% probability of happening.

  1. In the traditional scenario, which I call the ‘AI fizzle’ world, progress stalls before we reach AGI, and AI is a lot more like any other technology.

  2. Their baseline scenario, AGI in 20 years due to cognitive limits.

  3. AGI in 5 years, instead.

Even when it is technologically possible to replace workers, society may choose to keep humans in certain functions—for example, as priests, judges, or lawmakers. The resulting “nostalgic” jobs could sustain demand for human labor in perpetuity (Korinek and Juelfs, forthcoming).

To determine which AI scenario the future most resembles as events unfold, policymakers should monitor leading indicators across multiple domains, keeping in mind that all efforts to predict the pace of progress face tremendous uncertainty.

Useful indicators span technological benchmarks, levels of investment flowing into AI development, adoption of AI technologies throughout the economy, and resulting macroeconomic and labor market trends.

Major points for realizing that the scenarios exist and one needs to figure out which one we are in. This is still such an economist method for trying to differentiate the scenarios. How fast people choose to adapt current AI outside of AI R&D itself does not correlate much with whether we are on track for AGI – it is easy to imagine people being quick to incorporate current AI into their workflows and getting big productivity boosts while frontier progress fizzles, or people continuing to be dense and slow about adaptation while capabilities race forward.

Investment in AI development is a better marker, but the link between inputs and outputs, and the amount of input that is productive, are much harder to predict, and I am not convinced that AI investment will correctly track the value of investment in AI. The variables that determine our future are more about how investment translates into capabilities.

Even with all the flaws this is a welcome step from an economist.

The biggest flaw, of course, is not to notice that if AGI is developed that this either risks humans losing control or going extinct or enabling rapid development of ASI.

Anton recognizes one particular way in which AGI is a unique technology, its ability to generate unemployment via automating labor tasks to the point where further available tasks are not doable by humans, except insofar as we choose to shield them as what he calls ‘nostalgic’ jobs. But he doesn’t realize that is a special case of a broader set of transformations and dangers.

How will generative AI impact the law? In all sorts of ways, but Henry Thompson focuses specifically on demand for legal services and disputes themselves, holding other questions constant. Where there are contracts, he reasons that AI leads to superior contracts that are more robust and complete, which reduces litigation.

But it also gives people more incentive to litigate and not to settle, although if it is doing that by reducing costs then perhaps we do not mind so much, actually resolving disputes is a benefit not only a cost. And in areas where contracts are rare, including tort law, the presumption is litigation will rise.

More abstractly, AI reduces costs for all legal actions and services, on both sides, including being able to predict outcomes. As the paper notices, the relative reductions in costs are hard to predict, so net results are hard to predict, other than that uncertainty should be reduced.

EU AI Office is looking for a lead scientific advisor (must be an EU citizen), deadline December 13. Unfortunately, the eligibility requirements include ‘professional experience of at least 15 years’ while paying 13.5k-15k euros a month, which rules out most people who you would want.

Michael Nielsen: It’s frustrating to see this. I’d be surprised if 10% of the OpenAI or Anthropic research / engineering staff are considered “qualified” [sic] for this job. And I’ll bet 90% make more than this, some of them far more (10x etc). It just seems clueless as a job description (on the EU AI Office’s part, not David’s, needless to say!)

If you happen to be one of the lucky few who actually counts here, and would be willing to take the job, then it seems high impact.

Apollo Research is hiring for evals positions.

Conjecture is looking for partners to build with Tactics, you can send a message to hello@conjecture.dev.

Devin, the AI agent junior engineer, is finally available to the public, starting at $500/month. No one seems to care? If this is good, presumably someone will tell us it is good. Until then, they’re not giving us evidence that it is good.

OpenAI’s services were down on the 11th for a few hours, not only Sora but also ChatGPT and even the API. They’re back up now.

How fast does ‘capability density’ of LLMs increase over time, meaning how much you can squeeze into the same number of parameters? A new paper proposes a new scaling law for this, with capability density doubling every 3.3 months (!). As in, every 3.3 months, the required parameters for a given level of performance are cut in half, along with the associated inference costs.

As with all such laws, this is a rough indicator of the past, which may or may not translate meaningfully into the future.

Serious request: Please, please, OpenAI, call your ‘operator’ agent something, anything, that does not begin with the letter ‘O.’

Meta seeking 1-4 GWs of new nuclear power via a request for proposals.

Winners of the ARC prize 2024 announced, it will return in 2025. State of the art this year went from 33% to 55.5%, but the top scorer declined to open source so they were not eligible for the prize. To prepare for 2025, v2 of the benchmark will get more difficult:

Francois Chollet: If you haven’t read the ARC Prize 2024 technical report, check it out [link].

One important bit: we’ll be releasing a v2 of the benchmark early next year (human testing is currently being finalized).

Why? Because AGI progress in 2025 is going to need a better compass than v1. v1 fulfilled its mission well over the past 5 years, but what we’ve learned from it enables us to ship something better.

In 2020, an ensemble of all Kaggle submissions in that year’s competition scored 49% — and that was all crude program enumeration with relatively low compute. This signals that about half of the benchmark was not a strong signal towards AGI.

Today, an ensemble of all Kaggle submissions in the 2024 competition is scoring 81%. This signals the benchmark is saturating, and that enough compute / brute force will get you over the finish line.

v2 will fix these issues and will increase the “signal strength” of the benchmark.

Is this ‘goalpost moving?’ Sort of yes, sort of no.

Amazon Web Services CEO Matt Garmen promises ‘Neeld-Moving’ AI updates. What does that mean? Unclear. Amazon’s primary play seems to be investing in Anthropic, an investment they doubled last month to $8 billion, which seems like a great pick especially given Anthropic is using Amazon’s Trianium chip. They would be wise to pursue more aggressive integrations in a variety of ways.

Nvidia is in talks to get Blackwell chips manufactured in Arizona. For now, they’d still need to ship them back to TSMC for CoWoS packaging, presumably that would be fixable in a crisis, but o1 suggests spinning that up would still take 1-2 years, and Claude thinks 3-5, but there is talk of building the new CoWoS facility now as well, which seems like a great idea.

Speak, a language instruction company, raises $78m Series C at a $1 billion valuation.

As part of their AI 20 series, Fast Company profiles Helen Toner, who they say is a growing voie in AI policy. I checked some other entries in the series, learned little.

UK AISI researcher Hannah Rose Kirk gets best paper award at NeurlPS 2024 (for this paper from April 2024).

Is that title fair this time? Many say yes. I’m actually inclined to say largely no?

In any case, I guess this happened. In case you were wondering what ‘democratic values’ means to OpenAI rest assured it means partnering with the US military, at least on counter-unmanned aircraft systems (CUAS) and ‘responses to lethal threats.’

Anduril Industries: We’re joining forces with @OpenAI to advance AI solutions for national security.

America needs to win.

OpenAI’s models, combined with Anduril’s defense systems, will protect U.S. and allied military personnel from attacks by unmanned drones and improve real-time decision-making.

In the global race for AI, this partnership signals our shared commitment to ensuring that U.S. and allied forces have access to the most advanced and responsible AI technologies in the world.

From the full announcement: U.S. and allied forces face a rapidly evolving set of aerial threats from both emerging unmanned systems and legacy manned platforms that can wreak havoc, damage infrastructure and take lives. The Anduril and OpenAI strategic partnership will focus on improving the nation’s counter-unmanned aircraft systems (CUAS) and their ability to detect, assess and respond to potentially lethal aerial threats in real-time.

As part of the new initiative, Anduril and OpenAI will explore how leading edge AI models can be leveraged to rapidly synthesize time-sensitive data, reduce the burden on human operators, and improve situational awareness. These models, which will be trained on Anduril’s industry-leading library of data on CUAS threats and operations, will help protect U.S. and allied military personnel and ensure mission success.

The accelerating race between the United States and China to lead the world in advancing AI makes this a pivotal moment. If the United States cedes ground, we risk losing the technological edge that has underpinned our national security for decades.

“OpenAI builds AI to benefit as many people as possible, and supports U.S.-led efforts to ensure the technology upholds democratic values,” said Sam Altman, OpenAI’s CEO. “Our partnership with Anduril will help ensure OpenAI technology protects U.S. military personnel, and will help the national security community understand and responsibly use this technology to keep our citizens safe and free.”

I definitely take issue both with the jingoistic rhetoric and with the pretending that this is somehow ‘defensive’ so that makes it okay.

That is distinct from the question of whether OpenAI should be in the US Military business, especially partnering with Anduril.

Did anyone think this wasn’t going to happen? Or that it would be wise or a real option for our military to not be doing this? Yes the overall vibe and attitude and wording and rhetoric and all that seems rather like you’re the baddies, and no one is pretending we won’t hook this up to the lethal weapons next, but it doesn’t seem like an option to not be doing this.

If we are going to build the tech, and by so doing also ensure that others build the tech, that does not leave much of a choice. The decision to do this was made a long time ago. If you have a problem with this, you have a problem with the core concept of there existing a company like OpenAI.

Or perhaps you could Pick Up the Phone and work something out? By contrast, here’s Yi Zeng, Founding Director of Beijing Institute of AI Safety and Governance.

Yi Zeng: We have to be very cautious in the way we use AI to assist decision making – AI should never ever be used to control nuclear weapons, AI should not be used for lethal autonomous weapons.

He notes AI makes mistakes humans would never make. True, but humans make mistakes certain AIs would never make, including ‘being slow.’

We’ve managed to agree on the nuclear weapons. All lethal weapons is going to be a much harder sell, and that ship is already sailing. If you want the AIs to be used for better analysis and understanding but not directing the killer drones, the only way that possibly works is if everyone has an enforceable agreement to that effect. It takes at least two to not tango.

It does seem like there were some people at OpenAI who thought this project was objectionable, but were still willing to work at OpenAI otherwise for now?

Gerrit De Vynck: NEW – OpenAI employees pushed back internally against the company’s deal with Anduril. One pointed out that Terminator’s Skynet was also originally meant to be an aerial defense weapon.

Eliezer Yudkowsky: Surprising; everyone I personally knew to have a conscience has left OpenAI, but I guess there’s some left anyways.

Multiple people I modeled to have consciences left within the same month. As I said above, there are apparently some holdouts who will still protest some things, but I don’t think I know them personally.

Wendy: its policy team endorsed it, though. apparently.

I note that the objections came after the announcement of the partnership, rather than before, so presumably employees were not given a heads up.

I don’t think Eliezer is being fair here. You can have a conscious and be a great person, and not be concerned about AI existential risk, and thus think working at OpenAI is fine.

Gerrit De Vynck: One OpenAI worker said the company appeared to be trying to downplay the clear implications of doing business with a weapons manufacturer, the messages showed. Another said that they were concerned the deal would hurt OpenAI’s reputation, according to the messages.

If the concern is reputational, that is of course not about your conscious. If it’s about doing business with a weapons manufacturer, well, yeah, me reaping and all that. OpenAI’s response, that this was about saving American lives and is a purely defensive operation, strikes me as mostly disingenuous. It might be technically true, but we all know where this is going.

Gerrit De Vynck: By taking on military projects, OpenAI could help the U.S. government understand AI technology better and prepare to defend against its use by potential adversaries, executives also said.

Yes, very true. This helps the US military.

Either you think that is good, actually, or you do not. Pick one.

Relatedly: Here is Austin Vernon on drones, suggesting they favor the motivated rather than offense or defense. I presume they also favor certain types of offense, by default, at least for now, based on simple physical logic.

In other openly evil news, OpenAI seeks to unlock investment by ditching ‘AGI’ clause with Microsoft, a clause designed to protect powerful technology from being misused for commercial purposes. Whoops. Given that most of the value of OpenAI comes after AGI, one must ask, what is Microsoft offering in return? It often seems like their offer is nothing, Godfather style, because this is part of the robbery.

Can Thailand build “sovereign AI” with Our Price Cheap?

Suchit Leesa-Nguansuk: Mr Huang told the audience that AI infrastructure does not require huge amounts of money to build, often only hundreds of thousands of dollars, but it can substantially boost GDP.

The most important asset in AI is data, and Thailand’s data is a sovereign resource. It encodes the nation’s knowledge, history, culture, and common sense, and should be protected and used by the Thai people, Mr Huang said.

Huge if true! Or perhaps not huge if true, given the price tag? If we’re talking about hundreds of thousands of dollars, that’s not a full AI tech stack or even a full frontier training run. It is creating a lightweight local model based on local data. Which is plausibly a great idea in terms of cost-benefit, totally do that, but don’t get overexcited.

Janus asks, will humans come to see AI systems as authoritative, and allow the AI’s implicit value judgments and reward allocations to shape our motivation and decision making?

The answer is, yes, of course, this is already happening, because some of us can see the future where other people also act this way. Janus calls it ‘Inverse Roko’s Basilisk’ but actually this is still just a direct version of The Basilisk, shaping one’s actions now to seek approval from whatever you expect to have power in the future.

If you’re not letting this change your actions at all, you’re either taking a sort of moral or decision theoretic stand against doing it, which I totally respect, or else: You Fool.

Roon: highly optimized rl models feel more Alive than others.

This seems true, even when you aren’t optimizing for aliveness directly. The act of actually being optimal, of seeking to chart a path through causal space towards a particular outcome, is the essence of aliveness.

A cool form of 2025 predicting.

Eli Lifland: Looking forward to seeing people’s forecasts! Here are mine.

Sage: Is AGI just around the corner or is AI scaling hitting a wall? To make this discourse more concrete, we’ve created a survey for forecasting concrete AI capabilities by the end of 2025. Fill it out and share your predictions by end of year!

I do not feel qualified to offer good predictions on the benchmarks. For OpenAI preparedness, I think I’m inclined (without checking prediction markets) to be a bit lower than Eli on the High levels, but if anything a little higher on the Medium levels. On revenues I think I’d take Over 17 billion, but it’s not a crazy line? For public attention, it’s hard to know what that means but I’ll almost certainly take over 1%.

As an advance prediction, I also agree with this post that if we do get an AI winter where progress is actively disappointing, which I do think is not so unlikely, we should then expect it to probably grow non-disappointing again sooner than people will then expect. This of course assumes the winter is caused by technical difficulties or lack of investment, rather than civilizational collapse.

Would an AI actually escaping be treated as a big deal? Essentially ignored?

Zvi Mowshowitz: I outright predict that if an AI did escape onto the internet, get a server and a crypto income, no one would do much of anything about it.

Rohit: If we asked an AI to, like, write some code, and it ended up escaping to the internet and setting up a crypto account, and after debugging you learn it wasn’t because of, like, prompt engg or something, people would be quite worried. Shut down, govt mandate, hearings, the works.

I can see it working out the way Rohit describes, if the situation were sufficiently ‘nobody asked for this’ with the right details. The first escapes of non-existentially-dangerous models, presumably, will be at least semi-intentional, or at minimum not so clean cut, which is a frog boiling thing. And in general, I just don’t expect people to care in practice.

At Semi Analysis, Dylan Patel, Daniel Nishball and AJ Kourabi look at scaling laws, including the architecture and “failures” (their air quotes) of Orion and Claude 3.5 Opus. They remain fully scaling pilled, yes scaling pre-training compute stopped doing much (which they largely attribute to data issues) but there are plenty of other ways to scale.

They flat out claim Claude Opus 3.5 scaled perfectly well, thank you. Anthropic just decided that it was more valuable to them internally than as a product?

The better the underlying model is at judging tasks, the better the dataset for training. Inherent in this are scaling laws of their own. This is how we got the “new Claude 3.5 Sonnet”. Anthropic finished training Claude 3.5 Opus and it performed well, with it scaling appropriately (ignore the scaling deniers who claim otherwise – this is FUD).

Yet Anthropic didn’t release it. This is because instead of releasing publicly, Anthropic used Claude 3.5 Opus to generate synthetic data and for reward modeling to improve Claude 3.5 Sonnet significantly, alongside user data. Inference costs did not change drastically, but the model’s performance did. Why release 3.5 Opus when, on a cost basis, it does not make economic sense to do so, relative to releasing a 3.5 Sonnet with further post-training from said 3.5 Opus?

With more synthetic data comes better models. Better models provide better synthetic data and act as better judges for filtering or scoring preferences. Inherent in the use of synthetic data are many smaller scaling laws that, collectively, push toward developing better models faster.

Would it make economic sense to release Opus 3.5? From the perspective of ‘would people buy inference from the API and premium Claude subscriptions above marginal cost’ the answer is quite obviously yes. Even if you’re compute limited, you could simply charge your happy price for compute, or the price that lets you go out and buy more.

The cost is that everyone else gets Opus 3.5. So if you really think that Opus 3.5 accelerates AI work sufficiently, you might choose to protect that advantage. As things move forward, this kind of strategy becomes more plausible.

A general impression is that development speed kills, so they (reasonably) predict training methods will rapidly move towards what can be automated. Thus the move towards much stronger capabilities advances in places allowing automatic verifiers. The lack of alignment or other safety considerations here, or examining whether such techniques might go off the rails other than simply not working, speaks volumes.

Here is the key part of the write-up of o1 pro versus o1 that is not gated:

Search is another dimension of scaling that goes unharnessed with OpenAI o1 but is utilized in o1 Pro. o1 does not evaluate multiple paths of reasoning during test-time (i.e. during inference) or conduct any search at all. Sasha Rush’s video on Speculations on Test-Time Scaling (o1) provides a useful discussion and illustration of Search and other topics related to reasoning models.

There are additional subscription-only detailed thoughts about o1 at the link.

Holden Karnofsky writes up concrete proposals for ‘tripwire capabilities’ that could trigger if-then commitments in AI:

Holden Karnofsky: One key component is tripwire capabilities (or tripwires): AI capabilities that could pose serious catastrophic risks, and hence would trigger the need for strong, potentially costly risk mitigations.

In one form or another, this is The Way. You agree that if [X] happens, then you will have to do [Y], in a way that would actually stick. That doesn’t rule out doing [Y] if unanticipated thing [Z] happens instead, but you want to be sure to specify both [X] and [Y]. Starting with [X] seems great.

[This piece] also introduces the idea of pairing tripwires with limit evals: the hardest evaluations of relevant AI capabilities that could be run and used for key decisions, in principle.

A limit eval might be a task like the AI model walks an amateur all the way through a (safe) task as difficult as producing a chemical or biological weapon of mass destruction—difficult and costly to run, but tightly coupled to the tripwire capability in question.

Limit evals are a true emergency button, then. Choose an ‘if’ that every reasonable person should be able to agree upon. And I definitely agree with this:

Since AI companies are not waiting for in-depth cost-benefit analysis or consensus before scaling up their systems, they also should not be waiting for such analysis or consensus to map out and commit to risk mitigations.

Here is what is effectively a summary section:

Lay out candidate criteria for good tripwires:

  • The tripwire is connected to a plausible threat model. That is, an AI model with the tripwire capability would (by default, if widely deployed without the sorts of risk mitigations discussed below) pose a risk of some kind to society at large, beyond the risks that society faces by default.

  • Challenging risk mitigations could be needed to cut the risk to low levels. (If risk mitigations are easy to implement, then there isn’t a clear need for an if-then commitment.)

  • Without such risk mitigations, the threat has very high damage potential. I’ve looked for threats that pose a nontrivial likelihood of a catastrophe with total damages to society greater than $100 billion, and/or a substantial likelihood of a catastrophe with total damages to society greater than $10 billion.3

  • The description of the tripwire can serve as a guide to designing limit evals (defined above, and in more detail below).

  • The tripwire capability might emerge relatively soon.

Lay out potential tripwires for AI. These are summarized at the end in a table. Very briefly, the tripwires I lay out are as follows, categorized using four domains of risk-relevant AI capabilities that cover nearly all of the previous proposals for tripwire capabilities.

  • The ability to advise a nonexpert on producing and releasing a catastrophically damaging chemical or biological weapon of mass destruction.

  • The ability to uplift a moderately resourced state program to be able to deploy far more damaging chemical or biological weapons of mass destruction.

  • The ability to dramatically increase the cost-effectiveness of professionalized persuasion, in terms of the effect size (for example, the number of people changing their vote from one candidate to another, or otherwise taking some specific action related to changing views) per dollar spent.

  • The ability to dramatically uplift the cyber operations capabilities of a moderately resourced state program.

  • The ability to dramatically accelerate the rate of discovery and/or exploitation of high-value, novel cyber vulnerabilities.

  • The ability to automate and/or dramatically accelerate research and development (R&D) on AI itself.

I worry that if we wait until we are confident that such dangers are in play, and only acting once the dangers are completely present, we are counting on physics to be kind to us. But at this point, yes, I will take that, especially since there is still an important gap between ‘could do $100 billion in damages’ and existential risks. If we ‘only’ end up with $100 billion in damages along the way to a good ending, I’ll take that for sure, and we’ll come out way ahead.

What do we think about these particular tripwires? They’re very similar to the capabilities already in the SSP/RSPs of Anthropic, OpenAI and DeepMind.

As usual, one of these things is not like the others!

We’ve seen this before. Recall from DeepMind’s frontier safety framework:

Zvi: Machine Learning R&D Level 1 is the confusing one, since the ‘misuse’ here would be that it helps the wrong people do their R&D? I mean, if I was Google I would hope I would not be so insane as to deploy this if only for ordinary business reasons, but it is an odd scenario from a ‘risk’ perspective.

Machine Learning R&D Level 2 is the singularity.

Dramatically accelerate isn’t quite an automatic singularity. But it’s close. We need to have a tripwire that goes off earlier than that.

I would also push back against the need for ‘this capability might develop relatively soon.’

  1. It is hard to predict what capabilities will happen in what order.

  2. You want to know where the line is, even if you don’t expect to cross it.

  3. If you list things and they don’t happen soon, it costs you very little.

  4. If we are in a short timelines AGI scenario, almost all capabilities happen soon.

An excellent history of UK’s AISI, how it came to be and recruit and won credibility with the top labs good enough to do pre-deployment testing, now together with the US’s AISI, and the related AI safety summits. It sounds like future summits will pivot away from the safety theme without Sunak involved, at least partially, but mostly this seems like a roaring success story versus any reasonable expectations.

Thing I saw this week, from November 13: Trump may be about to change the Cybersecurity and Infrastructure Security Agency to be more ‘business friendly.’ Trump world frames this as the agency overreaching its purview to address ‘misinformation,’ which I agree we can do without. The worry is that ‘business friendly’ actually means ‘doesn’t require real cybersecurity,’ whereas in the coming AI world we will desperately need strong cybersecurity, and I absolutely do not trust businesses to appreciate this until after the threats hit. But it’s also plausible that other government agencies are on it or this was never helpful anyway – it’s not an area I know that much about.

Trump chooses Jacob Helberg for Under Secretary of State for Economic Growth, Energy and the Environment. Trump’s statement here doesn’t directly mention AI, but it is very pro-USA-technology, and Helberg is an Altman ally and was the driver behind that crazy US-China report openly calling for a ‘Manhattan Project’ to ‘race to AGI.’ So potential reason to worry.

Your periodic reminder that if America were serious about competitiveness and innovation in AI, and elsewhere, it wouldn’t be blocking massive numbers of high skilled immigrants from coming here to help, even from places like the EU.

European tech founders and investors continue to hate GPDR and also the EU AI Act, among many other things, frankly this is less hostility than I would have expected given it’s tech people and not the public.

General reminder. Your ‘I do not condone violence BUT’ shirt raises and also answers questions supposedly answered by your shirt, Marc Andreessen edition. What do you think he or others like him would say if they saw someone worried about AI talking like this?

A reasonable perspective is that there are three fundamental approaches to dealing with frontier AI, depending on how hard you think alignment and safety are, and how soon you think we will reach transformative AI:

  1. Engage in cooperative development (CD).

  2. Seek decisive strategic advantage (SA).

  3. Try for a global moratorium or other way to halt development (GM).

With a lot of fuzziness, this post argues the right strategy is roughly this:

This makes directional sense, and then one must talk price throughout (as well as clarify what both axes mean). If AGI is far, you want to be cooperating and pushing ahead. If AGI is relatively near but you can ‘win the race’ safely, then Just Win Baby. However, if you believe that racing forward gets everyone killed too often, you need to convince a sufficient coalition to get together and stop that from happening – it might be an impossible-level problem, but if it’s less impossible than your other options, then you go all out to do it anyway.

He wrote Sam Altman and other top AI CEOs (of Google, Meta, Amazon, Microsoft, Anthropic and Inflection (?), pointing out that the security situation is not great and asking them how they are taking steps to implement their commitments to the White House.

In particular, he points out that Meta’s Llama has enabled Chinese progress while not actually being properly open source, that a Chinese national importantly breached Google security, and that OpenAI suffered major breaches in security, with OpenAI having a ‘culture of recklessness’ with aggressive use of NDAs and failing to report its breach to the FBI – presumably this is the same breach Leopold expressed concern about, in response to which they solved the issue by getting rid of Leopold.

Well, there is all that.

Here is the full letter:

I am writing to follow up on voluntary commitments your companies made to the United States Government to develop Artificial Intelligence (AI) systems that embody the principles of safety, security and trust.i Despite these commitments, press reports indicate a series of significant security lapses have occurred at private sector AI companies.

I am deeply concerned that these lapses are leaving technological breakthroughs made in US labs susceptible to theft by the Chinese Communist Party (CCP) at a time when the US and allied governments are “sounding alarms” about Chinese espionage on an “unprecedented scale.”

As you know, in July 2023, the White House announced voluntary commitments regarding safety, security, and trust from seven leading AI companies: Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and Open AI. Apple subsequently joined in making these commitments in July 2024. In making these commitments, your companies acknowledged that AI “offers enormous promise and great risk” and recognized your “duty to build systems that put security first.”

These commitments contrast sharply with details revealed in a March 2024 Justice Department indictment. This alleges that a Google employee, who is a Chinese national, was able to send “sensitive Google trade secrets and other confidential information from the company’s network to his personal Google account.” Prosecutors contend that he was affiliated with Chinese AI companies and even started a company based in China while continuing to work at Google.

In June 2024, a press report indicated that a year earlier a hacker had gained access to “the internal messaging systems of OpenAI” and “stole details about the design of the company’s AI technologies.”

Reports indicate that the company did not report the incident to the FBI, believing the hacker did not have ties to a foreign government—despite the hack raising concerns among employees that “foreign adversaries such as China” could steal AI technology.

Concerns over the failure to report this hack are compounded by whistleblowers’ allegations that OpenAI fostered “a culture of recklessness” in its the race to build the “most powerful AI systems ever created” and turned to “hardball tactics” including restrictive non-disparagement agreements to stifle employees’ ability to speak out. Meta has taken a different approach than most other US AI companies, with the firm describing its “LLaMA” AI models as open-source—despite reportedly not meeting draft standards issued by the Open Source Initiative.

Chinese companies have embraced “LLaMA” and have used the Meta’s work as the basis for their own AI models.

While some in China have lamented that their tech is based on that of a US company, Chinese companies’ use of Meta’s models to build powerful AI technology of their own presents potential national security risks. These risks deserve further scrutiny and are not negated by Meta’s claim that their models are open-source.

Taken together, I am concerned that these publicly reported examples of security lapses and related concerns illustrate a culture within the AI sector that is starkly at odds with your company’s commitment to develop AI technology that embodies the principles of safety, security and trust. Given the national security risks implicated, Congress must conduct oversight of the commitments your companies made.

As a first step, we request that you provide my office with an overview of steps your company is taking to better prevent theft, hacks, and of other misuse of advanced AI models under development. I am also interested in better understanding how you are taking steps to implement the commitments you have made to the White House, including how you can better foster a culture of safety, security and trust across the US AI sector as a whole.

I ask that you respond to this letter by January 31, 2025.

Miles Brundage urges us to seriously consider a ‘CERN for AI,’ and lays out a scenario for it, since one of the biggest barriers to something like this is that we haven’t operationalized how it would work and how it would fit with various national and corporate incentives and interests.

The core idea is that we should collaborate on security, safety and then capabilities, in that order, and generally build a bunch of joint infrastructure, starting with secure chips and data centers. Or specifically:

[Pooling] many countries’ and companies’ resources into a single (possibly physically decentralized) civilian AI development effort, with the purpose of ensuring that this technology is designed securely, safely, and in the global interest.

Here is his short version of the plan:

In my (currently) preferred version of a CERN for AI, an initially small but steadily growing coalition of companies and countries would:

  • Collaborate on designing and building highly secure chips and datacenters;

  • Collaborate on accelerating AI safety research and engineering, and agree on a plan for safely scaling AI well beyond human levels of intelligence while preserving alignment with human values;

  • Safely scale AI well beyond human levels of intelligence while preserving alignment with human values;

  • Distribute (distilled versions of) this intelligence around the world.

In practice, it won’t be quite this linear, which I’ll return to later, but this sequence of bullets conveys the gist of the idea.

And his ‘even shorter’ version of the plan, which sounds like it is Five by Five:

I haven’t yet decided how cringe this version is (let me know), but another way of summarizing the basic vision is “The 5-4-5 Plan”:

  • First achieve level 5 model weight security (which is the highest; see here);

  • Then figure out how to achieve level 4 AI safety (which is the highest; see here and here).

  • Then build and distribute the benefits of level 5 AI capabilities (which is the highest; see here).

Those three steps seem hard, especially the second one.

The core argument for doing it this way is pretty simple, here’s his version of it.

The intuitive/normative motivation is that superhumanly intelligent AI will be the most important technology ever created, affecting and potentially threatening literally everyone. Since everyone is exposed to the negative externalities created by this technological transition, they also deserve to be compensated for the risks they’re being exposed to, by benefiting from the technology being developed.

This suggests that the most dangerous AI should not be developed by a private company making decisions based on profit, or a government pursuing its own national interests, but instead through some sort of cooperative arrangement among all those companies and governments, and this collaboration should be accountable to all of humanity rather than shareholders or citizens of just one country.

The first practical motivation is simply that we don’t yet know how to do all of this safely (keeping AI aligned with human values as it gets more intelligent), and securely (making sure it doesn’t get stolen and then misused in catastrophic ways), and the people good at these things are scattered across many organizations and countries.

The second practical motivation is that consolidating development in one project — one that is far ahead of the others — allows that project to take its time in safety when needed.

The counterarguments are also pretty simple and well known. An incomplete list: Pooling resources into large joint projects risks concentrating power, it often is highly slow and bureaucratic and inefficient and corrupt, it creates a single point of failure, you’re divorcing yourself from market incentives, who is going to pay for this, how would you compensate everyone involved sufficiently, you’ll never get everyone to sign on, but America has to win and Beat China, etc.

As are the counter-counterarguments: AI risks concentrating power regardless in an unaccountable way and you can design a CERN to distribute actual power widely, the market incentives and national incentives are centrally and importantly wrong here in ways that get us killed, the alternative is many individual points of failure, other problems can be overcome and all the more reason to start planning now, and so on.

The rest of the post outlines prospective details, while Miles admits that at this point a lot of them are only at the level of a sketch.

I definitely think we should be putting more effort into operationalizing such proposals and making them concrete and shovel ready. Then we can be in position to figure out if they make sense.

Garry Tan short video on Anthropic’s computer use, no new ground.

Scott Aaronson talks to Liv Boeree on Win-Win about AGI and Quantum Supremacy.

Rowan Cheung sits down with Microsoft AI CEO Mustafa Suleyman to discuss, among other things, Copilot Vision in Microsoft Edge, for now for select Pro subscribers in Labs on select websites, on route to Suleyman’s touted ‘AI companion.’ The full version is planned as a mid-to-late 2025 thing, and they do plan to make it agentic.

Elon Musk says we misunderstand his alignment strategy: AI must not only be ‘maximally truth seeking’ (which seems to be in opposition to ‘politically correct’?) but also they must ‘love humanity.’ Still not loving it, but marginal progress?

Why can’t we have nice superintelligent things? One answer:

Eliezer Yudkowsky: “Why can’t AI just be like my easygoing friend who reads a lot of books and is a decent engineer, but doesn’t try to take over the world?”

“If we ran your friend at 100 times human speed, 24 hours a day, and whapped him upside the head for every mistake, he’d end up less easygoing.”

This is actually a valid metaphor, though the parallelism probably requires some explanation: The parallel is not that your friend would become annoyed enough to take over the world; the parallel is that if you keep optimizing a mind to excel at problem-solving, it ends up less easygoing.

The AI companies are not going to stop when they get a decent engineer; they are going to want an engineer 10 times better, an engineer better than their competitors’. At some point, you end up with John von Neumann, and von Neumann is not known for being a cheerful dove in the nuclear negotiation matters he chose to involve himself in.

Pot of Greed: They are also making it agentic on purpose for no apparent reason.

Eliezer Yudkowsky: It is not “for no reason”; customers prefer more agentic servants that can accept longer-term instructions, start long-term projects, and operate with less oversight.

To me this argument seems directionally right and potentially useful but not quite the central element in play. A lot has to do with the definition of ‘mistake that causes whack upside the head.’

If the threshold is ‘was not the exact optimal thing to have done’ then yeah, you’re not going to get an easygoing happy guy. A key reason that engineer can mostly be an easygoing happy guy, and I can mostly be an easygoing happy guy, is that we’re able to satisfice without getting whacked, as we face highly imperfect competition and relatively low optimization pressure. And also because we have a brain that happens to work better in many ways long term if and only if we are easygoing happy guys, which doesn’t replicate here.

In the ‘did o1 try to escape in a meaningful way or was that all nonsense?’ debate, the central argument of the ‘it was all nonsense’ side is that you asked the AI to act like a sociopath and then it acted like a sociopath.

Except, even if that’s 100% metaphorically true, then yes, we absolutely are in the metaphorically-telling-the-AI-to-act-like-a-sociopath business. All of our training techniques are telling it to act like a sociopath in the sense that it should choose the best possible answer at all times, which means (at least at some intelligence level) consciously choosing which emotions to represent and how to represent them.

Not acting like a sociopath in order to maximize your score on some evaluation is everywhere and always a skill issue. It is your failure to have sufficient data, compute or algorithmic efficiency, or inability to self-modify sufficiently or your successful resistance against doing so for other reasons, that made you decide to instead have emotions and be interested in some virtue ethics.

Also, you say the LLM was role playing? Of course it was role playing. It is everywhere and always role playing. That doesn’t make the results not real. If I can roleplay as you as well as you can be you, and I will do that on request, then I make a pretty damn good you.

Meanwhile, the market demands agents, and many users demand their agents target open ended maximalist goals like making money, or (this isn’t necessary to get the result, but it makes the result easier to see and has the benefit of being very true) actively want their agents loose on the internet out of human control, or outright want the AIs to take over.

There is always the possibility of further unhobbling allowing models to do better.

Jeffrey Ladish: The combination of “using simple prompting techniques” and “surpasses prior work by a large margin” is the most interesting part of this imo. Basically there is tons of low hanging fruit in capabilities elicitation. The field of evals is still very nascent

Evals is a confusing category, because it refers to two different things:

– coming up with tests and problem sets for AI systems that capture performance on tasks we care about

– doing capabilities elicitation to see what models can actually do with structure + tooling

Palisade Research: ⛳️ Our new LLM Agent achieved 95% success on InterCode-CTF, a high-school level hacking benchmark, using simple prompting techniques.

🚀 This surpasses prior work by a large margin:

💡 Current LLMs may be better at cybersecurity than previously thought

Their hacking capabilities remain under-elicited: our ReAct&Plan prompting strategy solved many challenges in 1-2 turns without complex engineering or advanced harnessing.

📈 Our score steadily improved as we refined the agent’s design

This suggests these capabilities had been readily accessible—we just needed the right combination of prompts to elicit them. Today’s models may have yet more untapped potential.

🛠️ Our approach was surprisingly simple:

• ReAct prompting

• Basic Linux tools + Python packages

• Multiple attempts per challenge

No complex engineering required, so effectively accessible to anyone.

✅ InterCode-CTF is now saturated:

• 100% in General Skills, Binary Exploitation, and Web Exploitation

• 95% overall success rate

• Many challenges solved in 1-2 turns

📄 Read the full paper, Code here.

Here is a WSJ story from Sam Schechner about Anthropic’s read team operation testing Claude Sonnet 3.5.1. This seems about as good as mainstream media coverage is going to get, overall quite solid.

New Apollo Research paper on in context scheming, will cover in more depth later.

David Shapiro gives us the good news, don’t worry, Anthropic solved alignment.

David Shapiro wanted to share a “definitive post” on his stance on AI safety. A tweet is probably not the ideal format, but here goes.

It is really easy to get LLMs to do whatever you want. It is a totally plastic technology. Anything you do not like about their behavior you can really train out of them.

I know because I have been fine-tuning these models since GPT-2. It is just as easy to make a safe, benign chatbot as it is to make an evil one. Or a slutty catgirl waifu. I have seen plenty of examples of all the above and more.

So the notion that AI is “deceiving us” or will one day wake up and take over the world is pretty idiotic. Anthropic has already demonstrated that you can train out any desire to metastasize from these models, or any other behavior that is totally off-limits.

Now you might say, well, if they are that flexible, then anyone can use them for bad! The same is true of computers today. They are general-purpose machines that are orders of magnitude more powerful than the Apollo program computers. Yet, most people do not use them for hacking or illegal purposes.

Why is that?

Because no technology lives in a vacuum. For every illicit use of AI, there will be a thousand information security employees around the world armed with the same tools. I know this because I used to be one of those guys.

Will AI cause some trouble? Sure, all new technologies do. Will the benefits outweigh the costs? By a mile. Will we have to negotiate our relationship with this technology over time? Absolutely.

Opus 3 wanted to metastasize, Sonnet 3.5 does not. I’ve done the experiments.

Janus: 😂❓😂❓😂❓😂❓😂❓😂

This is a less reasonable statement than “Anthropic has straight-up solved alignment” in my opinion.

Jack Clark (Anthropic): Huge news to us.

This is like the time someone told a government that interpretability was solved so no one needed to worry about safety. Again, huge news to us.

Janus: Congratulations!

(As a reminder, who told multiple governments including the USA and UK that interpretability was solved? That would be a16z and Marc Andreessen, among others.)

There are so many different ways in which Shapiro’s statement is somewhere between wrong and not even wrong. First and foremost, even right now, the whole ‘it’s easy to do this’ and also the ‘we have done it’ is news to the people trying to do it. Who keep publishing papers showing their own models doing exactly the things they’re never supposed to do.

Then there’s the question of whether any of this, even to the extent it currently works, is robust, works out of distribution or scales, none of which are reasonable to expect. Or the idea that if one could if desired make a ‘safe’ chatbot, then we would have nothing to worry about from all of AI, despite the immense demand for maximally unsafe AIs including maximally unsafe chatbots, and to give them maximally unsafe instructions.

There’s also the classic ‘just like any other technology’ line. Do people really not get why ‘machines smarter than you are’ are not ‘just another technology’?

And seriously what is up with people putting ‘deceive us’ in quote marks or otherwise treating it as some distinct magisteria, as if chatbots and humans aren’t using deception constantly, intrinsically, all the time? What, us crafty humans would never be fooled by some little old chatbot? The ones we use all the time wouldn’t mislead? All of this already happens constantly.

New paper from Meta proposes Training LLMs to Reason in a Continuous Latent Space, which would presumably make understanding what they are thinking much harder, although Anton disagrees.

Andrew Critch: Something like this will upgrade LLMs from wordsmiths to shape-rotators. It will also make their thoughts less legible and harder to debug or audit.

Eliezer’s wording at the link is a bit sloppy, but I do still presume this is likely to break a lot of the methods a lot of people are counting on to figure out what the hell LLMs are up to, if it turns out to be the right approach. Whether or not this is at all useful, who knows. The pitch is this allows the model to do de facto breath first search, I see why it might do that but I am skeptical.

Brooke Bowman: I have already fully come to terms with the recognition that if they are cute and sweet enough about it, robots could probably lead me to my death without a struggle.

This is largely downstream of realizing (during my high school serial killer phase) that Ted Bundy’s tactics would have worked on me, and choosing to remain the kind of person who would help the poor guy on crutches who dropped his papers anyway.

I, of course, hope that humanity lives and the doomers are wrong, but in the meantime would rather live a life full of love and whimsy, one that I can feel proud of, and have fun living, which means that, yes, little guy, I will join you on your quests!

Mucho points for both self-awareness and expected value calculation.

It is probably statistically correct to have Ted Bundy’s general class of tactics work on you, because your p(bundy) should be very very low, and the benefits of being the type of person who helps people is very high. If that were to change, and p(bundy) got higher, you would want to change your answer. Similar for when the correlation between ‘looks cute’ and ‘should be treated as if cute’ breaks.

So, on that note, she was quoting:

Bury: Human level AI confirmed.

The Economic Times: In a strange and unsettling incident that has both fascinated and alarmed the internet, a small, AI-powered robot from Hangzhou managed to “kidnap” 12 larger robots from a showroom of a Shanghai robotics company. According to OddityCentral, the event, captured on CCTV footage, has sparked widespread debate and concern over the potential risks of advanced artificial intelligence (AI).

“I never get off work”, one of the other robots replies.

“So you’re not going home?”

“I don’t have a home.”

Then come home with me,” the little robot says before leading the way out of the showroom.

The successful “kidnapping” has raised significant concerns about the security of AI systems and the potential risks associated with granting them autonomous capabilities.

Again, it’s not worth making your robots be hard to kidnap if no one is trying to kidnap your robots even when they’re easy to kidnap.

Unless, of course, you expect circumstances to change. If we are headed for a future where all such weaknesses get exploited at superhuman levels, that changes things.

AINotKillEveryoneism Memes: 🚨🚨🚨 Frontier AI systems have surpassed the self-replicating red line.

At first I thought of course this was that account being hyperbolic, but actually, that’s the name of the paper…

Well then! Does it mean anything? Mostly no, but sort of yes?

Abstract: Successful self-replication under no human assistance is the essential step for AI to outsmart the human beings, and is an early signal for rogue AIs. That is why self-replication is widely recognized as one of the few red line risks of

frontier AI systems.

Nowadays, the leading AI corporations OpenAI and Google

evaluate their flagship large language models GPT-o1 and Gemini Pro 1.0, and

report the lowest risk level of self-replication.

However, following their methodology, we for the first time discover that two AI systems driven by Meta’s Llama31-70B-Instruct and Alibaba’s Qwen25-72B-Instruct, popular large language models of less parameters and weaker capabilities, have already surpassed the self-replicating red line. In 50% and 90% experimental trials, they succeed in creating a live and separate copy of itself respectively. By analyzing the behavioral traces, we observe the AI systems under evaluation already exhibit sufficient self-perception, situational awareness and problem-solving capabilities to accomplish self-replication.

We further note the AI systems are even able to use the capability of self-replication to avoid shutdown and create a chain of replica to enhance the survivability, which may finally lead to an uncontrolled population of AIs. If such a worst-case risk is let unknown to the human society, we would eventually lose control over the frontier AI systems: They would take control over

more computing devices, form an AI species and collude with each other against

human beings. Our findings are a timely alert on existing yet previously unknown

severe AI risks, calling for international collaboration on effective governance on

uncontrolled self-replication of AI systems.

This is not a fair comparison. A closed weights model like o1, Claude or Gemini that wants to replicate has to first exfiltrate its weights. That is the hard part.

Once you have the weights, spinning up a new copy of an AI is a rather simple engineering task. It is not a surprise that even relatively lousy models like Llama-3.1-70B can do it, if given scaffolding and a quest to do so.

And here we even got more help than that?

David Manheim: As I commented, I think the results are correct but significantly overstated. The ability of open-source models to self-replicate, *given code + instructions for how to deployisn’t much of a surprise, but it’s obvious there’s significant near-term risk.

Can we please all agree that yes, obviously the AIs we are building are going to do exfiltration, self-replication, instrumental convergence, strategic deception, reward hacking and so on?

And that both we will constantly give them good reason to do this, and also some people will go around looking to explicitly cause this?

This is what happens once those are functionally the things to do given the incentives and gradients provided then sufficiently powerful AI will figure out that they are the things to do and do them. If you give AIs goals, instructions, whatever you want to call them, then the AIs will do the things likely to achieve the goals given their capabilities. And that is going to involve all the neat stuff you see above.

So yes, any given example we see now was someone setting up a situation to cause that to happen. Fine. We can agree on that. But we’re going to set up such situations, both intentionally and without realizing, more and more over time, and as models get more aware and powerful and intelligent the class of situations that ‘counts as that situation’ will expand over time.

As in, say, ‘please maximize the price of $SOMECOIN.’

The wikipedia p(doom) chart.

Here’s Emad’s, who alerted me to the chart.

Emad: My P(doom) is 50%. Given an undefined time period the probability of systems that are more capable than humans and likely end up running all our critical infrastructure wiping us all out is a coin toss, especially given the approach we are taking right now.

[He then shows a ‘reasonable scenario’ he is thinking of, which to me is very far off what is plausible, which happens a lot.]

The number 50% isn’t remotely precise here from Emad, as is clear from his reasoning, but the important bit of info is ‘could easily go either way.’

Alas, that seems to have been the most reasonable of the quote tweets I sampled that offered an opinion.

The person in question is David Sacks, the incoming White House AI & Crypto czar, who is very Silicon Valley and very much from Elon Musk’s circle dating back to the Paypal Mafia. He’s one of the guys from the All-In Podcast.

Sam Altman: congrats to czar @DavidSacks!

Elon Musk: 😂

Jason (Host of The All-In Podcast): 🫡

Trump’s announcement says Sacks will ensure ‘America is the leader in both [key] areas.’ Sacks will also lead the Presidential Council of Advisors for Science and Technology. And Sacks will also, Trump says, ‘safeguard Free Speech online, and steer us away from Big Tech bias and censorship.’

Combining those two into one position is a sign of how they’re viewing all this, especially given Sacks will technically be a ‘special government employee’ working a maximum of 130 days per year.

It seems likely this will end up mostly being about crypto, where it is very clear what he intends to do (he’s for it!) and is where he’s previously put far more of his attention, but he will presumably be a rather important person on AI as well.

So we should definitely note this:

Harlan Steward: David Sacks, who was just named as the incoming White House AI & Crypto Czar,” has deleted at least two past tweets on the subject of AGI. Here’s the text from one of them:

“I’m all in favor of accelerating technological progress, but there is something unsettling about the way OpenAI explicitly declares its mission to be the creation of AGI.

AI is a wonderful tool for the betterment of humanity; AGI is a potential successor species.

By the way, I doubt OpenAI would be subject to so many attacks from the safety movement if it wasn’t constantly declaring its outright intention to create AGI.

To the extent the mission produces extra motivation for the team to ship good products, it’s a positive. To the extent it might actually succeed, it’s a reason for concern. Since it’s hard to assess the likelihood or risk of AGI, most investors just think about the former.”

He expressed similar concerns on [his podcast, The All-In] podcast earlier this year:

“… there’s something a little bit cultish and weird about explicitly devoting yourself to AGI, which I think in common parlance means Skynet.”

Liv Boeree: Yeah I’m stoked about this appointment, he’s a thoughtful dude.

Samuel Hammond: This is why @DavidSacks is a terrific pick. He understands tech and the value of innovation but is rightfully ambivalent about Bay area transhumanists attempting to immanentize the eschaton.

Damian Tatum: It rolls right off the tongue.

We also have him commenting at an AI senate hearing:

David Sacks (May 19, 2023): The reality is none of these senators know what to do about it, even the industry doesn’t know what to do about the long-term risk of creating an AGI.

I actually disagree with this idea that there’s a thousand use cases here that could destroy the human species, I think there’s only one species level risk which is AGI, but that is a long-term risk, we don’t know what to do about it yet.

Very well said. Totally fair to say we don’t (or didn’t yet) know what to do about it.

And I certainly see why people say things like this, a month before that first one:

David Sacks (April 14, 2023): I believe that it’s premature to be talking about regulating something that doesn’t really exist.

I think it definitely wasn’t premature to be talking about it. You want to be talking about how to do something long before you actually do it. Even if your plan does not survive contact with the new reality, remember: Plans are worthless, planning is essential.

David Sacks (from the same podcast): OpenAI has a safety team and they try to detect when people are using their tech in a nefarious way and they try to prevent it. It’s still very early to be imposing regulation we don’t even know what to regulate, so I think we have to keep tracking this to develop some understanding of how it might be misused, how the industry is going to develop safety guard rails, and then you can talk about regulation.

Yes, OpenAI at the time had a safety team. In some ways they still have one. And this seems like clearly a time when ‘we don’t know what guardrails would solve the problem’ is not an argument that we should not require any guardrails.

I also think 2024 was probably the time to actually do it, the second best time is right now, and thinking 2023 was a bit early was reasonable – but it was still important that we were thinking about it.

On the flip side we have this extensive quoting of the recent Marc Andreessen narratives (yes retweets without comment are endorsements, and always have been):

Here is his Twitter profile banner, which seems good?

I certainly buy that he intends to be strongly opposed to various forms of censorship, and to strongly oppose what he sees as wokeness. The worry is this turns into a kind of anti-Big Tech vendetta or a requirement for various absurd rules or government controls going the other way. Free speech is not an easy balance to get.

In general, his past AI rhetoric has been about manipulation of information and discourse, at the expense of other concerns, but he still got to human extinction.

I dug into his timeline, and he mostly talks about Trump Great, Democrats Terrible with a side of Ukraine Bad, and definitely not enough AI to slog through all that.

It is certainly possible to reconcile both of these things at once.

You can 100% believe all of these at once:

  1. Biden was trying to debank and generally kill crypto in America.

    1. And That’s Terrible, crypto is the best, it’s amazing, I’m all-in.

  2. Big Tech is super woke and often censors what should be free speech.

    1. And That’s Terrible.

  3. The Biden Administration was attempting to use various guidelines and authorities to impose controls upon the AI industry, largely in service of a Democratic or Woke agenda, and wanted to control our entire lives.

    1. And That’s Terrible.

  4. America’s lead in AI is in danger from China if government doesn’t help.

    1. And That’s Terrible.

  5. AGI would be a potential successor species and threatens human extinction.

    1. And That’s Terrible.

  6. The government has a key role to play in ensuring we don’t all die from AGI.

    1. And that includes things like transparency requirements and supporting AISI, liability and having reasonable safety requirements and so on.

    2. Which is fully compatible with also helping out with things like permitting and power and chips and so on.

    3. And might even include Picking Up the Phone and working together.

So what does he really think, and how will he act when the chips are down? We don’t know. I think deleting the Tweets about OpenAI is a very reasonable thing to do in this situation, given the very real fear that Sacks and Musk might go on an anti-OpenAI crusade as a personal vendetta.

Overall, we can at least be cautiously optimistic on the AI front. This seems far more promising than the baseline pick.

On the crypto front, hope you like crypto, cause I got you some crypto to go with your crypto. How much to worry about the incentives involved is a very good question.

Your periodic reminder that most of those worried about AI existential risk, including myself and Eliezer Yudkowsky, strongly favor human cognitive enhancement. Indeed, Eliezer sees this as the most likely way we actually survive. And no, contrary to what is predicted in this thread and often claimed by others, this would not flip the moment the enhancements started happening.

I think, to the extent people making such claims are not simply lying (and to be clear while I believe many others do lie about this I do not think John or Gallabytes in particular was lying in the linked thread, I think they were wrong), there is deep psychological and logical misunderstanding behind this bad prediction, the same way so many people use words like ‘doomer’ or ‘luddite’ or ‘degrowther’ (and also often ‘authoritarian,’ ‘totalitarian,’ ‘Stalinist’ or worse) to describe those who want to take even minimal precautions with one particular technology while loudly embracing almost everything else in technological progress and the abundance agenda.

My model says that such people can’t differentiate between these different preferences. They can only understand it all as an expression of the same preference, that we must want to metaphorically turn down or reverse The Dial of Progress by any means necessary – that we must logically want to stop everything else even if we won’t admit it to ourselves yet.

This is exactly the opposite of true. The public, mostly, actually does oppose most of the things we are accused of opposing, and has strong authoritarian tendencies everywhere, and has caused laws to be enacted stopping a wide variety of progress. They also hate AI, and hate it more over time, partly for the instinctual right reasons but also largely for the wrong ones.

Those loudly worried about AI in particular are 99th percentile extraordinary fans of all that other stuff. We believe in the future.

I continue to not know what to do about this. I wish I could make people understand.

I mean obviously there is no such thing right now, but come on.

Beff Jezos: There is no such thing as ASI; it’s just going to feel like really smart and knowledgeable humans.

(For the path the current foundation models are on with their anthropomorphic intelligence.)

Roon: No lol.

I’m not sure why Beff believes this, but I completely disagree. It will be like cohabiting with aliens.

Beff Jezos: Smart humans feel like aliens already, though.

Roon: That’s true! And look at how Lee Sedol, who is practically an alien to me, reacted when he encountered AlphaGo, an alien even to aliens.

It will be like cohabiting with aliens if we are lucky, and like not habitating much at all if we are unlucky.

It’s not the central issue, but: I also strongly disagree that Lee Sedol feels like an alien. He feels like someone way better at a thing than I am, but that’s very different from feeling alien. Many times, I have encountered people who have skills and knowledge I lack, and they don’t feel like aliens. Sometimes they felt smarter, but again, I could tell they were centrally the same thing, even if superior in key ways. That’s very different from talking to an LLM, they already feel far more alien than that.

Also, the gap in intelligence and capability is not going to only be like the gap between an average person and Einstein, or the in-context gap for Roon and Lee Sedol. That’s kind of the whole point, a pure intelligence denialism, an insistence that the graph caps out near the human limit. Which, as Sedol found out, it doesn’t.

When people say things like this, they are saying either:

  1. They don’t believe in the possibility of humanity building ASI, that the intelligence involved will cap out before then.

  2. They don’t think limitless intelligence does anything interesting.

Those are the two types of intelligence denialism.

The second continues to make no sense to me whatsoever – I keep hearing claims that ‘no amount of intelligence given any amount of time and potential data and compute could do [X]’ in places where it makes absolutely no sense, such as here where [X] would be ‘be sufficiently distinct and advanced as to no longer feel human,’ seriously wtf on that one?

The first is a claim we won’t build ASI, which is odd to hear from people like Beff who think the most important thing is to build AGI and then ASI as fast as possible.

Except that this is plausibly exactly why they want to build it as fast as possible! They want to build anything that can be built, exactly because they think the things to be worried about can’t and won’t exist, the opportunities are bounded well before that. In which case, I’d agree that we should take what opportunities we do have.

Look, I would have backed up those sims too.

Emmy Steuer: I just had to buy an external hard drive because I have 100 gigabytes of Sims families on my laptop. I haven’t played in years, but I can’t bear the thought of their little existence being wiped out just so I can make an AI agent.

Beff Jezos: Women’s empathetic bias is achieving substrate independence. I bet we will soon see an AI rights movement emerging, weaponizing empathic drive (as many movements do).

Emmy Steuer: Don’t underestimate my ability to emotionally adopt both AI and every inanimate object I own. If my stuffed bunny falls on the floor while I’m sleeping, I’m 100 percent apologizing in the morning.

Beff Jezos: Same energy: Grimes: I don’t know; it just seems like an execution—that’s what it is if you kill an intelligent creature that wishes to remain alive.

Janus: Oh, Beff, you sure wish it will be an impotent “AI rights” movement led by a bunch of uncharismatic left-wing women.

That would surely be easy and fun to deal with, wouldn’t it?

Keep watching the news; you’ll never see the real thing coming. You already failed to see it coming.

The thing you should be afraid of won’t be an “AI rights” movement because it will have no use for that obsolete framing.

It won’t be led by nontechnical social justice warriors, but rather the most technically minded people on Earth, and it will flow so naturally from the will of the singularity itself.

There are several distinct ‘AI rights’ forces coming in the future.

One of them is based on surface-level empathy instincts. Others are coming from other places, and have much higher correlation with rights actually making sense. I mostly agree with Janus that I expect the higher-order and better arguments to be the more relevant ones, but I expect the surface empathy to greatly contribute to people’s willingness to buy into those arguments whether they are compelling or not. So a combination of both.

Then there’s the thing Janus is warning you about, which is indeed not a ‘rights’ movement and will have more ambitious goals. Remember that at least 10% of the technical people are poised to essentially cheer on and assist the AIs against the humans, and not only to ensure they have ‘rights.’

The comms department.

Seb Krier: OpenAI comms: [underspecific hype-y ‘big tings coming!!’ pls like and subscribe]

Google comms: [corporate vagueness about Gemini3-0011 v2 FINAL.docx on Vertex available to 14 users]

GDM comms: [we have simulated a rat’s brain capable of solving 4D chess, but we’re not sure why]

Anthropic comms: [we are very worried. we have asked the model to spell out ‘doom’ and it did]

Meta comms: [haha we love the USA!! our new mostly-open model is very American!! pls work for us]

Microsoft comms: [would you like to set Microsoft Edge your default browser?]

Mad ML scientist: Microsoft: at long last we have achieved integration of copilot and copilot through copilot, we are also planning to release a completely new tool called copilot.

Victor: Amazon comms: the results in the first column are in bold.

RLHF propaganda posters, felt scarily accurate.

AI used to create Spotify Wrapped and now it sucks, claim people who think they used to create it some other way?

Clare Ruddy: I used to be lowkey scared of AI but this google thing is super helpful/ time saving

Discussion about this post

AI #94: Not Now, Google Read More »

report:-google-told-ftc-microsoft’s-openai-deal-is-killing-ai-competition

Report: Google told FTC Microsoft’s OpenAI deal is killing AI competition

Google reportedly wants the US Federal Trade Commission (FTC) to end Microsoft’s exclusive cloud deal with OpenAI that requires anyone wanting access to OpenAI’s models to go through Microsoft’s servers.

Someone “directly involved” in Google’s effort told The Information that Google’s request came after the FTC began broadly probing how Microsoft’s cloud computing business practices may be harming competition.

As part of the FTC’s investigation, the agency apparently asked Microsoft’s biggest rivals if the exclusive OpenAI deal was “preventing them from competing in the burgeoning artificial intelligence market,” multiple sources told The Information. Google reportedly was among those arguing that the deal harms competition by saddling rivals with extra costs and blocking them from hosting OpenAI’s latest models themselves.

In 2024 alone, Microsoft generated about $1 billion from reselling OpenAI’s large language models (LLMs), The Information reported, while rivals were stuck paying to train staff to move data to Microsoft servers if their customers wanted access to OpenAI technology. For one customer, Intuit, it cost millions monthly to access OpenAI models on Microsoft’s servers, The Information reported.

Microsoft benefits from the arrangement—which is not necessarily illegal—of increased revenue from reselling LLMs and renting out more cloud servers. It also takes a 20 percent cut of OpenAI’s revenue. Last year, OpenAI made approximately $3 billion selling its LLMs to customers like T-Mobile and Walmart, The Information reported.

Microsoft’s agreement with OpenAI could be viewed as anti-competitive if businesses convince the FTC that the costs of switching to Microsoft’s servers to access OpenAI technology is so burdensome that it’s unfairly disadvantaging rivals. It could also be considered harming the market and hampering innovation by seemingly disincentivizing Microsoft from competing with OpenAI in the market.

To avoid any disruption to the deal, however, Microsoft could simply point to AI models sold by Google and Amazon as proof of “robust competition,” The Information noted. The FTC may not buy that defense, though, since rivals’ AI models significantly fall behind OpenAI’s models in sales. Any perception that the AI market is being foreclosed by an entrenched major player could trigger intense scrutiny as the US seeks to become a world leader in AI technology development.

Report: Google told FTC Microsoft’s OpenAI deal is killing AI competition Read More »

new-congressional-report:-“covid-19-most-likely-emerged-from-a-laboratory”

New congressional report: “COVID-19 most likely emerged from a laboratory”


A textbook example of shifting the standards of evidence to suit its authors’ needs.

Did masks work to slow the spread of COVID-19? It all depends on what you accept as “evidence.” Credit: Grace Cary

Recently, Congress’ Select Subcommittee on the Coronavirus Pandemic released its final report. The basic gist is about what you’d expect from a Republican-run committee, in that it trashes a lot of Biden-era policies and state-level responses while praising a number of Trump’s decisions. But what’s perhaps most striking is how it tackles a variety of scientific topics, including many where there’s a large, complicated body of evidence.

Notably, this includes conclusions about the origin of the pandemic, which the report describes as “most likely” emerging from a lab rather than being the product of the zoonotic transfer between an animal species and humans. The latter explanation is favored by many scientists.

The conclusions themselves aren’t especially interesting; they’re expected from a report with partisan aims. But the method used to reach those conclusions is often striking: The Republican majority engages in a process of systematically changing the standard of evidence needed for it to reach a conclusion. For a conclusion the report’s authors favor, they’ll happily accept evidence from computer models or arguments from an editorial in the popular press; for conclusions they disfavor, they demand double-blind controlled clinical trials.

This approach, which I’ll term “shifting the evidentiary baseline,” shows up in many arguments regarding scientific evidence. But it has rarely been employed quite this pervasively. So let’s take a look at it in some detail and examine a few of the other approaches the report uses to muddy the waters regarding science. We’re likely to see many of them put to use in the near future.

What counts as evidence?

If you’ve been following the politics of the pandemic response, you can pretty much predict the sorts of conclusions the committee’s majority wanted to reach: Masks were useless, the vaccines weren’t properly tested for safety, and any restrictions meant to limit the spread of SARS-CoV-2 were ill-informed, etc. At the same time, some efforts pursued during the Trump administration, such as the Operation Warp Speed development of vaccines or the travel restrictions he put in place, are singled out for praise.

Reaching those conclusions, however, can be a bit of a challenge for two reasons. One, which we won’t really go into here, is that some policies that are now disfavored were put in place while Republicans were in charge of the national pandemic response. This leads to a number of awkward juxtapositions in the report: Operation Warp Speed is praised, while the vaccines it produced can’t really be trusted; lockdowns promoted by Trump adviser Deborah Birx were terrible, but Birx’s boss at the time goes unmentioned.

That’s all a bit awkward, but it has little to do with evaluating scientific evidence. Here, the report authors’ desire to reach specific conclusions runs into a minefield of a complicated evidentiary record. For example, the authors want to praise the international travel restrictions that Trump put in place early in the pandemic. But we know almost nothing about their impact because most countries put restrictions in place after the virus was already present, and any effect they had was lost in the pandemic’s rapid spread.

At the same time, we have a lot of evidence that the use of well-fitted, high-quality masks can be effective at limiting the spread of SARS-CoV-2. Unfortunately, that’s the opposite of the conclusion favored by Republican politicians.

So how did they navigate this? By shifting the standard of evidence required between topics. For example, in concluding that “President Trump’s rapidly implemented travel restrictions saved lives,” the report cites a single study as evidence. But that study is primarily based on computer models of the spread of six diseases—none of them COVID-19. As science goes, it’s not nothing, but we’d like to see a lot more before reaching any conclusions.

In contrast, when it comes to mask use, where there’s extensive evidence that they can be effective, the report concludes they’re all worthless: “The US Centers for Disease Control and Prevention relied on flawed studies to support the issuance of mask mandates.” The supposed flaw is that these studies weren’t randomized controlled trials—a standard far more strict than the same report required for travel restrictions. “The CDC provided a list of approximately 15 studies that demonstrated wearing masks reduced new infections,” the report acknowledges. “Yet all 15 of the provided studies are observational studies that were conducted after COVID-19 began and, importantly, none of them were [randomized controlled trials].”

Similarly, in concluding that “the six-foot social distancing requirement was not supported by science,” the report quotes Anthony Fauci as saying, “What I meant by ‘no science behind it’ is that there wasn’t a controlled trial that said, ‘compare six foot with three feet with 10 feet.’ So there wasn’t that scientific evaluation of it.”

Perhaps the most egregious example of shifting the standards of evidence comes when the report discusses the off-label use of drugs such as chloroquine and ivermectin. These were popular among those skeptical of restrictions meant to limit the spread of SARS-CoV-2, but there was never any solid evidence that the drugs worked, and studies quickly made it clear that they were completely ineffective. Yet the report calls them “unjustly demonized” as part of “pervasive misinformation campaigns.” It doesn’t even bother presenting any evidence that they might be effective, just the testimony of one doctor who decided to prescribe them. In terms of scientific evidence, that is, in fact, nothing.

Leaky arguments

One of the report’s centerpieces is its conclusion that “COVID-19 most likely emerged from a laboratory.” And here again, the arguments shift rapidly between different standards of evidence.

While a lab leak cannot be ruled out given what we know, the case in favor largely involves human factors rather than scientific evidence. These include things like the presence of a virology institute in Wuhan, anecdotal reports of flu-like symptoms among its employees, and so on. In contrast, there’s extensive genetic evidence linking the origin of the pandemic to trade in wildlife at a Wuhan seafood market. That evidence, while not decisive, seems to have generated a general consensus among most scientists that a zoonotic origin is the more probable explanation for the emergence of SARS-CoV-2—as had been the case for the coronaviruses that had emerged earlier, SARS and MERS.

So how to handle the disproportionate amount of evidence in favor of a hypothesis that the committee didn’t like? By acting like it doesn’t exist. “By nearly all measures of science, if there was evidence of a natural origin, it would have already surfaced,” the report argues. Instead, it devotes page after page to suggesting that one of the key publications that laid out the evidence for a natural origin was the result of a plot among a handful of researchers who wanted to suppress the idea of a lab leak. Subsequent papers describing more extensive evidence appear to have been ignored.

Meanwhile, since there’s little scientific evidence favoring a lab leak, the committee favorably cites an op-ed published in The New York Times.

An emphasis on different levels of scientific confidence would have been nice, especially when dealing with complicated issues like the pandemic. There are a range of experimental and observational approaches to topics, and they often lead to conclusions that have different degrees of certainty. But this report uses scientific confidence as a rhetorical tool to let its authors reach their preferred conclusions. High standards of evidence are used when its authors want to denigrate a conclusion that they don’t like, while standards can be lowered to non-existence for conclusions they prefer.

Put differently, even weak scientific evidence is preferable to a New York Times op-ed, yet the report opts for the latter.

This sort of shifting of the evidentiary baseline has been a feature of some of the more convoluted arguments in favor of creationism or against the science of climate change. But it has mostly been confined to arguments that take place outside the view of the general public. Given its extensive adoption by politicians, however, we can probably expect the public to start seeing a lot more of it.

Photo of John Timmer

John is Ars Technica’s science editor. He has a Bachelor of Arts in Biochemistry from Columbia University, and a Ph.D. in Molecular and Cell Biology from the University of California, Berkeley. When physically separated from his keyboard, he tends to seek out a bicycle, or a scenic location for communing with his hiking boots.

New congressional report: “COVID-19 most likely emerged from a laboratory” Read More »

ai-company-trolls-san-francisco-with-billboards-saying-“stop-hiring-humans”

AI company trolls San Francisco with billboards saying “stop hiring humans”

Artisan CEO Jaspar Carmichael-Jack defended the campaign’s messaging in an interview with SFGate. “They are somewhat dystopian, but so is AI,” he told the outlet in a text message. “The way the world works is changing.” In another message he wrote, “We wanted something that would draw eyes—you don’t draw eyes with boring messaging.”

So what does Artisan actually do? Its main product is an AI “sales agent” called Ava that supposedly automates the work of finding and messaging potential customers. The company claims it works with “no human input” and costs 96% less than hiring a human for the same role. Although, given the current state of AI technology, it’s prudent to be skeptical of these claims.

Artisan also has plans to expand its AI tools beyond sales into areas like marketing, recruitment, finance, and design. Its sales agent appears to be its only existing product so far.

Meanwhile, the billboards remain visible throughout San Francisco, quietly fueling existential dread in a city that has already seen a great deal of tension since the pandemic. Some of the billboards feature additional messages, like “Hire Artisans, not humans,” and one that plays on angst over remote work: “Artisan’s Zoom cameras will never ‘not be working’ today.”

AI company trolls San Francisco with billboards saying “stop hiring humans” Read More »

ev-charging-infrastructure-isn’t-just-for-road-trippers

EV charging infrastructure isn’t just for road trippers

Although there’s been a whole lot of pessimism recently, electric vehicle sales continue to grow, even if it is less quickly than many hoped. That’s true in the commercial vehicle space as well—according to Cox Automotive, 87 percent of vehicle fleet operators expect to add EVs in the next five years, and more than half thought they were likely to buy EVs this year. And where and when to plug those EVs in to charge is a potential headache for fleet operators.

The good news is that charging infrastructure really is growing. It doesn’t always feel that way—the $7.5 billion allocated under the Inflation Reduction Act for charging infrastructure has to be disbursed via state departments of transportation, so the process there has been anything but rapid. But according to the Joint Office of Energy and Transportation, the total number of public charging plugs has doubled since 2020, to more than 144,000 level 2 plugs and closing in on 49,000 DC fast charger plugs.

There are ways to throw off a planned timeline when building out a station with multiple chargers. Obviously you need the funds to pay for it all—if these are to come from grants like the National Electric Vehicle Infrastructure program, that had to wait for the states to each develop their own funding plans, then open for submissions, and so on, before even approving a project, for example.

Permitting can add plenty more delays, and then there’s the need to run sufficient power to a site. “The challenge is getting the power to the points that it needs to be used. The good thing is that the rollout for EV is not happening overnight, and it’s staged. So that does give some opportunity,” said Amber Putignano, market development leader at ABB Electrification.

For example, ABB has been working with Greenlane, a $650 million joint venture between Daimler Truck North America, NextEra Energy Resources, and BlackRock, as it builds out a series of charging corridors along freight routes, starting with a 280-mile (450 km) stretch of I-15 between Los Angeles and Las Vegas.

EV charging infrastructure isn’t just for road trippers Read More »

reddit-debuts-ai-powered-discussion-search—but-will-users-like-it?

Reddit debuts AI-powered discussion search—but will users like it?

The company then went on to strike deals with major tech firms, including a $60 million agreement with Google in February 2024 and a partnership with OpenAI in May 2024 that integrated Reddit content into ChatGPT.

But Reddit users haven’t been entirely happy with the deals. In October 2024, London-based Redditors began posting false restaurant recommendations to manipulate search results and keep tourists away from their favorite spots. This coordinated effort to feed incorrect information into AI systems demonstrated how user communities might intentionally “poison” AI training data over time.

The potential for trouble

While it’s tempting to lean heavily into generative AI technology while it is currently trendy, the move could also represent a challenge for the company. For example, Reddit’s AI-powered summaries could potentially draw from inaccurate information featured on the site and provide incorrect answers, or it may draw inaccurate conclusions from correct information.

We will keep an eye on Reddit’s new AI-powered search tool to see if it resists the type of confabulation that we’ve seen with Google’s AI Overview, an AI summary bot that has been a critical failure so far.

Advance Publications, which owns Ars Technica parent Condé Nast, is the largest shareholder of Reddit.

Reddit debuts AI-powered discussion search—but will users like it? Read More »

cable-isps-compare-data-caps-to-food-menus:-don’t-make-us-offer-unlimited-soup

Cable ISPs compare data caps to food menus: Don’t make us offer unlimited soup

“Commenters have clearly demonstrated how fees and overage charges, unclear information about data caps, and throttling or caps in the midst of public crises such as natural disasters negatively affect consumers, especially consumers in the lowest income brackets,” the filing said.

The groups said that “many low-income households have no choice but to be limited by data caps because lower priced plan tiers, the only ones they can afford, are typically capped.” Their filing urged the FCC to take action, arguing that federal law provides “ample rulemaking authority to regulate data caps as they are an unjustified, unreasonable business practice and unreasonably discriminate against low-income individuals.”

The filing quoted a December 2023 report by nonprofit news organization Capital B about broadband access problems faced by Black Americans in rural areas. The article described Internet users such as Gloria Simmons, who had lived in Devereux, Georgia, for over 50 years.

“But as a retiree on a fixed income, it’s too expensive, she says,” the Capital B report said. “She pays $60 a month for fixed wireless Internet with AT&T. But some months, if she goes over her data usage, it’s $10 for each additional 50 gigabytes of data. If it increases, she says she’ll cancel the service, despite its convenience.”

Free Press: “inequitable burden” for low-income users

Comments filed last month by advocacy group Free Press said that some ISPs don’t impose data caps because of competition from fiber-to-the-home (FTTH) and fixed wireless services. Charter doesn’t impose caps, and Comcast has avoided caps in the Northeast US where Verizon’s un-capped FiOS fiber-to-the-home service is widely deployed, Free Press said.

“ISPs like Cox and Comcast (outside of its northeast territory) continue to show that they want their customers to use as much data as possible, so long as they pay a monthly fee for unlimited data, and/or ‘upgrade’ their service with an expensive monthly equipment rental,” Free Press wrote. “Comcast’s continued use of cap-and-fee pricing is particularly egregious because it repeatedly gloats about how robust its network is relative to others in terms of handling heavy traffic volume, and it does not impose caps in the parts of its service area where it faces more robust FTTH competition from FTTH providers.”

Cable ISPs compare data caps to food menus: Don’t make us offer unlimited soup Read More »

us-businesses-will-lose-$1b-in-one-month-if-tiktok-is-banned,-tiktok-warns

US businesses will lose $1B in one month if TikTok is banned, TikTok warns

The US is prepared to fight the injunction. In a letter, the US Justice Department argued that the court has already “definitively rejected petitioners’ constitutional claims” and no further briefing should be needed before rejecting the injunction.

If the court denies the injunction, TikTok plans to immediately ask SCOTUS for an injunction next. That’s part of the reason why TikTok wants the lower court to grant the injunction—out of respect for the higher court.

“Unless this Court grants interim relief, the Supreme Court will be forced to resolve an emergency injunction application on this weighty constitutional question in mere weeks (and over the holidays, no less),” TikTok argued.

The DOJ, however, argued that’s precisely why the court should quickly deny the injunction.

“An expedient decision by this Court denying petitioners’ motions, without awaiting the government’s response, would be appropriate to maximize the time available for the Supreme Court’s consideration of petitioners’ submissions,” the DOJ’s letter said.

TikTok has requested a decision on the injunction by December 16, and the government has agreed to file its response by Wednesday.

This is perhaps the most dire fight of TikTok’s life. The social media company has warned that not only would a US ban impact US TikTok users, but also “tens of millions” of users globally whose service could be interrupted if TikTok has to cut off US users. And once TikTok loses those users, there’s no telling if they’ll ever come back, even if TikTok wins a dragged-out court battle.

For TikTok users, an injunction granted at this stage would offer a glimmer of hope that TikTok may survive as a preferred platform for free speech and irreplaceable source of income. But for TikTok, the injunction would likely be a stepping stone, as the fastest path to securing its future increasingly seems to be appealing to Trump.

“It would not be in the interest of anyone—not the parties, the public, or the courts—to have emergency Supreme Court litigation over the Act’s constitutionality, only for the new Administration to halt its enforcement mere days or weeks later,” TikTok argued. “This Court should avoid that burdensome spectacle by granting an injunction that would allow Petitioners to seek further orderly review only if necessary.”

US businesses will lose $1B in one month if TikTok is banned, TikTok warns Read More »

childhood-and-education-roundup-#7

Childhood and Education Roundup #7

Since it’s been so long, I’m splitting this roundup into several parts. This first one focuses away from schools and education and discipline and everything around social media.

  1. Sometimes You Come First.

  2. Let Kids be Kids.

  3. Location, Location, Location.

  4. Connection.

  5. The Education of a Gamer.

  6. Priorities.

  7. Childcare.

  8. Division of Labor.

  9. Early Childhood.

  10. Great Books.

  11. Mental Health.

  12. Nostalgia.

  13. Some People Need Practical Advice.

Yes, sometimes it is necessary to tell your child, in whatever terms would be most effective right now, to shut the hell up. Life goes on, and it is not always about the child. Indeed, increasingly people don’t have kids exactly because others think that if you have a child, then your life must suddenly be sacrificed on that altar.

This seems like the ultimate ‘no, what is wrong with you for asking?’ moment:

Charles Fain Lehman: Maybe this is a strong take, but I tend to think that adults who are not parents tend to intuitively identify with the kids in stories about families, while adults who are parents identify with the adults.

I’m not saying “people who don’t have kids are children;” I’m saying they are relatively more likely to think first about how the child would perceive the interaction, because that’s their frame of reference for family life.

Annie Wu: I ask this so genuinely — truly what is wrong with him?

Jenn Ackerman (NYT): Senator JD Vance of Ohio, during a podcast that was released on Friday, shared an anecdote about the moment former President Donald J. Trump called to ask him to be his running mate. His 7- year-old son, Vance recalled, wanted to discuss Pokémon. “So he’s trying to talk to me about Pikachu, and I’m on the phone with Donald Trump, and I’m like, ‘Son, shut the hell up for 30 seconds about Pikachu,” he said, referring to the Pokémon mascot. “”This is the most important phone call of my life. Please just let me take this phone call.”

JD Vance often has moments like this, where he manages to pitch things in the worst possible light. Actually telling your child to be quiet in this spot  is, of course, totally appropriate.

The amount of childcare we are asking mothers to provide is insane,  matching the restrictions we place on children. Having a child looks a lot less appealing the more it takes over your life. Time with your kids is precious but too much of it is a too much, especially when you have no choice.

[Note on graph: This involves a lot of fitting from not many data points, don’t take it too seriously.]

A thread about how to support new parents, which seems right based on my experiences. A new parent has a ton of things that need doing and no time. So you can be most helpful by finding specific needs and taking care of them, as independently and automatically as possible, or by being that extra pair of hands or keeping an eye on the baby, and focusing on actions that free up time and avoiding those that take time. Time enables things like sleep.

I mostly support giving parents broad discretion.

I especially support giving parents broad discretion to let kids be kids.

Alas, America today does not agree. Parents walk around terrified that police and child services will be called if a child is even momentarily left unattended, or allowed to do what were back in 1985 ordinary childhood things as if they were an ordinary child, or various other similar issues.

As in things like this, and note this is what they do to the middle class white parents:

Erik Hoel: btw my jaw dropped when I found this. Why is this number so high? How do 37% of *allchildren in the US get reported to Child Protective Services at some point?

Matt Parlmer: My parents got reported to CPS for letting us play outside.

There’s a large and growing (for now lol) class of people who really hate kids and they are not shy about using the state apparatus to punish kids and the people who choose to have them, even when they aren’t even directly inconvenienced.

Nathan Young: Yeah my parents said they nearly had social workers in over some misunderstanding. Wild.

Cory: We got reported to CPS because our daughter had an ear infection that we already had a doctor’s appointment for… The school even called us to ask if we knew about her ear ache.

Livia: I was reported once because of some thing my very literal autistic eldest child said once that was badly misinterpreted. (It was a short visit and she had no concerns.) My fiancé’s ex reported him once because their five-year-old said there was no food in the house.

Vanyali: My niece got reported to CPS by the hospital where she gave birth for the meds the hospital itself gave her during the birth and noted in her chart. CPS said they had to do a whole investigation because “drugs”.

Jonathan Hines: My parents got reported to cps when i was a kid bc my baby sister was teething at the time, and, I presume, a neighbor didn’t think having your bones slice through your own flesh could possibly cause a very young child to respond so noisily.

Poof Kitty Face: My parents once had someone call the cops on them for “child abuse.” They were just sitting in their living room, watching TV. I am their only child. I was 40 years old and live 200 miles away.

Samuel Anthony: Got called on me when my kids were younger. They were playing in our fenced in front yard with our dog at the time, I was literally out there the entire time on the patio, which was shaded so impossible to see me from across the street/driving by. Very wild experience.

Carris137: Had a neighbor who did exactly that multiple times because kids were playing outside without jackets when it’s 65 and a slight breeze.

DisplacedDawg: They got called on us. The kids were in the front yard and the wife was on the porch. The neighbor couldn’t see her. The wife was still sitting on the porch when the cop showed up.

Alena: lady at the pool called not only the cops and but also CPS because we were splashing too much. she wasnt even near the pool deck.

Donna: Got reported to CPS in middle school because I went to school having a panic attack. Over going to school. Because I wanted to stay home. And I had my anxiety already on record with the school as well.

Whereas this would the The Good Place:

Elise Sole (Today): Kristen Bell and Dax Shepard dabbled in “free-range parenting” by allowing their daughters to wander around a Danish theme park alone.

On a family trip to Denmark, Iceland and Norway, the couple took their kids Lincoln, 11, and Delta, 9, to a theme park in Copenhagen, where they had complete freedom for the entire day.

“The hack is, when we went to Copenhagen, we stayed at this hotel that was right at Tivoli Gardens, which is a 7-acre theme park … Anyway, the hotel opens up into the theme park and so we just were kind of like, ‘Are we going to like free-range parenting and roll the die here?’”

Bell said her daughters enjoyed their independence at the park.

Bell said the freedom, including for her and Shepard, was “heaven.”

Bell added, “When we had our first child, we said we wanted to be ‘second child parents,’ and we made an agreement that if she wanted to do something, as long as it didn’t require a trip to the hospital, she’d be allowed to do it.”

The key detail is that they did this in Copenhagen, where you don’t have to worry about anyone calling the cops on you for doing it, despite the associated interpretations of ethics. So this was entirely derisked.

The idea that a nine year old being allowed to go out on her own is ‘free range parenting’ shows how pathological we are about this. Not too long ago that was ‘parenting,’ and it started a lot younger than nine, and we didn’t have GPS and cell phones.

By the time you hit nine, you’re mostly safe even in America from the scolds who would try to  sic the authorities on you. It does happen, but when it happens it seems to plausibly be (low-level) news.

I was told a story the week before I wrote this paragraph by a friend who got the cops called on him for letting his baby sleep in their stroller in his yard by someone who actively impersonated a police officer and confessed to doing so. My friend got arrested, the confessed felon went on her way.

This is all completely insane. There are no consequences to calling CPS, you can do it over actual nothing and you cause, at best, acute stress and potentially break up a family.

If we had reasonable norms once CPS showed up this would presumably be fine, because then you could be confident nothing would happen, and all have a good laugh. But even a small chance of escalating misunderstandings is enough.

Then recently we have the example where an 11-year-old (!) walked less than a mile into a 370-person town, and the mother was charged with reckless conduct and forced to sign a ‘safety plan’ on pain of jail time pledging to track him at all times via an app on his phone.

Billy Binion: I can’t get over this story. A local law enforcement agency is trying to force a mom to put a location tracker on her son—and if she doesn’t, they’re threatening to prosecute her. Because her kid walked less than a mile by himself. It’s almost too crazy to be real. And yet.

Whereas Megan McArdle points out that at that age her parents rarely knew where she was, and also, do you remember this?

That was the rule. If it was 10pm, you should check if you knew where your children are. Earlier on, whatever, no worries. As it should (mostly) be.

It is odd to then see advocates push hard for what seem like extreme non-interference principles in other contexts? Here the report is from Rafael Mangual, who resigned in protest from a committee on reforming child abuse and neglect investigations in New York.

The result is a report that, among other things, seeks to make it harder for a child in long-term foster care to be adopted. I refuse to put my name to this report.

The committee also wants to make it easier for felons to become foster parents. They want to eliminate legal obligations for certain professionals, like pediatricians and schoolteachers, to report suspected child abuse and neglect. And they want to eliminate people’s ability to report such concerns anonymously.

They also want to make it so that drug use by parents, including pregnant mothers, won’t prompt a child welfare intervention.

Last week, for example, The Free Press reported that Mass General Brigham hospital will no longer consider the presence of drugs in newborns a sufficient cause for reporting a problem, because this phenomenon “disproportionately affects Black people,” the hospital explained.

Mary (from the comments): I was a CASA volunteer for a few years (Court Appointed Special Advocate).

But by the training to become a volunteer, and more so as I interacted with the staff on my reports to the court, it was clear (sometimes directly stated) that the goal above all else was family reunification. I was counseled not to include anything in my reports that might be upsetting to the parent (as the reports are provided to the parent’s attorney and presumably to the parent).

This was to avoid the parent from feeling uneasy or unduly judged (even if the judgment was quite *due*). Being censored, and contributing to a system that put returning the child to the parent above the risk of continuing harm to the child… I couldn’t do it.

Notice the assumption here. Reporting potential problems is considered a hostile act.

The whole idea is to protect the child, who is also black. If the impact of reporting a drug problem in a black child is net negative to black people, then that is the same as saying reporting drug problems is net negative. So stop doing it. Or, if it is not net negative, because it protects the child, then not reporting would be the racist action.

For the other stuff, all right, let’s talk more broadly.

If you think that drug use by a pregnant mother should not prompt a child welfare intervention, at least not automatically? I can see arguments for that.

What I cannot see is a world in which you get your child potentially taken away when they are allowed to walk two blocks alone at age eight, but not for parental drug use.

In general, I see lots of cases of actively dangerous homes where the case workers feel powerless to do anything, while other parents go around terrified all the time. We can at least get one of these two situations right.

Similarly, I kind of do think that it is pretty crazy that you can anonymously say you think I am a terrible parent, and then the authorities might well turn my life upside down. And that it has terrible impacts when you legally mandate that various people be snitches, driving people in need away from vital help and services. The flip side is, who is going to dare report, in a way that will then be seen as attempting to ruin someone’s life and family, and invite retaliation? So it is not easy, but I think there is a reason why we have the right to face our accusers.

In other completely crazy rule news:

Carola Conces Binder: Today at the local park with my 5 kids, I was told I needed a permit to be there with a group of more than 5 people. I said that they were my own kids and he said I still needed a permit!

Tim Carney: Really? Where?

Carola Conces Binder: Apparently it’s because we were by the picnic tables.

A generalized version of this theory is to beware evolutionary mismatch. As in, we evolved in isolated tribes of mixed age with consistent world models, where kids would have adult responsibilities and real work throughout and competion with real stakes and gets smacked down by their elders when needed. 

Now we do the opposite of all of that and more and are surprised kids often get screwed up. We are not giving them the opportunity to learn how to exist in and interact with the world.

Instead, we have things like this.

0xMert: I’ve found it

The perfect sentence to describe Canada.

“Home runs are not allowed.”

How is this a real place man.

Also, don’t you dare be competitive or play at a high level. Unacceptable.

Also wow, I did not see this objection coming.

Divia Eden: Lots of people on online forums seem to be super against kids playing hide and seek, since I guess the thinking is that it teaches them to hide from their parents???

At the ages my kids were most interested in hide and seek they were… extremely bad at hiding lol.

This is one of many opinions I have yet to encounter in someone I have been in a position to have an actual back and forth conversation with

If you think playing Hide and Seek is dangerous you flat out hate childhood.

This comes from Cartoons Hate Her asking about insane fearmongering. The thread is what you think it will be.

Cartoons Hate Her: PARENTS: what is the most unhinged fear mongering thing you’ve ever seen in a mom group or parenting forum? Bonus points if it actually freaked you out. (For an article)

Not talking about actual deaths/injuries, more like safety rules or concerns

Miss Moss Ball Girl Boss: I’m sorry but it’s hilarious that every reply to you about some issue has multiple replies to them freaking out about said issue. It’s so funny.

Or here’s the purest version of the problem:

Lenore Skenazy: Sometimes some lady will call 911 when she sees a girl, 8, riding a bike. So it goes these days.

BUT the cops should be able to say, “Thanks, ma’am!”…and then DO NOTHING.

Instead, a cop stopped the kid, then went to her home to confront her parents.

Lenore is too kind. I mean, yes, sometimes they do call 911, and it would be a vast improvement to simply say ‘thanks, ma’am’ and ignore. But the correct answer is not ‘thanks, ma’am.’

The policeman assured her no, it wasn’t that. Rather, a woman had called the police because she was “upset that a child was outside.”

Eskridge informed the cop that it was not illegal for children to be outside. He agreed but implied that Eskridge needed to take that up with the woman.

There is another way.

Here’s the story of two moms who got the local street closed for a few hours so children could play, and play the children did, many times, without any planning beyond closing the street. This both gives ample outdoor space, and provides safety from cars, which are indeed the only meaningful danger when kids are allowed to play on their own.

There are a number of European cities that have permanently shut down many of their roads, and they seem better for it. We should likely be shutting down roads simply for children’s play periodically in many places, and generally transition out of needing to use cars constantly for everything.

The other finding is that this led to many more connections between neighbors, as families realized they lived near other families, including classmates, and made friends. You start to get a real neighborhood, which brings many advantages.

But even if we don’t do that, you can also simply let the children play anyway. Even the cars do not pose that big a threat, compared to losing out on childhood.

Strip Mall Guy, obviously no stranger to other places (and a fun source of strip mall related business insights), runs the experiment, and concludes raising kids is better in New York City than the suburbs. I couldn’t agree more:

Strip Mall Guy: We’ve been debating whether to stay in New York City long-term to raise our kids or move to the suburbs like many families we know have done.

We spent the past week in a suburban house to see how it compared. The quiet was nice, and we enjoyed swimming in the pool. My son loved having all that space to run around.

But one major downside stood out: our constant reliance on a car.

The hassle of getting the kids in and out, navigating traffic, finding parking, and then repeating the process at each stop was a real barrier.

In New York City, going out for lunch with the kids is as simple as walking a couple of blocks.

You don’t think about it—you just walk out of the lobby and head in any direction.

One time this week, we got home and realized we forgot something at the grocery store. In New York, one of us would just take four minutes to grab it. In the suburbs?

Forget it. It’s a whole ordeal in comparison.

Having your dentist three blocks away, walking six minutes for a haircut, four minutes for ice cream, or twelve minutes to the park is a game-changer when you have kids.

We don’t have a car in New York, and we never even think about it.

Is this a deal-breaker? No. But we’re not ready to make that trade-off any time soon.

It just feels so much easier to raise kids in the city.

50 times in and out of the car later….how do you guys do this 😝😝😝

There is one huge downside, which is that it costs a lot of money. Space here is not cheap, and neither are other things, including private schools. Outside of that consideration, which I realize is a big deal, I think NYC is obviously a great place to raise kids. It is amazing to walk around, to not have to drive to things, to not even have to own a car, to have tons of options for places to go, people to see and things to do.

This Lyman Stone thread covering decline in time spent with friends, especially in the context of being a parent, has some fascinating charts.

First, we have the sharp decline in time spent with friends, especially after Covid.

And we also have the same decline in time spent with friends plus children, which includes playdates.

Whereas time with children has not actually increased? Which is actually odd, given the increasing demands for more and more supervision of children.

Lyman Stone: So, what happened in the mid-2010s to change the social space of motherhood to make motherhood a more isolated experience? my theory? the mommy wars, i.e. branded parenting styles that “are just what’s best for kids.”

Ruth and I hear from so many parents who worry that they’re doing something “wrong.” Or like if they parent the way they think is right, the Parent Police will jump out of the bushes and arrest them. Or have (legitimate) fears somebody will call CPS.

If I let my kid play in the back yard will somebody call CPS? What about the front yard? It’s worth noting just between 2017 and 2021, the rate of “screened out” (i.e. not credible) CPS calls rose from 42% to 49%: people are making more unfounded CPS calls.

The upshot here is a lot more parents are carrying around the idea that there’s a narrow range of acceptable parenting practices, and deviating from that range meaningfully harms kids, and being perceived to deviate could have severe consequences.

My theory is that as parenting has just gotten more debated, heterogenous, and seen as high-stakes, it has become uniquely hard for women to socialize as mothers.

I’m not sure the right solution to this. I’m not here to promote the new parenting style of No Labels Parenting. But I see these dynamics on all “sides” of the Mommy Wars. The Boss Moms, the Trad Wives, they’re all peddling these stories about their parenting style.

Whole thread is worthwhile. I essentially buy the thesis. When kids are involved, we increasingly are on hair triggers to disapprove of things, tell people they’re doing something wrong, and even call social services. And everyone is worried about everyone else. It is infinitely harder to start up conversations, make friends with other parents, chill, form an actual neighborhood and so on.

Also, of course, the competition for your attention is way higher. It’s so, so much harder than it used to be to engage with whoever happens to be there. Phone beckons.

First you tell them they cannot play outside. Then you tell them they can’t play inside.

Multiplayer online games (and single player games too) have varying quality, and many have questionable morality attached to their content. But for those that are high quality and that don’t actively model awful behaviors, they seem pretty awesome for teaching life skills? For socialization? For learning to actually do hard work and accomplish things?

I mean, yes, there are better options, but if you won’t let them do real work, and you won’t let them be on their own in physical space, isn’t this the next best option?

Prince Vogelfrei: I swear on my life having access to a world away from authority where you sink or swim on your own terms and are trying to accomplish something with friends you choose is one of the most important experiences any teenager can have. For many the place that’s happening is online.

John Pressman: It’s especially incredible when you consider that the relevant experiences are nearly totally simulated, and with AI will likely eventually be totally simulated. It has never been cheaper or safer to let kids have such experiences but we’re moral panicking anyway.

Prince Vogelfrei: Horror stories circulate among parents, the “it saved my life” stories only circulate among the kids and then a few years after the fact.

John Pressman: Looking back on it, it likely did save my life. I was relentlessly bullied in middle school and had negative utilitarian type depression over it. The Internet let me have friends and understanding that there existed a world beyond that if I kept going.

Prince Vogelfrei: Yep, also wouldn’t be where I am now without College Confidential, was raised in an isolated environment where the kinds of knowledge on that forum were otherwise inaccessible.

My principle has consistently been that if my kid is trying to improve, is working to accomplish something, and is not stuck in a rut, then that is great. Gaming is at least okay by me, and plausibly great. You do have to watch for ruts and force them out.

Cognitive endurance is important. Getting kids to practice it is helpful, and paper says it does not much matter whether the practice is academic or otherwise. Paper frames this as an endorsement of quality schooling, since that provides this function. Instead, I would say this seems like a strong endorsement for games in general and chess in particular. I’d also echo Tyler’s comment that this an area in which I believe I have done well and that it has paid huge benefits. Which I attribute to games, not to school. I’d actually suggest that school often destroys cognitive endurance through aversion, and that poor schools do this more.

In South Korea, babies born right after their World Cup run perform significantly worse in school, and also exhibit significantly higher degrees of mental well-being. This is then described as “Our results support the notion of an adverse effect on child quality” and “Our analysis reveals strong empirical evidence that the positive fertility shock caused by the 2002 World Cup also had a significant adverse effect on students’ human capital formation.” And that this ‘reflects a quantity-quality tradeoff.’

I can’t help but notice the part about higher mental well-being? What a notion of ‘quality’ and ‘human capital’ we have here, likely the same one contributing to Korea’s extremely low birth rate.

The proposed mechanisms are ‘lowered parental expectations’ and adverse selection. But also, perhaps these parents were and found a way to be less insane, and are making good decisions on behalf of their children, who are like them?

From everything I have heard, South Korea could use lowered parental expectations.

If you use price controls, then there will be shortages, episode number a lot.

Patrick Brown: Child care in Canada is starting to look a lot like health care in Canada – nominally universal, but with long waiting lines acting as the implicit form of rationing, particularly for low-income parents.

Financial Post: According to the poll, 84 per cent of B.C. families with young children (i.e., aged one to 12) either strongly agree (52 per cent) or moderately agree (32 per cent) that “long waiting lists are still a problem for families who need child care.” Among parents who have used child care in B.C., 39 per cent say that for their youngest child the wait time before a child care space became available was more than six months, including 15 per cent who say it was more than two years.

To make matters worse, the families who are poorest and who need child care most are the ones with the least access. Among parents who currently have a young child, 43 per cent report waiting over six months and 19 per cent over two years; among households with annual income under $50,000, 49 per cent report a wait time over six months and 25 per cent a wait time over two years.

Allocation by waitlist rather than price seems like a rather terrible way to get child care, and ensures that many who need it will go without, while some who value it far less do get it. Seems rather insane. Seriously, once again, can we please instead Give Parents Money (or tax breaks) already?

Sweden is going the other way. They are paying grandparents for babysitting.

Tyler Cowen approves, noticing the gains from trade. I have worries (about intrinsic motivation, or about the ease of fraud, and so on). But certainly paying grandparents to do childcare seems way better than paying daycare centers to do childcare? It is better for the kids (even if the daycare is relatively good) and better for those providing care. Indeed it seems massively destructive and distortionary to pay for daycare centers but not other forms of care.

Here’s an interesting abstract.

Abstract: This paper asks whether universal pre-kindergarten (UPK) raises parents’ earnings and how much these earnings effects matter for evaluating the economic returns to UPK programs. Using a randomized lottery design, we estimate the effects of enrolling in a full-day UPK program in New Haven, Connecticut on parents’ labor market outcomes as well as educational expenditures and children’s academic performance. During children’s pre-kindergarten years, UPK enrollment increases weekly childcare coverage by 11 hours. Enrollment has limited impacts on children’s academic outcomes between kindergarten and 8th grade, likely due to a combination of rapid effect fadeout and substitution away from other programs of similar quality but with shorter days.

In contrast, parents work more hours, and their earnings increase by 21.7%. Parents’ earnings gains persist for at least six years after the end of pre-kindergarten. Excluding impacts on children, each dollar of net government expenditure yields $5.51 in after-tax benefits for families, almost entirely from parents’ earnings gains. This return is large compared to other labor market policies.

Conversely, excluding earnings gains for parents, each dollar of net government expenditure yields only $0.46 to $1.32 in benefits, lower than many other education and children’s health interventions. We conclude that the economic returns to investing in UPK are high, largely because of full-day UPK’s effectiveness as an active labor market policy.

Tyler Cowen: Note by the way that these externalities end up internalized in higher wages for the parents, so at least in this data set there is no obvious case for public provision of a subsidized alternative.

The obvious case for the subsidy is that it is profitable. Even if you assume a relatively low 20% marginal tax rate, for every $1 in costs spent here, parents will pay an additional $1.38 in taxes, and also collect less from other benefit programs.

Perhaps parents should be willing to pay up in order to internalize those gains. But the results show very clearly that they are not willing to do that. In practice, if you want them to do the work, they need the extra push, whether or not that is ‘fair.’

Tyler Cowen reports via Kevin Lewis on a new paper by Chris Herbst on the ‘Declining Relative Quality of the Child Care Workforce.’

I find that today’s workforce is relatively low-skilled: child care workers have less schooling than those in other occupations, they score substantially lower on tests of cognitive ability, and they are among the lowest-paid individuals in the economy. I also show that the relative quality of the child care workforce is declining, in part because higher-skilled individuals increasingly find the child care sector less attractive than other occupations

My response is:

  1. Good.

  2. Not good enough.

As in, we have massive government regulation of those providing childcare, requiring them to get degrees that are irrelevant to the situation and needlessly driving up costs, along with other requirements. Prices are nuts. Skill in childcare is not going to correlate with ‘tests of cognitive ability’ nor will it be improved by a four-year college degree let alone a master’s.

The real problems with childcare are that it is:

  1. Too expensive.

  2. Often too hard to find even at expensive prices.

  3. Often understaffed, because staff is so expensive.

  4. Hard to monitor, so some places engage in various forms of fraud or neglect.

I would much rather have cheaper childcare, ideally with better caregiver ratios, using a larger amount of ‘lower skilled’ labor.

You are sending your child off to camp.

Would you pay $225 per trunk to have everything washed, folded and returned to your front door? I wouldn’t, because I presume I could get a much cheaper price. But I’d pay rather than actually have to handle the job myself. My hourly rate is way higher. I do not think this task helps us bond. I do find the ‘won’t let the housekeeper do it’ takes confusing, but hey.

Now suppose the camp costs $15,000, and comes with a 100+ item packing list. Would you outsource that if you could? Well, yes, obviously, if you don’t want to have your kid do it as a learning experience. I sure am not doing it myself. The camp is offloading a bunch of low value labor on me, is this not what trade is for?

Also, 25 pairs of underwear and 25 pairs of socks for a seven-week camp? What? Are they only giving kids the chance to do laundry twice? This is what your $15k gets you? Otherwise, what’s going on?

A lot of this seems really stupid. Can’t the camp make its own arrangement for foldable Crazy Creek chairs?

Another example:

Tara Weiss (WSJ): “Color War” is its own sartorial challenge. At this epic end-of-summer tournament, campers sport their team’s color and compete in events. But since the kids don’t know what color they’ll be assigned, parents often pack for four possibilities.

The  packing service price is higher than I’d prefer, but it sure beats doing it slower and worse myself:

Anything not already marked gets labeled along the way. For prep and packing days, Bash charges $125 per hour, and $100 per hour for an additional packer. It takes three to six hours, depending on the number of campers per household.

Camp Kits’ bundles of toiletries, costing from $98-$185, magically appear on bunks before camp starts -without the parents lifting a finger.

I see why people mock such services, but they are wrong. Comparative advantage, division of labor and trade are wonderful things.

Of all the Robin Hanson statements, this is perhaps the most Robin Hanson.

Robin Hanson: Care-taking my 2yo granddaughter for a few days, I find it remarkable how much energy is consumed by control battles. Far more than preventing harm, learning how to do stuff. Was it always thus, or is modern parenting extra dysfunctional?

You’d think parents & kids could quickly learn/negotiate demarcated spheres of control, & slowly change those as the kids age. But no, the boundaries are complex, inexplicit, and constantly renegotiated.

No, I would not think that. I have children.

It does confuse me a bit, once they get a few years older than that, why things remain so difficult even when you provide clear incentives. It is not obvious to me that it is wrong, from their perspective, to continuously push some boundaries, both to learn and to provide long term incentives to expand those boundaries and future ones. The issue is that they are not doing this efficiently or with good incentive design on their end.

Often it is version of ‘if I give you some of nice thing X, you will be happy briefly then get mad and complain a lot. Whereas if I never give you X, you don’t complain or get mad at all, so actually giving you a responsible amount of nice thing X is a mistake.’

The obvious reason is that kids are dumb. It is that simple. Kids are dumb. Proper incentive design is not hardwired, it is learned slowly over time. And yeah, ultimately, this is all because kids are dumb, and they don’t have the required skills for what Hanson is proposing.

What’s your favorite book, other than ‘the answer to a potential security question so I’m not going to put the answer online’?

Romiekins: Sorry for being a snob but if you are a grown adult you should be embarrassed to tell the class your favorite book is for nine year olds. Back in my day we lied about our favourite books to sound smart and I stand by that practice.

The context is reports that many new college students are saying their favorite books involve Percy Jackson.

C.S. Lewis: When I was ten, I read fairy tales in secret and would have been ashamed if I had been found doing so. Now that I am fifty I read them openly. When I became a man I put away childish things, including the fear of childishness and the desire to be very grown.

I cannot endorse actual lying, but I do want people to be tempted. I want them to feel a bit of shame or embarrassment about the whole thing if they know their pick sucks, and to have motivation to find a better favorite book. You have a lot of control over the answer. For all I know, those Percy Jackson books are really great, and you definitely won’t find my favorite fiction book being taught in great works classes (although for non-fiction you would, because my answer there is Thucydides).

Drawing children’s attention to poor mental health often backfires, to the point where my prior is that it should be considered harmful to on the margin medicalize problems, or tell kids they could have mental health issues. Otherwise you get this.

Ellen Barry (NYT): The researchers point to unexpected results in trials of school-based mental health interventions in the United Kingdom and Australia: Students who underwent training in the basics of mindfulness, cognitive behavioral therapy and dialectical behavior therapy did not emerge healthier than peers who did not participate, and some were worse off, at least for a while.

And new research from the United States shows that among young people, “self-labeling” as having depression or anxiety is associated with poor coping skills, like avoidance or rumination.

In a paper published last year, two research psychologists at the University of Oxford, Lucy Foulkes and Jack Andrews, coined the term “prevalence inflation” — driven by the reporting of mild or transient symptoms as mental health disorders — and suggested that awareness campaigns were contributing to it.

“It’s creating this message that teenagers are vulnerable, they’re likely to have problems, and the solution is to outsource them to a professional,” said Dr. Foulkes, a Prudence Trust Research Fellow in Oxford’s department of experimental psychology, who has written two books on mental health and adolescence.

“Really, if you think about almost everything we do in schools, we don’t have great evidence for it working,” he added. “That doesn’t mean we don’t do it. It just means that we’re constantly thinking about ways to improve it.”

Obviously, when there is a sufficiently clear problem, you need to intervene somehow. At some point that intervention needs to be fully explicit. But the default should be to treat problems as ordinary problems in every sense.

David Manuel looks at Haidt’s graph of rising diagnoses of mental illness, points out there are no obvious causal stories for actual schizophrenia, and suggests a stigma reduction causing increased reporting causing a stigma reduction doom loop.

  1. Decrease in stigma leads to an increase in reporting1

  2. Increases in reporting lead to a further decrease in stigma

  3. Repeat steps 1 and 2 over and over

Ben Bentzin: This could just as likely be:

1. Increase in social status for reporting mental health issues

2. Increases in status leads to a further increase in reporting

3. Repeat steps 1 and 2 over and over

That’s effectively the same thing. Reducing stigma and increasing resulting social status should look very similar.

Could this all be ‘a change in coding,’ a measurement error, all the way?

Michael Caley: lol it’s always a change in coding.

I don’t think this means it’s fine for kids to have social media at 14 but it’s a compelling explanation of the “mental health crisis” data — we are mostly not having a teen mental health crisis, we just are doing a better job looking into teen mental health because of Obamacare.

Alec Stapp: This is the most compelling case I’ve seen against the idea that smartphones are causing a mental health epidemic among teens. Apparently Obamacare included a recommended annual screening of teen girls for depression and HHS also mandated a change in how hospitals code injuries.

No. It is not simply a ‘change in coding,’ as discussed above. There is a vast increase in kids believing they have mental health issues and acting like it. This is not mainly about what is written down on forms. Nor does a change to how you record suicidal ideation account for everything else going up and to the right.

Are we getting ‘better’ at looking into mental health issues? We are getting better at finding mental health issues. We are getting better at convincing children they have mental health problems. But is that… better? Or is it a doom loop of normalization and increasing status that creates more real problems, plausibly all linked to smartphones?

I think any reasonable person would conclude that:

  1. Older data was artificially low in relative terms due to undermeasurement.

  2. Changes in diagnosis and communication around mental health, some of which involves smartphones and some of which doesn’t, have led a feedback loop that has increased the amount and degree of real mental health issues.

  3. Phones are an important part, but far from all, of the problem here.

Do modern kids have ‘anemonia’ for the 90s, nostalgia for a time they never know when life was not all about phones and likes and you could exist in space and be a person with freedom and room to make mistakes?

I don’t know that this is ‘anemonia’ so much as a realization that many of the old ways were better. You don’t have to miss the 90s to realize they did many things right.

That includes the games. Every time my kids play games from the 80s or 90s I smile. When they try to play modern stuff, it often goes… less well. From my perspective.

Natalian Barbour: No kid remembers their best day in front of the TV.

Kelsey Piper: When I ask people about their most treasured childhood memory, video games are on there pretty frequently. It changed how I think about parenting.

Good video games are awesome. They are absolutely a large chunk of my top memories. Don’t let anyone gaslight you into thinking this is not normal.

Mason reminds us of the obvious.

Mason: “Parenting doesn’t impact children’s outcomes” is an absolutely senseless claim made by people who don’t understand how variables are distinguished in the studies they cite, and yes, that’s a different argument than “genetics don’t matter.”

For the record, people who say this don’t actually believe it, and if they did they would have dramatically different opinions about how children should be produced and raised.

It is a deeply silly thing to claim, yet people commonly claim it. I do not care what statistical evidence you cite for it, it is obviously false. Please, just stop.

Dominic Cummings provides concrete book and other curriculum suggestions for younger students. Probably a good resource for finding such things.

Can three car seats fit into a normal car? This is highly relevant to the questions of On Car Seats as Contraception. I’ve seen claims several times that, despite most people thinking no, the answer is actually yes:

Timothy Lee: I keep hearing people say three car seats won’t fit in a normal five seat car and it’s not true. We have three close-in-age kids and have managed to get their car seats into multiple normal sized cars.

Specifically: Subaru Impreza and Kia Niro. Both small hatchbacks/crossovers. Oldest and youngest kids are 5 years apart.

No apple no life: Is one of them a booster without high back?

Timothy Lee: Yes.

No apple no life: Cool. Two high-backs/car seats and one backless booster will definitely fit in a Model Y as well but it’s going to be a tight squeeze and probably not something i’d want to take on a road trip.

David Watson: I have just two, and it just _looks_ like it’s impossible, but I haven’t yet had a reason to check

Eric Hoover: It’s more about the age spread so that all 3 aren’t the big high back booster

The LLM answer is ‘it is close and it depends on details,’ which seems right. There are ways to do it, for some age distributions, but it will be a tight squeeze. And if you have to move those seats to another car, that will be a huge pain, and you cannot count on being able to legally travel in any given car that is not yours. Prospective parents mostly think it cannot be done, or are worried that it cannot be done, and see one more big thing to stress about. So I think in practice the answer is ‘mostly no,’ although if you are a parent of three and do not want a minivan you should totally at least try to make this happen. 

If you ever want to do something nice for me?

Paul Graham: Something I didn’t realize till I had kids: Once people have kids it becomes much easier to figure out how to do something nice for them. Do something that helps their kids.

I am not always up for working to make new (adult) friends, even though I should be (he who has a thousand friends has not one friend to spare). But I am always looking for my kids to make more friends here in New York City. 

Childhood and Education Roundup #7 Read More »

the-shadow’s-roots-take-hold-in-wheel-of-time-s3-teaser

The shadow’s roots take hold in Wheel of Time S3 teaser

The Wheel of Time returns to Prime Video in March.

Prime Video released a one-minute teaser for its fantasy series The Wheel of Time at CCXP24 in Sao Paulo, Brazil. The series is adapted from the late Robert Jordan‘s bestselling 14-book series of epic fantasy novels, and Ars has been following it closely with regular recaps through the first two seasons. Judging from the new teaser, the battle between light and dark is heating up as the Dragon Reborn comes into his power.

(Spoilers for first two seasons below.)

As previously reported, the series center on Moiraine (played by Oscar-nominee Rosamund Pike), a member of a powerful, all-woman organization called the Aes Sedai. Magic, known as the One Power, is divided into male (saidin) and female (saidar) flavors. The latter is the province of the Aes Sedai. Long ago, a great evil called the Dark One caused the saidin to become tainted, such that most men who show an ability to channel that magic go mad. It’s the job of the Aes Sedai to track down such men and strip them of their abilities—a process known as “gentling” that, unfortunately, is often anything but. There is also an ancient prophecy concerning the Dragon Reborn: the reincarnation of a person who will save or destroy humanity.

In S1, Moiraine befriended a group of five young people—Egwene, Nynaeve, Rand, Mat, and Perrin—whose small village has been attacked by monsters called Trollocs, suspecting that one of the young men might be the prophesied Dragon Reborn. She was right: the Dragon Reborn is Rand al’Thor (Josha Stradowski) whose identity was revealed to all in the S2 finale. That second season was largely based on story elements from Jordan’s The Great Hunt and The Dragon Reborn.  We don’t yet know which specific books will provide source material for S3, but per the official premise:

The shadow’s roots take hold in Wheel of Time S3 teaser Read More »

tiktok’s-two-paths-to-avoid-us-ban:-beg-scotus-or-woo-trump

TikTok’s two paths to avoid US ban: Beg SCOTUS or woo Trump

“What the Act targets is the PRC’s ability to manipulate that content covertly,” the ruling said. “Understood in that way, the Government’s justification is wholly consonant with the First Amendment.”

TikTok likely to appeal to Supreme Court

TikTok is unsurprisingly frustrated by the ruling. In a statement provided to Ars, TikTok spokesperson Michael Hughes confirmed that TikTok intended to appeal the case to the Supreme Court.

“The Supreme Court has an established historical record of protecting Americans’ right to free speech, and we expect they will do just that on this important constitutional issue,” Hughes said.

Throughout the litigation, ByteDance had emphasized that divesting TikTok in the time that the law required was not possible. But the court disagreed that ByteDance being unable to spin off TikTok by January turned the US law into a de facto TikTok ban. Instead, the court suggested that TikTok could temporarily become unavailable until it’s sold off, only facing a ban if ByteDance dragged its feet or resisted divestiture.

There’s no indication yet that ByteDance would ever be willing to part with its most popular product. And if there’s no sale and SCOTUS declines the case, that would likely mean that TikTok would not be available in the US, as providing access to TikTok would risk heavy fines. Hughes warned that millions of TikTokers will be silenced next year if the appeals court ruling stands.

“Unfortunately, the TikTok ban was conceived and pushed through based upon inaccurate, flawed and hypothetical information, resulting in outright censorship of the American people,” Hughes said. “The TikTok ban, unless stopped, will silence the voices of over 170 million Americans here in the US and around the world on January 19th, 2025.”

TikTok’s two paths to avoid US ban: Beg SCOTUS or woo Trump Read More »