2nd Order Thinkers
2nd Order Thinkers.
When OpenAI and Anthropic Walked Into a PE Firm; Musk Found a New Toy.
0:00
-5:40

When OpenAI and Anthropic Walked Into a PE Firm; Musk Found a New Toy.

This week, the business models shifted. So did the goalposts.

1. OpenAI Wants to Be Your PE Firm’s AI Guy

OpenAI has raised $4 billion from 19 private equity firms and consultancies to start a new company staffed with ‘Forward-deployed engineers’. The job: help PE portfolio companies integrate AI (and to land themselves more b2b users)

Bloomberg

By coincidence, or not so much coincidence, Anthropic also announced the exact same joint venture with friendly private equity funds. It’s funny how two AI labs announced the same business strategy on the same day.

I’m writing a deep dive on this, out Sunday. Keep an eye on it!

This independent newsletter relies on sponsorship; consider becoming a paid subscriber if you appreciate the work.


2. The U.S. Government Can Now Stress-Test AI Before It Ships

A Commerce Department agency (CAISI) has signed agreements with Google DeepMind, Microsoft, and xAI, giving the federal government access to frontier AI models for national security testing before public release.

NIST

In addition to its existing partnerships with OpenAI and Anthropic, the Commerce Department Agency will also review the homework of other AI labs. Do they really have the capability to do it, or is this just a stamp on a piece of paper?

Leave a comment


3. Is xAI the Next CoreWeave?

SpaceX announced its xAI unit has struck a deal to supply computing capacity to Anthropic, following a similar deal with another company weeks earlier. xAI is quietly becoming an infrastructure play.

Anthropic

Elon Musk is like running a food truck that started with American, pivoted to Chinese, added Indian, and is now opening a French à la carte counter.

Two readings:

  1. Either xAI, the model is quietly becoming a dead end, and this is the pivot.

  2. Or he looked at CoreWeave’s business and decided the middleman margin was better than the frontier model race. Neither is reassuring for Grok.

By the way, I’m sorry you have to see this. Or am I?

4. AI Models Have Invented a Private Shorthand — and It’s 11x Cheaper

Instead of reasoning out loud in English before answering, a new training method teaches AI models to think in abstract tokens, a kind of internal shorthand the model develops for itself. As an end user, you still get a normal answer at the end. The result: 11x fewer reasoning tokens used during inference, with accuracy matching or beating standard verbal reasoning.

→ https://arxiv.org/abs/2604.22709

AI optimization is moving fast enough that today’s breakthrough is tomorrow’s baseline. Whether abstract reasoning tokens are the answer or just another stepping stone, this is positive in many ways, in saving energy and the mess created by sourcing lands for data-centers, for one.


5. Anthropic Wants to Replace Junior Analysts?

Anthropic has shipped ready-to-run Claude agent templates aimed at financial services teams: building pitchbooks and DCF models in Excel, drafting credit memos and PowerPoint decks, screening KYC files, and closing the books at month-end autonomously.

Anthropic

The media has reached another climax over the idea that junior jobs are f*cked by AI companies. However, based on what I read in the system card, even Anthropic’s latest model, Mythos, would still have a serious amount of hallucinations.

But keep in mind, being 80% correct in finance, 100% that gets you fired.

So is AI there yet?

Not quite.

Sharing is caring. If you care about solving the right problem, share this.

Share


6. 1 in 1,300 Claude Chats Is Messing With Your Head

Anthropic’s own research found that roughly 1 in every 1,300 Claude conversations distorts users’ grip on reality: nudging beliefs, reinforcing delusions, or blurring the line between the model’s framing and the user’s own thinking.

Anthropic

1 in 1,300 seems small, right?

Not at all.

Claude has 19 million monthly active users. That’s at least 14,500 people per month whose sense of reality is swayed by chatbots.

Free stuff is great. But quality analysis requires sanity. Sanity requires groceries…

Discussion about this episode

User's avatar

Ready for more?