Feed

Join the Hivemind

[
]
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
See catalogue
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
by Pondering Durian
11.08.2025

As an (at least aspiring) high-end content producer on the interwebs, I found this episode of No Priors with Cloudflare CEO Matthew Prince to be a welcome shot of optimism in a sea of AI doomerism around content production.

The gist of the episode is that the historical social contract of the web is breaking down. Content-producers used to produce content, and Google would route them traffic which could then be monetized by ads or merch or just plain clout. Increasingly though, AI models mine the content and provide derivative answers to users without routing any traffic back to users or even attribution.

Obviously, this new status quo significantly reduces the incentives to provide new content, something which, in the long term, will also impact model performance.

Cloudflare's unique position may enable it to enforce a new status quo: one which provides more granular permissions to content creators around who can view their content and at what price (e.g. humans view for free, but bots have to pay).

This identity layer is something the crypto world has spent a lot of time thinking about with networks like World being the most notable examples, but perhaps Cloudflare's existing position at the network's edge puts it in a very unique position to enforce a new social contract on today's rails: one which creates a market for truly novel human knowledge creation online.

The old world of Google indexing the internet and routing traffic will die a slow death. We have a chance to rethink monetization online - one in which markets are likely to be a better solution for content creators as opposed to regulatory protection (unlikely to hold up under fair use).

In an optimistic scenario, this will be a market that stops rewarding dopamine and outrage and starts rewarding truly novel forms of human knowledge which, in turn, fill in the gaps of foundational models.

Like Kevin Lu's latest essay, I wonder if Cloudflare is in a position to enforce a globally, distributed version of scale AI / surge AI. Perhaps the future of the web is internet scale markets for truly novel human knowledge; something I think we can all get behind.  

This is some text inside of a div block.
by Lex
10.08.2025

The release of GPT5 sparked significant controversy, leading Sam Altman to promise the return of GPT4o during Friday's Reddit AMA (ask me anything)—a rather embarrassing situation.

While GPT5's performance and the drastic downgrade of previous models represents the primary controversy, I'm more interested in the less-discussed aspects of OpenAI's most significant release since 2023, particularly how it reflects a shift in the company's strategic focus.

First, on the developer's end, the enhancements are very agent-friendly:

  1. Tools: custom tools now support any text input (code, SQL, shell, configuration, etc.) beyond just JSON; Allowed tools mechanism lets you dynamically specify which subset of tools can be used in each round; Parallel tool invocation enables the model to use multiple tools simultaneously, greatly improving efficiency for complex tasks and multi-tasking; Preamble feature automatically generates concise explanations before tool use, improving transparency and debugging.
  2. Structured Output: Support for CFG (context-free grammar) allows developers to use lark/regex to precisely constrain output formats. Combined with strict mode, JSON output success rates have improved from 40% to over 90%. The model now strictly adheres to parameter formats, making outputs more predictable and safer.
  3. Performance/Latency/Cost Control: New "Reasoning Effort" settings (minimal/low/medium/high) let developers fine-tune the depth of model "thinking"—minimal for ultra-low latency needs, high for complex reasoning tasks. Output verbosity settings (low/medium/high) control detail levels; for code, low generates refined code while high provides detailed explanations with structured formatting.
  4. Context: Chain-of-Thought cross-round delivery enables reasoning chains to persist across conversation rounds, giving developers fine-grained control over context management and multi-step tasks.
  5. Pricing: Token caching support for high-frequency scenarios, dramatically lower API pricing (now aligned with Gemini and approaching 1/10th of Claude's rates).

These API enhancements focus on being "agent-friendly" and "developer-controllable." Experienced model API developers will recognize that while these features aren't directly related model intelligence, they address significant headaches in day-to-day engineering applications. When model intelligence itself isn't the bottleneck, these practical features often become decisive factors in choosing which AI system to use.

Second is the ecosystem. Few may have noticed that Cursor released CLI mode on the second day after GPT5's launch, quickly integrating with GPT5 in response to Claude Code's impact. Frequent users recognize that Claude Code's true killer feature is its $200 max subscription that provides thousands of dollars worth of tokens. Now that Anthropic can no longer afford such subsidies and has begun limiting usage, the affordable GPT5 API paired with Cursor CLI offers an immediate counterattack just as the competition pauses to regroup. Additionally, the open-weight version released recently features strategically selected parameter sizes—120B for high-performance workstations and 20B for consumer PCs and even MacBooks. Its optimization goals and marketing clearly target the recent surge in China's open-source models.

Finally, the most radical change to the consumer experience is ChatGPT's completely revamped user interface. GPT5 represents more than just a model—it's a UNIFIED system that combines a fast model, a thinking model, and real-time routing between them. OpenAI has eliminated all other models, offering only GPT5. This simplification actually benefits the "silent majority" of users who were previously confused by multiple model options and likely never understood which one to choose—much like Word users who only know how to copy/paste using right-clicks. The backlash against these changes stems primarily from GPT5's underwhelming performance—many users believe it's inferior to GPT4o—rather than from the interface overhaul itself. The system also features four preset answer personalities, enhanced memory capabilities, better third-party integration (Gmail/calendar), improved healthcare answer accuracy, significantly reduced hallucinations, and stronger emphasis on secure output. While power users skilled with prompts might find these features trivial, they're genuinely transformative for novices. Additionally, for the first time, the flagship model was made immediately available to free users, andcharging U.S. government employees just $1 serves as both a political statement and a strategic move to capture institutional mindshare.

Analyzing the current competitive landscape helps us understand OpenAI's actions. As AI models have converged on the general LLM + RL formula, capturing users' mindshare has become the key competitive advantage. This is crucial not only for direct revenue but also for the second phase of modeling training —learning in real environments

On one hand, OpenAI leads in consumer products with 700 million weekly users (85% international). METR's report shows their user stickiness (DAU/MAU) exceeds 20%, while no close competitor surpasses 10% (excluding China's DeepSeek and MS Copilot, which also uses OpenAI's model). They've doubled annualized revenue to $12B with 90% penetration among Global 500 companies.

On the other hand, Anthropic's relative growth rate is significantly faster—growing more than 10x for two consecutive years and quadrupling in the first half of 2025 to reach $4B annually. An estimated 60-70% comes from API usage (despite having the industry's most expensive API pricing). Claude Code has reached nearly $400 million in annualized revenue within months of launch. Seeing developers gravitating toward Anthropic must concern OpenAI, as historical experience shows that platform-type products always win by attracting creators.

Meanwhile, Google remains a formidable competitor with modeling technology that potentially surpasses GPT. Their innovations span LLM, video, world models, and AI4S. Though slower in product development, Google's robust platforms—Search, Chrome, Android, YouTube—provide substantial resources to sustain a prolonged competitive battle. Additional threats include Musk-backed Grok from x.com, a regrouped Meta, and Chinese players fighting for ecological and enterprise developers with open-weight models.

OpenAI aims to establish itself as both the dominant global consumer AI portal and preferred assistant while maintaining its strong position with agent developers and enterprise markets. Their strategy is well-executed; however, ironically, what holds them back this time is their traditional strength—modeling capability...

This is some text inside of a div block.
by Lex
08.08.2025

In his article "How AI Conquered the US Economy—and What Happens Next" Derek Thompson uses a historical yardstick to place the AI investment boom within the panorama of U.S. economic development, giving us a more comprehensive understanding of these figures. In Q2 and Q3 of 2025, tech giants' capital expenditure on AI data centers has reached $400 billion, nearly $100 billion per quarter, accounting for approximately 1.36% of U.S. GDP during the same period. This figure not only exceeds the U.S. government's annual budget for education, employment, and social services, but can also be compared to the entire European defense spending.

Thompson emphasizes that AI-related capital expenditure approaches 2% of U.S. GDP. While the investment intensity has not yet reached the peak of the 19th-century railroad boom (in terms of GDP proportion, it's currently about 20% of the railroad investment peak), it has far surpassed the historical levels of the internet bubble and telecommunications investment.

More remarkably, in the first half of 2025, AI capital expenditure's marginal contribution to GDP growth rate surpassed consumer spending, the traditional engine, for the first time, becoming the biggest driver of economic expansion this year.

These funds primarily come from tech giants' free cash flow, capital market financing, and high investor expectations for the AI sector. Microsoft, Amazon, Google, Meta, and other companies have bet almost all their profits and financing on data center and AI computing power expansion, creating a concentrated flow of capital to very few sectors. Meanwhile, traditional manufacturing and consumer-oriented startups face cold financing conditions, with the economy's "diversified engines" being replaced by a "computing power monopoly."

Can these investments justify their returns? Not yet. AI hardware suppliers like NVIDIA have achieved doubled revenue and profit growth in 2025, driving related companies' market values to soar. The rises in the S&P 500 and NASDAQ are almost entirely driven by the "AI Seven Giants," while traditional consumer and manufacturing sectors show weakness. However, this income structure is highly dependent on the continued expansion of AI computing power demand. Once application-layer innovation or demand expectations cool down, related sectors may face dramatic adjustments in revenue and market value.

The employment landscape similarly shows structural bias. In May and June 2025, U.S. non-farm employment increases were only 19K and 14K respectively, both experiencing significant downward revisions, while July's new employment was only 73K, of which healthcare contributed 55K, accounting for 75%. Except for healthcare and social services, other industries showed almost zero or negative growth, with employment growth increasingly dependent on a few sectors. Meanwhile, the talent war in the AI field has reached an "NBA level." Meta has offered compensation packages worth hundreds of millions or even billions to top researchers, far exceeding the compensation levels of historical national technology projects such as the Manhattan Project and Apollo moon landing. These sky-high salaries not only make AI researchers worth more than most NBA stars but also highlight Silicon Valley's crazy bet on "superintelligence."

Finally, a bonus point: since the rise of LLMs, academic research has undergone tremendous changes. In 2024, the occurrence of the word "delves" exceeded the historical average by 2700%... It's estimated that 1/7 of abstracts have been processed by AI (including this article, of course)...

Regardless of the outcome of this investment boom, this is definitely no longer just a game that only affects the capital markets...

== Sources ==

This is some text inside of a div block.
Link Copied