The post The productivity bull case for almost everything appeared on BitcoinEthereumNews.com. This is a segment from The Breakdown newsletter. To read full editions, subscribe.  “Productivity isn’t everything, but in the long run it is almost everything.” — Paul Krugman “Total factor productivity” (TFP) is how economists measure the contribution of technological innovation to economic growth — the sustained ability of an economy to produce more output with the same amount of inputs. As such, it’s arguably economists’ most important measurement, because the continual process of producing more with less is how life gets better. “A country’s ability to improve its standard of living over time depends almost entirely on its ability to raise its output per worker,” Paul Krugman explains.  Technology is what makes that happen and TFP is how it’s measured. To get a more tangible sense of how important technology-generated productivity is, consider this: A recent paper from the National Bureau of Economic Research (NBER) estimates that an additional 0.5% of annual TFP growth would stabilize the US government’s finances at today’s level of debt-to-GDP. 0.5%! It doesn’t sound like a lot, but if sustained over the next 10 years, NBER estimates that would reduce the baseline forecast for US government debt by $2 trillion. Over 30 years, a sustained 0.5% boost to TFP would make the US government’s debt-to-GDP ratio 42 percentage points lower than NBER’s baseline forecast (and 80 percentage points lower than its pessimistic one). Given the seemingly hopeless state of government finances, maintaining today’s level of indebtedness is a dream scenario that seems too good to be true. But researchers at Anthropic think we can do even better. Anthropic conducted a study of 100,000 Claude.ai conversations to “estimate how long the tasks in these conversations would take with and without AI assistance, and study the productivity implications across the broader economy.”  Its conclusion? LLMs could… The post The productivity bull case for almost everything appeared on BitcoinEthereumNews.com. This is a segment from The Breakdown newsletter. To read full editions, subscribe.  “Productivity isn’t everything, but in the long run it is almost everything.” — Paul Krugman “Total factor productivity” (TFP) is how economists measure the contribution of technological innovation to economic growth — the sustained ability of an economy to produce more output with the same amount of inputs. As such, it’s arguably economists’ most important measurement, because the continual process of producing more with less is how life gets better. “A country’s ability to improve its standard of living over time depends almost entirely on its ability to raise its output per worker,” Paul Krugman explains.  Technology is what makes that happen and TFP is how it’s measured. To get a more tangible sense of how important technology-generated productivity is, consider this: A recent paper from the National Bureau of Economic Research (NBER) estimates that an additional 0.5% of annual TFP growth would stabilize the US government’s finances at today’s level of debt-to-GDP. 0.5%! It doesn’t sound like a lot, but if sustained over the next 10 years, NBER estimates that would reduce the baseline forecast for US government debt by $2 trillion. Over 30 years, a sustained 0.5% boost to TFP would make the US government’s debt-to-GDP ratio 42 percentage points lower than NBER’s baseline forecast (and 80 percentage points lower than its pessimistic one). Given the seemingly hopeless state of government finances, maintaining today’s level of indebtedness is a dream scenario that seems too good to be true. But researchers at Anthropic think we can do even better. Anthropic conducted a study of 100,000 Claude.ai conversations to “estimate how long the tasks in these conversations would take with and without AI assistance, and study the productivity implications across the broader economy.”  Its conclusion? LLMs could…

The productivity bull case for almost everything

2025/12/10 00:20

This is a segment from The Breakdown newsletter. To read full editions, subscribe.


“Total factor productivity” (TFP) is how economists measure the contribution of technological innovation to economic growth — the sustained ability of an economy to produce more output with the same amount of inputs.

As such, it’s arguably economists’ most important measurement, because the continual process of producing more with less is how life gets better.

“A country’s ability to improve its standard of living over time depends almost entirely on its ability to raise its output per worker,” Paul Krugman explains. 

Technology is what makes that happen and TFP is how it’s measured.

To get a more tangible sense of how important technology-generated productivity is, consider this: A recent paper from the National Bureau of Economic Research (NBER) estimates that an additional 0.5% of annual TFP growth would stabilize the US government’s finances at today’s level of debt-to-GDP.

0.5%!

It doesn’t sound like a lot, but if sustained over the next 10 years, NBER estimates that would reduce the baseline forecast for US government debt by $2 trillion.

Over 30 years, a sustained 0.5% boost to TFP would make the US government’s debt-to-GDP ratio 42 percentage points lower than NBER’s baseline forecast (and 80 percentage points lower than its pessimistic one).

Given the seemingly hopeless state of government finances, maintaining today’s level of indebtedness is a dream scenario that seems too good to be true.

But researchers at Anthropic think we can do even better.

Anthropic conducted a study of 100,000 Claude.ai conversations to “estimate how long the tasks in these conversations would take with and without AI assistance, and study the productivity implications across the broader economy.” 

Its conclusion? LLMs could raise total factor productivity by 1.1 percentage points.

1.1%!

If 0.5% would stabilize the US government’s finances for decades, what would 1.1% do? It would probably fix almost everything. 

There are reasons to be skeptical of this optimistic forecast, of course. 

The study finds, for example, that Claude saves teachers four hours of labor by creating curricula in just 11 minutes. But estimating how such time-savings might lead to higher economic output requires the kind of economic modelling that’s full of best-guess assumptions and false precision.

So, even if Anthropic is right about the time savings, it might be wrong about productivity: It might be that we use all the time AI saves us to do something economically unproductive, like watch more TikTok videos or read more newsletters.

In that case, AI would raise our welfare (more free time) but not our wealth (more economic output) — still great news for people, but no help to governments hoping for a silver bullet solution to their debt problem.

Conversely, there are reasons to think Anthropic’s model is being too pessimistic: “We don’t take into account the rate of adoption” it explains, “or the larger productivity effects that would come from much more capable AI systems.”

In other words, its study assumes we continue to use AI only as we do now and that we’re still using today’s language models, unimproved, for another 10 years.

Language models get noticeably better every few months and we’ve only just started learning how to use them — so Anthropic is right to say its estimate might represent an “approximate lower bound on the productivity effects of AI.”

If so — if 1.1% is the lower bound for AI-induced productivity — we might pay down government debt and have much more time for TikTok.

And that’s only taking into consideration AI’s impact on non-physical work — just wait until we get robots! 

To dismiss such optimism entirely is to think the trillions of dollars that corporations are planning to spend on AI capex and R&D will all be wasted.

Which it might be — technology revolutions don’t always arrive on schedule.

But the biggest reason for optimism is that Anthropic’s 1.1% estimate is based solely on AI “making existing tasks faster to complete” —  its model does not account for AI’s potential to completely change the way we complete those tasks.

“Historically,” Anthropic notes, “transformative productivity improvements — from electrification, computing, or the internet — came not from speeding up old tasks, but from fundamentally reorganizing production.”

There’s no way to model these new ways of doing things, but it seems likely its impact will be bigger than the one Anthropic has tried to measure.

Anthropic’s researchers are careful to caveat their hopeful findings by enumerating the limitations of their methodology and documenting the many assumptions they’re making. 

And even if all those assumptions work out and AI productivity solves the US government’s debt problem, lawmakers will probably spend their way right back into it.

But given the 100% probability everyone seems to put on looming fiscal disaster, even a small chance Anthropic’s estimates prove correct is a reason to update our priors: The US government’s finances are not as intractable as we think, and the US dollar is not as doomed as we think.

In the long run, productivity is almost everything — and AI might be on the verge of making us a lot more productive.


Get the news in your inbox. Explore Blockworks newsletters:

Source: https://blockworks.co/news/productivity-bull-case

Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Ethereum unveils roadmap focusing on scaling, interoperability, and security at Japan Dev Conference

Ethereum unveils roadmap focusing on scaling, interoperability, and security at Japan Dev Conference

The post Ethereum unveils roadmap focusing on scaling, interoperability, and security at Japan Dev Conference appeared on BitcoinEthereumNews.com. Key Takeaways Ethereum’s new roadmap was presented by Vitalik Buterin at the Japan Dev Conference. Short-term priorities include Layer 1 scaling and raising gas limits to enhance transaction throughput. Vitalik Buterin presented Ethereum’s development roadmap at the Japan Dev Conference today, outlining the blockchain platform’s priorities across multiple timeframes. The short-term goals focus on scaling solutions and increasing Layer 1 gas limits to improve transaction capacity. Mid-term objectives target enhanced cross-Layer 2 interoperability and faster network responsiveness to create a more seamless user experience across different scaling solutions. The long-term vision emphasizes building a secure, simple, quantum-resistant, and formally verified minimalist Ethereum network. This approach aims to future-proof the platform against emerging technological threats while maintaining its core functionality. The roadmap presentation comes as Ethereum continues to compete with other blockchain platforms for market share in the smart contract and decentralized application space. Source: https://cryptobriefing.com/ethereum-roadmap-scaling-interoperability-security-japan/
Share
BitcoinEthereumNews2025/09/18 00:25
Vitalik Buterin Suggests Ethereum Security Intact Amid Recent Glitch

Vitalik Buterin Suggests Ethereum Security Intact Amid Recent Glitch

The post Vitalik Buterin Suggests Ethereum Security Intact Amid Recent Glitch appeared on BitcoinEthereumNews.com. Ethereum remains secure despite a recent network glitch caused by a Prysm client bug that temporarily halted block finalization. Vitalik Buterin emphasized that this does not undermine the network’s core security, as blocks continue to be produced and executed, behaving like Bitcoin’s probabilistic model during such pauses. Vitalik Buterin assures that temporary loss of finality does not compromise Ethereum’s overall security model. The glitch primarily impacted secondary systems like bridges and Layer 2 solutions, not the base chain. Experts compare Ethereum’s response to Bitcoin’s, where probabilistic finality prevents chain rewrites while allowing continued operations. Ethereum secure despite recent glitch: Vitalik Buterin explains why the network’s resilience shines through temporary finality pauses. Discover key insights on blockchain reliability. Stay informed on crypto updates—read more now. What Did Vitalik Buterin Say About Ethereum’s Security After the Recent Glitch? Ethereum remains secure even amid the recent network disruption, according to Vitalik Buterin, Ethereum’s co-founder. He clarified that the Prysm client bug, which briefly interrupted block finalization, does not pose a threat to the protocol’s integrity. Instead, it highlights the network’s design for graceful degradation, where core functions persist without deterministic certainty. How Does Ethereum Behave During Finality Pauses? During the incident, Ethereum temporarily shifted to a probabilistic security model similar to Bitcoin’s, as noted by blockchain researchers. Fabrizio Romano Genovese, an Oxford PhD and Ethereum protocol specialist, explained that many blockchains, including Bitcoin, rely on growing difficulty in rewriting history rather than instant finality. In Ethereum’s case, blocks kept being created and executed, preventing any chain halt, though secondary services like cross-chain bridges experienced delays. This behavior underscores the network’s robustness, with no risk of approving incorrect transaction histories. Genovese added that such events reveal the need for better fallback mechanisms in dependent infrastructure, ensuring smoother operations in future occurrences. Statistics from the…
Share
BitcoinEthereumNews2025/12/11 16:40