Skip to content

Instantly share code, notes, and snippets.

@buwilliams
Last active April 10, 2025 19:02
Show Gist options
  • Save buwilliams/2a3b42d49ec0e9f466abc7cce8fd4ed1 to your computer and use it in GitHub Desktop.
Save buwilliams/2a3b42d49ec0e9f466abc7cce8fd4ed1 to your computer and use it in GitHub Desktop.

5-minute conversation, with enough evidence to spark discussion:

  • Natural Intelligence Took Ages to Evolve

    • Human intelligence developed over ~300,000 years as Homo sapiens emerged, shaped by survival pressures.
    • Fire, used ~1 million years ago, boosted brainpower by providing more energy (Wrangham, Catching Fire, 2009).
    • The brain eats up ~20% of our body’s energy, showing biology’s limits (Raichle & Gusnard, Nature Reviews Neuroscience, 2002).
  • AI Skips the Slow Grind

    • AI trains neural networks with gradient descent, mimicking evolution but running billions of iterations in years (Goodfellow et al., Deep Learning, 2016).
    • No biological constraints—AI uses raw compute power to scale fast and outpace human reasoning.
  • Compute Power Proves the Leap

    • GPT-3: ~3 × 10²³ FLOPS; GPT-4: ~2 × 10²⁵ FLOPS; Agent-1: ~3 × 10²⁷ FLOPS (Epoch AI, 2025 Compute Trends).
    • This explosive growth shows AI compressing millennia of natural trial-and-error into a few years.
  • Why It Fits Together

    • Natural intelligence: slow, energy-hungry, boosted by fire over eons.
    • AI: fast, unconstrained, engineered evolution on steroids—matching the story of intellect’s leap from biology to silicon.

Over the last two years (April 2023 to April 2025), the rapid advancement of AI capabilities driven by compute scaling has addressed and, in many cases, disproved several common objections that were prevalent earlier in the decade. These objections typically centered on diminishing returns, practical limitations, or overstated risks. Below, I’ll outline some of these objections and how evidence from the past two years has countered them, drawing on general trends and milestones in AI development up to the current date of April 10, 2025.

1. Objection: Diminishing Returns on Compute Scaling

  • Claim: Critics argued that increasing compute would yield progressively smaller gains in AI performance, as predicted by some interpretations of scaling laws (e.g., Chinchilla’s optimal scaling). They suggested that beyond a certain point, more compute wouldn’t translate into meaningful capability improvements.
  • Disproof: Over the last two years, models like OpenAI’s o1 (2024) and xAI’s Grok 3 (2024) have demonstrated that massive compute increases—sometimes orders of magnitude higher—continue to unlock new abilities. For example, o1’s enhanced reasoning in math and science (reportedly ~88% on MMLU) and Grok 3’s speculated near-human reasoning show that scaling compute still pushes boundaries. Benchmarks like MMLU have seen jumps from ~80% (GPT-4, 2023) to ~90%+ (2024 models), suggesting returns remain substantial when paired with architectural innovation.

2. Objection: Compute Scaling Hits Hardware Limits

  • Claim: Skeptics pointed to physical constraints—chip shortages, power consumption, and cooling challenges—as insurmountable barriers to scaling compute beyond 2022-2023 levels.
  • Disproof: Advances in hardware and infrastructure over the last two years have mitigated these concerns. Companies like NVIDIA and TSMC have rolled out next-gen chips (e.g., H200 GPUs in 2024) with higher efficiency, while AI labs have adopted distributed computing across massive clusters. Power usage, while still a challenge, has been addressed with renewable energy investments and optimized training algorithms (e.g., sparse models), allowing compute to scale from ~100,000 petaFLOP/s-days (GPT-4, 2023) to millions (Grok 3, 2024) without hitting a hard ceiling.

3. Objection: Data Bottlenecks Outweigh Compute Gains

  • Claim: Critics argued that even with more compute, the lack of high-quality training data would cap AI performance, rendering compute scaling ineffective without proportional data growth.
  • Disproof: The past two years have seen innovations in synthetic data generation and data-efficient learning that counter this objection. Models like GPT-4o (2024) and o1 leveraged self-generated data and improved pretraining strategies to achieve gains beyond what raw data availability would suggest. Additionally, multimodal models have expanded the data pool (e.g., images, audio), while techniques like retrieval-augmented generation (RAG) have reduced reliance on static datasets, proving compute can amplify capability even with finite data.

4. Objection: Scaling Compute Increases Risk Uncontrollably

  • Claim: Some warned that scaling compute would lead to unpredictable, uninterpretable, or dangerous AI systems, amplifying risks like bias, misalignment, or unintended behavior faster than safety measures could catch up.
  • Disproof: While risks remain a topic of debate, the last two years have shown that scaled models can be paired with improved interpretability and safety mechanisms. For instance, o1-preview (2024) reportedly included “reasoning traces” to make its decision-making more transparent, and Grok 3 (2024) integrates tools for analyzing content and user interactions, suggesting that safety research has kept pace. Performance gains haven’t led to widely reported catastrophic failures, and deployment has been cautious yet successful, undermining the idea of uncontrollable risk.

5. Objection: Compute Scaling is Economically Unsustainable

  • Claim: Detractors argued that the cost of scaling compute (e.g., billions for training runs) would become prohibitive, limiting its feasibility to only a few players and stalling progress.
  • Disproof: Over 2023-2025, the economics have shifted favorably. Training costs per FLOP have dropped due to hardware efficiencies and competition among chipmakers, while the value of AI outputs (e.g., in automation, research, and revenue generation) has soared. Companies like xAI and OpenAI have sustained massive compute investments—e.g., Grok 3’s speculated 5e26 FLOP (~$500M+ in training cost)—while delivering practical applications, proving economic viability. Smaller players have also entered the fray with efficient models, broadening access.

Key Takeaways from 2023-2025

  • Evidence of Progress: Compute scaling from ~10^5 petaFLOP/s-days (2023) to ~10^7 (2025) has coincided with AI moving from strong language tasks to advanced reasoning and multimodal mastery, disproving stagnation fears.
  • Adaptation: Hardware, software, and deployment strategies have evolved to handle scale, countering practical objections.
  • Balance: Safety and economic concerns have been addressed through parallel advancements, not just blind scaling.

These objections, once reasonable, have been largely disproved by the empirical success of AI systems over the last two years. Compute scaling isn’t limitless—new challenges may emerge—but the past 24 months show it’s far from hitting a wall. Let me know if you’d like a deeper dive into any specific objection!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment