Home
Article
From GPT-4 to AGI

From GPT-4 to AGI

From GPT-4 to AGI

GET THE #1 EMAIL FOR EXECUTIVES

Subscribe to get the weekly email newsletter loved by 1000+ executives. It's FREE!

Chris Dunlop's photo
Author
Chris Dunlop

Link

Article Description

This article makes the case that a leap to AGI by 2027 is possible, predicting and explaining that we should see another leap equivalent to the leap from GPT2 to GPT4 over the next 3 years.

AI Trends

Counting the Efficiency Increase

The author has a theory, that the progress in AI will come from a combination of:

  • How much more powerful the computers are (compute).
  • How much smarter the algorithms are getting (algorithmic efficiencies), which makes the computers feel even more powerful (we call this "effective compute").
  • How fixing small issues can make AI even better (unhobbling gains).

They sum all of this up into a framework called OOM (Order of Magnitude) where gains happen in multiples of 10x.

So they expect an increase in 100,000 in effective compute scaleup over the next 4 years.

AI Speed of Progress

Look at the chart below for the improvment tha thas been made in a single year.

GPT Improvement

Over and over again, year after year, skeptics have claimed “deep learning won’t be able to do X” and have been quickly proven wrong.If there’s one lesson we’ve learned from the past decade of AI, it’s that you should never bet against deep learning.
Leopold Aschenbrenner
Leopold Aschenbrenner

Follow the OOM to see a glimpse of the future

With each OOM of effective compute, models predictably, reliably get better. If we can count the OOMs, we can (roughly, qualitatively) extrapolate capability improvements. That’s how a few prescient individuals saw GPT-4 coming.
Leopold Aschenbrenner
Leopold Aschenbrenner

We can break down the progress from GPT-2 to GPT-4 into three big improvements:

  1. Compute: We're using much bigger and more powerful computers to train these AI models.
  2. Algorithmic efficiencies: The algorithms are getting better and smarter, making the computers seem even more powerful. We can measure this as growing "effective compute."
  3. Unhobbling gains: AI models have lots of potential, but they are held back by simple issues. By making small fixes like using human feedback, step-by-step thinking, and adding helpful tools, we can unlock a lot of hidden abilities and make the AI much more useful.

Efficiency Gain

The data wall and how to smash through it

There is a potentially important source of variance for all of this: we’re running out of internet data. That could mean that, very soon, the naive approach to pretraining larger language models on more scraped data could start hitting serious bottlenecks.
Leopold Aschenbrenner
Leopold Aschenbrenner

But maybe there is a way to be more efficient with the data we have.

What a modern AI model (LLM) does during training is like skimming through a textbook really fast without much thinking.

When we read a math textbook, we do it slowly, think about it, discuss it with friends, and try practice problems until we understand. We wouldn't learn much if we just skimmed through it like the AI models do.

But, there are ways to help AI models learn better by making them do what we do: think about the material, discuss it, and keep trying problems until they get it. This is what synthetic data, self-play, and reinforcement learning approaches aim to achieve.

A common pattern in deep learning is that it takes a lot of effort (and many failed projects) to get the details right, but eventually some version of the obvious and simple thing just works. Given how deep learning has managed to crash through every supposed wall over the last decade, my base case is that it will be similar here.
Leopold Aschenbrenner
Leopold Aschenbrenner

A counter argument to stalled progress

Moreover, it actually seems possible that cracking one of these algorithmic bets like synthetic data could dramatically improve models. Here’s an intuition pump. Current frontier models like Llama 3 are trained on the internet—and the internet is mostly crap, like e-commerce or SEO or whatever. Many LLMs spend the vast majority of their training compute on this crap, rather than on really high-quality data (e.g. reasoning chains of people working through difficult science problems). Imagine if you could spend GPT-4-level compute on entirely extremely high-quality data—it could be a much, much more capable model.
Leopold Aschenbrenner
Leopold Aschenbrenner

AlphaGo, the first AI to beat world champions at Go, is a great example.

  1. Step 1: AlphaGo learned by watching expert human Go games. This gave it a basic understanding.
  2. Step 2: AlphaGo played millions of games against itself. This made it super good at Go, leading to moves like the famous move 37 against Lee Sedol, which was brilliant and unexpected. This self-play method allowed AlphaGo to explore new strategies and refine its skills beyond human capabilities. It shows how AI can advance rapidly by learning from its own experiences, potentially leading to breakthroughs in other fields as well.
As an aside, this also means that we should expect more variance between the different labs in coming years compared to today. Up until recently, the state of the art techniques were published, so everyone was basically doing the same thing. (And new upstarts or open source projects could easily compete with the frontier, since the recipe was published.) Now, key algorithmic ideas are becoming increasingly proprietary. I’d expect labs’ approaches to diverge much more, and some to make faster progress than others—even a lab that seems on the frontier now could get stuck on the data wall while others make a breakthrough that lets them race ahead. And open source will have a much harder time competing. It will certainly make things interesting.
Leopold Aschenbrenner
Leopold Aschenbrenner

Unhobbling

Finally, let's talk about "unhobbling" - making AI models work better by removing simple limitations.

Imagine if you had to solve a hard math problem instantly, without working it out step-by-step. It would be really hard, right? That’s how we used to make AI solve math problems. But we figured out a better way: letting AI work through problems step-by-step, just like we do. This small change, called "Chain-of-Thought" prompting, made AI much better at solving difficult problems.

We've made big improvements in "unhobbling" AI models over the past few years:

  1. Reinforcement Learning from Human Feedback (RLHF): This technique helps AI learn from human feedback, making it more useful and practical. It's not just about censoring bad words; it helps the AI understand and answer questions better. For example, a small AI model trained with RLHF can perform as well as a much larger model without it.

  2. Chain of Thought (CoT): This technique lets AI think through problems step-by-step. It’s like giving the AI a scratchpad to work out math and reasoning problems, making it much more effective.

  3. Scaffolding: This involves using multiple AI models together. One model plans how to solve a problem, another proposes solutions, and another critiques them. This teamwork approach can make even smaller models perform better than larger ones working alone.

  4. Tools: Imagine if humans couldn't use calculators or computers. Similarly, giving AI models tools like web browsers or code execution capabilities helps them perform better. ChatGPT can now do things like browse the web and run code.

  5. Context Length: Early models could only remember a small amount of information at once. Now, models can remember much more (from 2k tokens to over 1 million tokens). This helps them understand and work on bigger tasks, like understanding a large codebase or writing a long document.

  6. Posttraining Improvements: Even after training, AI models can continue to improve. For example, the current GPT-4 has gotten much better at reasoning and other tasks compared to when it was first released.

By removing these limitations, we've made AI models much more powerful and useful.

Drivers of AI Progress

The possibilities here are enormous, and we’re rapidly picking low-hanging fruit here. This is critical: it’s completely wrong to just imagine “GPT-6 ChatGPT.” With continued unhobbling progress, the improvements will be step-changes compared to GPT-6 + RLHF. By 2027, rather than a chatbot, you’re going to have something that looks more like an agent, like a coworker.
Leopold Aschenbrenner
Leopold Aschenbrenner

oom trends

Thanks for reading! If you found this helpful, please share this article with 1 friend!

More Articles

The Hidden Psychology of Decision-Making: What Executives Can Learn from Hostage Negotiators

ARTICLE

The Hidden Psychology of Decision-Making: What Executives Can Learn from Hostage Negotiators

Explore how understanding emotional under currents can enhance decision-making in business...

No Rules Rules: Netflix and the Culture of Reinvention

BOOK REVIEW

No Rules Rules: Netflix and the Culture of Reinvention

An inside look into the culture of Netflix...

Olympic Glory in the Digital Age: How Paris 2024 Reshaped Athletes' Social Media Landscapes

ARTICLE

Olympic Glory in the Digital Age: How Paris 2024 Reshaped Athletes' Social Media Landscapes

We look at Instagram follower counts and see if there is a correlation between winning a medal or not...

Do you lead people? Your mood is like electricity - it spreads

ARTICLE

Do you lead people? Your mood is like electricity - it spreads

Discover how a leader's emotional state can spread through an organisation like wildfire, influencing performance at every level...

The 5 Word Meeting Technique

ARTICLE

The 5 Word Meeting Technique

Google, Apple and Amazon were told to run their companies this way...

What Every CEO Can Learn from GitHub's 100-Day Leadership Challenge

ARTICLE

What Every CEO Can Learn from GitHub's 100-Day Leadership Challenge

How Nat Friedman's Bold 100-Day Strategy Transformed GitHub and Redefined Leadership....

AI-Powered Networking: Building 50+ Connections in a New City

ARTICLE

AI-Powered Networking: Building 50+ Connections in a New City

Discover how I leveraged AI to transform networking in London, creating a scalable system for building meaningful professional relationships.

Struggling to Empower Your Team? Read This Book

BOOK REVIEW

Struggling to Empower Your Team? Read This Book

Learn how the best companies build products...

51 Books Every Executive Should Read in 2024

BOOKS

51 Books Every Executive Should Read in 2024

Hand picked, each of these has shaped us in some way...

25 Strategic Moves That Established Companies Need to Steal from Startups

ARTICLE

25 Strategic Moves That Established Companies Need to Steal from Startups

Sick of slow progress in your organisation? This will help you focus on what matters...

The Trillion Dollar Coach: Steve Job's Coach

BOOK REVIEW

The Trillion Dollar Coach: Steve Job's Coach

This is a simple book that is a must read for any leader...

Using a can of beans to figure out consulting pricing

ARTICLE

Using a can of beans to figure out consulting pricing

Get better at pricing your consulting jobs...

The Economics of Airbnb Icons

ARTICLE

The Economics of Airbnb Icons

Why exactly did they build the UP house?

How Youtube can help you to find your North Star Metric

ARTICLE

How Youtube can help you to find your North Star Metric

Learn why the biggest companies choose one thing to focus on...

Jeff Bezos is famous for reading slowly - here's why you need to do it too

ARTICLE

Jeff Bezos is famous for reading slowly - here's why you need to do it too

Honestly this is not an easy thing to do. I tried and it felt like the mental equivalent of deadlifting

Running a large organisation? You need to think this way

ARTICLE

Running a large organisation? You need to think this way

Day one thinking and why you need to make sure that your company thinks like this...

The Time a PhD Mathematician Won the Olympics

ARTICLE

The Time a PhD Mathematician Won the Olympics

The story of Anna Kiesenhofer's incredible victory in Tokyo

AI paying Humans?

ARTICLE

AI paying Humans?

A new company Payman is betting that the future of work involves AIs paying us to do their boring tasks ...

Company

Site Information

Fun Stuff

© 2024 Cub Digital. All Rights Reserved.