The implications of AGI: What comes after the era of LLMs
What occurs when our machines start to grasp us as naturally as we perceive one another?
That’s not a query for the future – it’s one we’re dwelling by proper now.
Training in the present day’s large language models already prices upwards of $100 million. Just final yr, two Nobel Prizes have been awarded for AI breakthroughs. That’s extraordinary. It indicators one thing profound: we’ve crossed a threshold the place synthetic intelligence isn’t simply fixing issues, it’s remodeling how we predict, create, and work together.
In my profession, from main analysis at Google DeepMind to my present work as Chief AI Officer at Genesis Therapeutics, I’ve seen AI evolve from brittle methods that adopted instructions to versatile companions succesful of reasoning, studying, and even displaying hints of character.
So on this article, I’ll replicate on the place that journey has taken us, and the place it’s main subsequent. We’ll discover how giant language fashions (LLMs) are altering pure interplay, unifying management throughout methods, and even studying autonomously.
Most importantly, we’ll think about what these breakthroughs imply for the path towards Artificial General Intelligence (AGI) – and for the security, duty, and humanity of the area we’re constructing collectively.
Let’s get began.
Teaching robots to grasp us
When I first began in robotics, giving directions to a robotic was about as intuitive as writing meeting code. You needed to specify coordinates, velocities, joint angles – each micro-step.
Now, think about as an alternative saying one thing like:
“Trot ahead slowly.” “Back off – don’t damage the squirrel.” “Act such as you’re limping.”
And the robotic merely understands.
That’s the leap we’ve made due to giant language fashions. In one of our early initiatives, we used GPT-4 to regulate a quadruped robotic. Underneath, a conventional controller dealt with the bodily contact patterns – blue meant contact, yellow meant elevate – whereas the LLM acted as an clever interface translating pure language into motor instructions.
What amazed us wasn’t simply that it labored – it was that it generalized. You might inform the robotic, “Good information, we’re happening a picnic!” and it will actually leap round.
That’s what I imply by pure interplay. For the first time, non-experts can entry complicated AI methods with out programming or robotics experience. It’s a basic shift – one which opens up AI to tens of millions extra individuals and use instances.

Code as a typical language
Across robotics, internet brokers, and digital assistants, one massive barrier has at all times been fragmentation. Every system speaks a unique “motion language.”
A self-driving car thinks in phrases of steering angle and acceleration.A quadruped robotic thinks in phrases of joint torques.An online agent manipulates HTML components.
There’s no common interface.
But code would possibly simply be that common motion area.
Let me offer you an instance. We constructed an internet navigation agent succesful of executing multi-step searches totally from pure directions like:
“Find one-bedroom flats in Ortonville for company housing, ranging from google.com.”
The agent reads the uncooked HTML (we’re speaking megabytes of unstructured knowledge), plans the subsequent steps, writes the essential code, executes it, and repeats – closing the loop autonomously.
With simply 200 self-collected demonstrations, this modular system discovered to generalize throughout totally new web sites. We achieved success charges between 65% and 80% on real-world domains like actual property, Reddit queries, and Google Maps.
However, this functionality additionally raised early considerations about AI security. I bear in mind vividly in late 2022, proper as ChatGPT launched, we have been discussing whether or not brokers ought to be allowed to jot down and execute code on their very own. That’s a robust and doubtlessly harmful means.
So whereas this experiment demonstrated how LLMs can unify motion throughout domains, it additionally reminded us that functionality should include management.



