Embracing open-endedness in the pursuit of creative AI
Why do humans keep inventing new musical genres, designing cities that float on water, and imagining life beyond our universe? It’s because we’ve inherited nature’s restless, boundary-pushing creativity. We are an open-ended species—driven by curiosity, dissatisfied with “good enough”, and always reaching for what’s next. But here’s the catch: most of our AI systems today are not quite there yet. They’re built for specific tasks, optimized for efficiency rather than exploration. While this approach has taken us far, researchers at Google DeepMind recently argued that it won’t be enough to reach artificial superintelligence (ASI). So, the big question is:
Can we teach AI to be as endlessly creative as humans or even nature itself?
In this post, we’ll dive into the fascinating world of open-ended AI and explore how embracing open-endedness might finally break through the walls of today’s AI systems.
The endless creativity of nature
Before we talk about AI, let’s take a moment to appreciate the masterpiece that is biological evolution. Over 3.5 billion years, evolution has produced an almost unimaginable diversity of life forms. From extremophile bacteria that party in boiling volcanic vents to the 86 billion neurons in your brain that somehow let you understand these words, nature has been running the ultimate open-ended algorithm. What makes evolution so special? It’s the ultimate improvisational artist. There’s no predetermined goal, no final “perfect organism” to achieve, no optimization objective written in stone. Instead, evolution continuously explores the space of possible life forms through three deceptively simple ingredients:
Variation (mutation and recombination create new possibilities)
Selection (the environment acts as a brutal but fair critic)
Retention (successful innovations become building blocks for the next experiment)
This process has given us everything from the compound eyes of dragonflies (which inspired fiber optic technology) to the echolocation abilities of bats (which influenced radar development) to the language capabilities of humans (which… well, led to this blog post). Each innovation becomes a stepping stone for further exploration, creating an open-ended spiral of increasing complexity and capability. But perhaps most remarkably, evolution never gets “stuck” in local optima for long. When faced with mass extinctions (e.g., ice ages, asteroid impacts, dramatic atmospheric changes), life doesn’t just survive, it reinvents itself. Not through careful planning, but through relentless exploration of what Stuart Kauffman calls “the adjacent possible”The "adjacent possible" is a concept introduced by theoretical biologist Stuart Kauffman, referring to the space of potential next steps that are one step away from the current state. In evolution, this represents all the possible immediate mutations, combinations, or adaptations that could emerge from existing life forms.. Thus, evolution is basically the universe’s way of asking: “I wonder what would happen if…?”, and now imagine if we could capture even a fraction of this creative restlessness in our AI systems. This is where open-endedness comes in.
What is open-endedness?
Imagine you’re playing with LEGO blocks. A closed system is like following the instruction manual to build the exact model shown on the box — step by step, brick by brick. You follow steps 1–47, and voilà: you’ve built a sleek-looking car that looks just like everyone else’s.
Figure 1. A pile of LEGO bricks.
But an open-ended system is what happens when you toss the instructions aside and let curiosity take the wheel. You start building a car, but then wonder: “What if it could fly?” So you add propellers. Then you think, “What if it could go underwater too?” So you give it fins and a submersible dome. Before long, you’ve created something entirely new — a flying, diving, transforming vehicle that reconfigures itself depending on where it is. The beauty of open-endedness is this endless sense of possibility — no fixed goal, no final answer. Just a continuous chain of ideas sparking even better ones.
More formally, researchers from Google DeepMind define a system as open-ended if it continuously produces artifacts that are both novel (new and different) and learnable (useful and comprehensible). Think of it as the sweet spot between pure randomness (novel but useless) and pure repetition (learnable but boring).
The artifacts must be genuinely new or different from what came before. This could mean statistical novelty (different from the training distribution) or functional novelty (performing new behaviors or solving problems in unexpected ways).
The observer must be able to learn something useful from the artifacts. Random noise might be novel, but it’s not learnable. The artifacts should provide insights, capabilities, or stepping stones for further development.
But here’s where it gets interesting: open-endedness is fundamentally observer-dependent. What seems groundbreaking to a beginner might be trivial to an expert, and vice versa.
Figure 2. Open-endedness is in the eye of the beholder. The same aircraft design system produces different levels of novelty and learnability for different observers.
Consider a system that generates aircraft designs. For a mouse, these flying machines might be utterly novel but completely incomprehensible. For a human aerospace engineer, they hit that sweet spot of being both novel and learnable. But for a hypothetical superintelligent alien civilization, even our most advanced designs might seem as primitive as paper airplanes.
This observer-dependence isn’t a bug, it’s a feature! It means that as we build more capable AI systems, they can continue to surprise and teach us in ways we never expected.
The quest for open-endedness
The quest for open-ended AI isn’t new. In fact, it has been a dream of computer scientists for decades. Let me take you on a whirlwind tour through the key milestones that brought us to where we are today.
The cybernetic dreamers (1940s-50s)
It all started with Norbert Wiener in the 1940s, who introduced the concept of cybernetics – self-regulating systems that could adapt through feedback loops. Around the same time, John von Neumann was dreaming up self-replicating automata that could evolve beyond their initial programming. These pioneers laid the theoretical groundwork for machines that could grow and adapt, but the technology wasn’t quite there yet Ironically, Wiener himself was skeptical about AI. He particularly doubted the feasibility of mechanical translation, given the ambiguous nature of language and its emotional connotations..
The evolution revolution (1960s-70s)
John Holland changed everything in the 1970s with genetic algorithms. For the first time, we had a computational process that could evolve solutions through variation, selection, and retention – just like biological evolution, but in silicon. It was a proof of concept that machines could indeed create endless novelty.
The artificial life explosion (1980s-90s)
The 1980s and 90s saw the birth of Artificial Life (ALife), spearheaded by researchers like Christopher Langton. But the real showstopper was Thomas Ray’sTierra, a digital world where self-replicating programs evolved through natural selectionIn Tierra, digital organisms evolved remarkable strategies including parasitism (programs that used other programs' code to reproduce), immunity (defenses against parasites), and even hyper-parasitism (parasites that specifically targeted other parasites). These behaviors emerged purely from the evolutionary process without any explicit programming.. Notably, Tierra had no explicit optimization function, the fitness function was determined purely by survival and reproduction.
Figure 3. Tierra, a virtual ecosystem where digital organisms could reproduce, mutate, and compete for computational resources. These digital creatures evolved increasingly sophisticated survival strategies, including parasitism, immunity, and even altruism, all without any explicit programming for these behaviors!
The modern renaissance (2010s-2020s)
Fast forward to the 2010s, and we hit the modern era of open-ended AI. Generative Adversarial Networks (GANs) showed us how two neural networks could engage in an endless creative arms race, generating increasingly realistic images through adversarial competition. Then came AlphaGo and AlphaZero, which didn’t just master board games, they invented entirely new strategies that human experts had never conceived. Move 37 in AlphaGo’s match against Lee Sedol wasn’t just a game-winning move; it was a moment when a machine demonstrated genuine creativityMove 37 was so unexpected that Go commentators initially thought it was a mistake. It violated centuries of established Go wisdom by playing on the fifth line early in the game. However, it proved to be brilliant strategic thinking that human experts later recognized as fundamentally changing how the game could be played..
Figure 4. Move 37, the famous move in AlphaGo’s match against Lee Sedol that changed the course of AI history.
Limitations of current AI systems
Most AI systems today follow a predictable pattern: they are trained to optimize for a specific objective, and then they are deployed to solve that problem. This is a very efficient way to solve problems, but it has a few limitations. As Kenneth Stanley and Joel Lehman brilliantly articulate in “Why Greatness Cannot Be Planned”, this goal-oriented approach creates the so-called objective deceptionObjective deception refers to the counterintuitive phenomenon where directly pursuing a specific goal often leads away from achieving it. Stanley and Lehman argue that the most significant discoveries and innovations often come as byproducts of exploration rather than targeted optimization., such that the pursuit of specific goals often leads us away from more interesting discoveries. Once these systems reach their optimization target, they become what I call intellectual fossilsBy "intellectual fossils," I mean AI systems whose capabilities are permanently fixed at the moment their training ended—like prehistoric creatures preserved in amber, they retain their form but cannot adapt or evolve further.– perfectly preserved specimens that can’t evolve further. Their capabilities are frozen at the moment their training ended, unable to learn from new experiences or develop novel skills.
Open-ended systems offer an elegant escape from these limitations. Instead of optimizing for fixed objectives, they embrace what we might call “productive uncertainty”, or the willingness to explore without knowing exactly what they’re looking for. Here’s how open-endedness may break through each barrier:
Barrier
Open-ended systems
Fixed Problem Space
Discover new problems worth solving rather than just solving predefined problems. Develop meta-skills to learn, adapt, and transfer knowledge across domains.
Narrow Objectives
Use novelty search2 and quality-diversity3 approaches. Ask “What’s interesting?” rather than “What optimizes metric X?”
Static Capabilities
Treat achievements as stepping stones rather than destinations. Use mastered skills as foundations for exploring new challenges.
2 Novelty search refers to an algorithm that searches with no objective other than continually finding novel behaviors in the search space.
3 Quality-diversity refers to an algorithm that aims at generating large collections of diverse solutions that are all high-performing.
Open-endedness in action
Enough theory let’s see open-endedness in action! Below are some of the most exciting demonstrations of open-ended AI systems that I have come across.
Voyager
Our first example is Voyager, an AI agent that explores Minecraft with endless curiosity. Unlike traditional game AI that follows scripted behaviors, Voyager is powered by large language models (specifically GPT-4) and consists of three key innovations: an automatic curriculum that maximizes exploration, a growing skill library that stores reusable code for complex behaviors, and an iterative prompting system that incorporates feedback to improve performance. What makes Voyager special is its ability to learn continuously without human intervention - it sets its own goals, acquires diverse skills, and makes novel discoveries through pure experimentationCompared to previous approaches, Voyager achieves impressive results: it collects 3.3× more unique items, travels 2.3× further distances, and reaches key technological milestones up to 15.3× faster, all while developing interpretable and transferable skills.. Most importantly, the skills it develops are interpretable, compositional, and transferable to new scenarios. Voyager can take what it learns in one Minecraft world and apply those capabilities to solve novel challenges in entirely new environments, demonstrating true open-ended learning.
Figure 5. Voyager, an AI agent that explores Minecraft with endless curiosity.
The AI Scientist
Perhaps the most ambitious demonstration of open-endedness is The AI Scientist, developed by Sakana AI in collaboration with the University of Oxford and University of British Columbia. This groundbreaking system represents the first comprehensive framework for fully automatic scientific discovery, enabling foundation models to perform research independently across the entire research lifecycle. The AI Scientist automates four main processes:
Idea generation (brainstorming novel research directions and checking novelty via Semantic Scholar)
Experimental iteration (executing experiments and generating visualizations)
Paper write-up (producing complete manuscripts in LaTeX)
Automated peer review (evaluating papers with near-human accuracy)
Figure 6. The AI Scientist automates the full research cycle: brainstorming ideas, implementing algorithms, running experiments, writing papers, and conducting peer review - all while learning from previous iterations to improve future research.
Remarkably, the system generates complete research papers at approximately $15 per paperThis cost estimate includes computational expenses for idea generation, experiment execution, paper writing, and automated review, making AI-driven research potentially 100-1000× more cost-effective than traditional human research in terms of pure paper production., demonstrating both the cost-effectiveness and scalability of AI-driven research. In its initial demonstration, The AI Scientist successfully conducted research across diverse machine learning subfields, discovering novel contributions in diffusion models, transformers, and grokking. The generated papers achieved “Weak Accept” ratings from automated reviewers based on top-tier conference standards, with some proposing genuinely interesting new directions backed by solid experimental results. While critics argue about the quality of its output, they’re missing the bigger picture: this represents the first step toward AI systems that can discover genuinely new knowledge without human guidance, potentially democratizing research and significantly accelerating scientific progress.
Even more impressively, The AI Scientist-v2 has recently achieved a historic milestone by becoming the first entirely AI-generated system to produce a peer-review-accepted workshop paper at ICLR 2025. This next-generation system eliminates reliance on human-authored code templates, generalizes across diverse ML domains, and employs a novel progressive agentic tree-search methodology. Enhanced with Vision-Language Model feedback loops for iterative refinement, one of its three submitted manuscripts exceeded the average human acceptance threshold – marking the first instance of a fully autonomous AI paper successfully navigating peer review.
Figure 7. An example of a peer-review-accepted paper generated by The AI Scientist-v2. This paper reported a negative result that The AI Scientist encountered while trying to innovate on novel regularization methods for training neural networks that can improve their compositional generalization. This manuscript received an average reviewer score of 6.33 at the ICLR workshop, placing it above the average acceptance threshold.
AI-Generating Algorithms
Jeff Clune has proposed one of the most ambitious visions: AI-Generating Algorithms (AI-GAs), representing an alternate paradigm to the dominant “manual AI approach” where researchers attempt to discover and combine individual pieces of intelligence. Instead of hand-designing solutions, AI-GAs automatically learn how to produce general AI, following the clear trend in machine learning where learned solutions eventually replace hand-designed ones. The AI-GA approach rests on 3 pillars:
Meta-learning architectures that can automatically discover neural network designs
Meta-learning the learning algorithms themselves rather than using fixed optimization methods
Generating effective learning environments that provide the right training conditions for developing intelligence
We’re already seeing promising implementations of this vision. For example, Automated Design of Agentic Systems (ADAS), which won Outstanding Paper at the NeurIPS 2024 Open-World Agent Workshop, demonstrates that agents can literally invent novel agent designs by programming in code. Using their “Meta Agent Search” algorithm, the system iteratively programs new agents, tests their performance, archives discoveries, and uses this growing knowledge base to inform subsequent iterations. Remarkably, these AI-discovered agents consistently outperform hand-designed baselines across multiple domains and maintain superior performance even when transferred across different tasks and models.
Figure 8. In Meta Agent Search, the “meta” agent is instructed to iteratively program new agents, test their performance on tasks, add them to an archive of discovered agents, and use this archive to inform the meta agent in subsequent iterations.
Similarly, Sakana AI’s Darwin-Gödel Machine (DGM) represents another step toward self-improving AI systems, exploring how AI can autonomously evolve and enhance its own capabilities through recursive self-improvement.
Interested readers can refer to my previous post for a comprehensive overview of the Darwin-Gödel Machine.
The full vision involves AI that can design not just its software and algorithms, but potentially even its hardware and architecture, making AI-GAs a compelling candidate for achieving more general intelligence through automated discovery rather than manual engineering.
Safety in an open-ended world
Now for the elephant in the room: if we build AI systems that can continuously discover new capabilities and behaviors, how do we ensure they remain safe and beneficial? Open-endedness is incredibly powerful, but as they say, “with great power comes great responsibility”. By definition, we can’t predict what open-ended systems will discover (because that’s the whole point!). This creates unique safety questions:
What if an open-ended system discovers new ways to manipulate humans?
How do we prevent it from finding exploits in its environment or constraints?
Can we ensure it maintains beneficial goals even as it develops new capabilities?
Most of these safety questions remain open challenges and active areas of research. One of the most feasible directions is to maintain meaningful human involvement throughout the development and deployment of open-ended AI systems. This “human-in-the-loop” approach emphasizes:
Collaborative discovery: Humans and AI working together to explore new possibilities, with humans providing guidance and oversight
Iterative feedback: Human operators continuously evaluating and shaping the direction of AI exploration
Democratic participation: Involving diverse communities of humans in decisions about AI capabilities and their deployment
The idea is that rather than trying to solve all safety challenges through purely technical means, we can create robust frameworks where human judgment and values remain central to how open-ended systems develop.
Figure 9. An illustration of how open-ended artifacts emerge from human-AI collaborative process, where AI systems engaged in self-improvement and automated discovery, and humans providing guidance and oversight.
Takeaways
Long ago on Earth, life was simple — just single-celled organisms. Then came the Cambrian ExplosionThe Cambrian Explosion was a relatively brief period when most major animal phyla appeared in the fossil record. In just 10-25 million years, life went from simple forms to complex organisms with eyes, shells, and sophisticated body plans—demonstrating nature's capacity for rapid, open-ended innovation., a time when life suddenly became much more complex and diverse as creatures of all shapes and sizes began to appear. Now imagine something similar happening with AI. We’re getting close to a moment when AI systems won’t just be tools — they’ll be creative, evolving partners that help us in many different areas. Here’s what that could look like:
AI Scientists who keep exploring and making discoveries around the clock
AI Artists who create new kinds of music, art, and stories we’ve never seen before
AI Engineers who design buildings, spaceships, and even new types of computers
AI Teachers who help every student learn in the way that works best for them
AI Doctors who find new treatments and maybe even help people live longer
These AIs won’t be fixed or limited — they’ll keep learning and improving over time. And importantly, they won’t just work in their own little corners - they’ll be able to connect ideas from different fields. The RenaissanceThe Renaissance exemplified cross-pollination of ideas: Leonardo da Vinci's anatomical studies informed his engineering designs, Islamic mathematical concepts influenced European art, and artistic perspective techniques advanced architectural planning. This interdisciplinary approach led to unprecedented innovation across multiple domains. was a time when great progress happened not just because of smart individuals, but because of ideas mixing together. Artists learned from science, scientists looked to nature, and engineers borrowed ideas from art. Open-ended AI could spark a new kind of Renaissance. Imagine an AI working on music discovering something useful for architecture. Or an AI studying animals finding a solution that helps computer design. With open-ended AI, the lines between different fields might blur in exciting ways — helping us solve problems and create things we can barely imagine today.