I’m really coming to appreciate the value of scaffolding in product development.
What do I mean by scaffolding? The structure that allows you to build the product effectively.
Some examples:
A group of early customers willing to give fast, high quality feedback
Sample inputs and outputs that allow you to verify quality
Analytics or telemetry that give you an early indication of success or failure
You set up scaffolding to help you build. It doesn’t need to be pretty, but it needs to be fast, cheap, and effective. At the end of the project, you take it down. Or maybe you incorporate it into the structure of the product, improving it to make it fit for purpose.
Sometimes the scaffolding feels like a distraction. I’m going to build a whole separate structure just to help me build? Only if you want to build it well.
The best projects I’ve worked on outline the scaffolding early. These are the support structures we’ll need to do good work fast.
One of the grand challenges of artificial general intelligence is developing agents capable of conducting scientific research and discovering new knowledge. While frontier models have already been used as aids to human scientists, e.g. for brainstorming ideas, writing code, or prediction tasks, they still conduct only a small part of the scientific process. This paper presents the first comprehensive framework for fully automatic scientific discovery, enabling frontier large language models to perform research independently and communicate their findings. We introduce The AI Scientist, which generates novel research ideas, writes code, executes experiments, visualizes results, describes its findings by writing a full scientific paper, and then runs a simulated review process for evaluation. In principle, this process can be repeated to iteratively develop ideas in an open-ended fashion, acting like the human scientific community. We demonstrate its versatility by applying it to three distinct subfields of machine learning: diffusion modeling, transformer-based language modeling, and learning dynamics. Each idea is implemented and developed into a full paper at a cost of less than $15 per paper. To evaluate the generated papers, we design and validate an automated reviewer, which we show achieves near-human performance in evaluating paper scores. The AI Scientist can produce papers that exceed the acceptance threshold at a top machine learning conference as judged by our automated reviewer. This approach signifies the beginning of a new era in scientific discovery in machine learning: bringing the transformative benefits of AI agents to the entire research process of AI itself, and taking us closer to a world where endless affordable creativity and innovation can be unleashed on the world's most challenging problems.
From The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery, full paper available here
I'm interested to read this one more closely and see the degree to which it does (or doesn't) rely upon having experiments that the LLM can execute without human intervention. Either way, an interesting result, but my hypothesis is that "places where the LLM can verify a result" is going to be the limiting factor.
The first 10 chapters of this book are among my favorite reading experiences ever; terrifying, mysterious, creative. I wondered how the author was going to keep up that pace for the rest of the book.
Ultimately, he didn't. I'm not sure if it was possible for him to. Part of what made the beginning part of the book so enticing was wondering if miracles were actually happening or if they were just coincidences. At some point, the author had to make a choice and from that point forward everything got less interesting.