Heliox: Where Evidence Meets Empathy π¨π¦β¬
We make rigorous science accessible, accurate, and unforgettable.
Produced by Michelle Bruecker and Scott Bleackley, it features reviews of emerging research and ideas from leading thinkers, curated under our creative direction with AI assistance for voice, imagery, and composition. Systemic voices and illustrative images of people are representative tools, not depictions of specific individuals.
We dive deep into peer-reviewed research, pre-prints, and major scientific worksβthen bring them to life through the stories of the researchers themselves. Complex ideas become clear. Obscure discoveries become conversation starters. And you walk away understanding not just what scientists discovered, but why it matters and how they got there.
Independent, moderated, timely, deep, gentle, clinical, global, and community conversations about things that matter. Breathe Easy, we go deep and lightly surface the big ideas.
Heliox: Where Evidence Meets Empathy π¨π¦β¬
π The Architecture of Innovation: On jokes, genius, and the AI economy we haven't built yet
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
There is a moment β you've felt it β when a joke lands just right. Not a polite chuckle, not a social reflex, but the real thing: a full-body release, something almost involuntary, like a hiccup of the soul. For a split second, your brain held two incompatible truths simultaneously and then, unable to contain them both, simply laughed. What if that moment β that tiny, human, ridiculous moment β turns out to be one of the most important cognitive events in the known universe?
See Substack for references
This is Heliox: Where Evidence Meets Empathy
Independent, moderated, timely, deep, gentle, clinical, global, and community conversations about things that matter. Breathe Easy, we go deep and lightly surface the big ideas.
Disclosure: This podcast uses AI-generated synthetic voices for a material portion of the audio content, in line with Apple Podcasts guidelines.
We make rigorous science accessible, accurate, and unforgettable.
Produced by Michelle Bruecker and Scott Bleackley, it features reviews of emerging research and ideas from leading thinkers, curated under our creative direction with AI assistance for voice, imagery, and composition. Systemic voices and illustrative images of people are representative tools, not depictions of specific individuals.
We dive deep into peer-reviewed research, pre-prints, and major scientific worksβthen bring them to life through the stories of the researchers themselves. Complex ideas become clear. Obscure discoveries become conversation starters. And you walk away understanding not just what scientists discovered, but why it matters and how they got there.
Independent, moderated, timely, deep, gentle, clinical, global, and community conversations about things that matter. Breathe Easy, we go deep and lightly surface the big ideas.
Spoken word, short and sweet, with rhythm and a catchy beat.
http://tinyurl.com/stonefolksongs
Imagine the exact moment you get a really, really good joke. Oh, yeah. You know that feeling, right? It's like it's almost like a physical snap of comprehension in your brain. Right. Like a release. Exactly. For a split second, you're completely confused. And then suddenly two entirely different ideas crash together. The tension releases and you just laugh. Yeah. Well, it's actually a physiological response to a cognitive shift. The punchline lands, the expected reality just sort of evaporates, and a new, completely unexpected reality instantly clicks into place. Now hold on to that very specific feeling, because next, I want you to imagine a completely different scenario. Imagine a scientist. Someone who has been, I don't know, staring at a blackboard full of equations for months, maybe years, just banging their head against a wall. Yeah, we've all been there. Right. And then out of nowhere, the clouds part, the math suddenly works. They realize a brand new law of physics. It's that ultimate world-shaking eureka moment. Oh, wow. Yeah. The feeling of the universe just revealing a hidden mechanism to you. A profound shift in understanding. Exactly. So here is the massive, slightly mind-bending question we are tackling in this deep dive today. What if those two experiences, laughing at a pun at a dinner party and uncovering the fundamental secrets of the universe, what if they aren't just, you know, metaphorically similar? Right. What if they are driven by the exact same cognitive machinery in our brains? That is the premise we are unpacking today. And we connect this to the bigger picture for you listening at home. This isn't just an abstract exploration of, you know, human psychology or neuroscience. We are looking at this through the lens of a technological roadmap. Because we have a stack of incredibly cutting edge research in front of us today and the mission is absolutely wild, we are looking at how humans identify important pivotal innovations. Starting with our baseline capacity for delight, curiosity, and humor. Right. And we are figuring out how to capture that exact essence in a machine learning system. We are literally talking about quantifying curiosity. It sounds impossible. It does. Quantifying that aha moment into hard mathematical metrics. Because if we can teach an AI to recognize true, original, thoughtful innovation, we can use that to build something unprecedented. I mean, think about the Internet as it exists right now for you, the listener. It's an economy based entirely on attention. Yep. Clicks. Right. Clicks, outrage, virality. If someone invents something brilliant and then, you know, 10,000 people copy it and make slightly better TikToks about it. the copiers get the money exactly but the research we are looking at today proposes the foundation for a brand new bottom-up economy an economy based on recognition on micropayments for every single use of an idea and on something called durability yes durability let's talk about that. Durability is really the key word there. It means ensuring that the true Nobel Prize level leaps of insight are rewarded disproportionately and forever. Rather than having their value completely diluted by the thousands of minor follow on papers or posts that just tweak the original concept. It's about tracing the DNA of an idea. From the exact moment you type a prompt into an AI, tracing the answer all the way back through the neural network to the human who originally had the insight and paying them for it. Exactly. But I mean, to build a machine that can recognize a brilliant idea, we first have to understand what a brilliant idea actually is. We have to start with the biology of the punchline. Right. And for that, the researchers point us back to a novelist and cultural critic named Arthur Koestler. OK. Arthur Koestler. Yeah. His 1964 book, The Act of Creation, is really the philosophical bedrock for this whole approach. He spent years analyzing human innovations. And he developed a theory that he called bisociation. Bisociation. Yeah, and we have to be careful here because bisociation is fundamentally different from mere association. Wait, let me stop you there because association is how we usually think about thinking, right? Exactly. If I say dog, you think cat. If I say rain, you think umbrella. It's routine. It's moving along a single established track of logic. That's a perfect way to describe it. Koestler called those single tracks matrices of thought. So association is just sliding along one matrix. You are following the rules of a familiar game. Okay. But Koestler argued that true creativity is entirely different. Creativity is the perceiving of a situation or an idea in two self-consistent but habitually incompatible frames of reference at the exact same time. So it's not moving along one track. No. It's the violent collision of two completely different tracks. He beautifully described a pun, for example, as two strings of thought tied together by an acoustic knot. Okay. An acoustic knot. I love that. So you are vibrating on two different wavelengths simultaneously. Exactly. And this brings us to what Kostler called the rounded triptych of creativity. Okay. The triptych. He believed that humor, scientific discovery, and art all share this exact same underlying psychological process of bisociation. Wait, really? All three of them? All three. The logical pattern finding hidden similarities in disparate worlds is identical across all three. The only thing that changes between them is what he termed the emotional climate. I want to break down that triptych because the way he categorizes human emotion here is fascinating. Let's start with the first panel, which is humor. Right. So in humor, the cognitive response is the ha ha. The emotional climate here is aggressive or self-assertive. It's a collision. Two incompatible codes clash. Yep. And because they can't be reconciled, the narrative tension simply explodes. The energy is discharged. And that explosion is the laugh. Exactly. Okay. So what about the second panel, Scientific Discover? The cognitive response there is the aha. The aha moment. Right. Here, the emotional climate is neutral or intellectual. Instead of a collision that explodes, it's a fusion. Okay. The two different matrices of thought are permanently integrated into a new, higher-level synthesis. The tension isn't discharged. It is utilized to build a new universal law. Ah, I see. And then the third panel, art. That is the, like a sigh. Yeah, exactly. The emotional climate is sympathetic and identificatory. It's not an explosion and it's not a complete fusion. It is a juxtaposition. So the artist places different planes of experience side by side to create an emotive resonance. You feel the connection between, I don't know, the paint on the canvas and a feeling of melancholy from your childhood. Beautifully put. Yes. This theory makes so much sense when you look at the historical examples in the research. They bring up St. Jerome in the fourth century. Oh, this is a great example. Right. So he's tasked with translating the old Latin Bible into the Vulgate. He gets to the story of the Garden of Eden, the fall of man. Now, originally there was no apple. Right. It was just fruit. It was just a generic fruit. But Jerome realizes that the Latin word malice means evil, but the word malum means apple. And he bissociates the two concepts. He ties the theology of evil and the physical object of an apple together with that acoustic knot, and boom, he alters the entire visual iconography of human history forever. Exactly. Or, you know, consider the classic example of Isaac Newton sitting in his garden. He sees an apple fall. A purely associative thinker just thinks, the apple is ripe, it fell, I should eat it. Right. But Newton bisocias. He understands the event simultaneously as the completely mundane fate of a piece of ripe fruit and as a startling demonstration of the cosmic law of gravity pulling the moon toward the Earth. He fuses the earthly and the celestial. That is by association in action. But I mean, what is actually happening in the brain during this? Why does this collision or this fusion feel so uniquely powerful to us biologically? That is explained by something called the incongruity resolution theory. Or IR theory. Yes, IR theory. This posits the humor and, by extension, insight results from the sudden surprising resolution of cognitive dissonance. Okay, unpack that. In the language of complexity science, your brain undergoes a literal phase transition. When you are listening to a joke, your brain builds an attractor-like script. Which is a fancy way of saying your brain is constantly aggressively predicting what will happen next based on patterns. experience. Exactly. We are prediction engines constantly. But then the punchline arrives and it actively destroys that expected script. It just wipes it out. Yep. It replaces it with a completely different, less probable script. And when your brain is forced to to make that sudden violent shift between realities, it releases what researchers call cognitive free energy. Cognitive free energy. And that free energy manifests physically. Your diaphragm spasms. You vocalize. You experience amusement. Like a spark coming off a short circuit. A very joyful short circuit. And we can actually see this happening in neuroimaging. When you process a joke, very specific brain regions light up.- Which ones?- The temporal parietal junction, or TPJ, and the precuneus become highly active.- Wait, what are those areas actually doing, though?- Well, the TPJ is deeply involved in perspective taking and integrating multiple streams of information, which makes sense, you're holding two realities in your head.- Right, right.- Meanwhile, your middle temporal gyrus, which is involved in semantic processing, detects the incongruity itself. It's the alarm bell that says, wait, this doesn't fit the pattern. Okay, let me make sure I'm visualizing this correctly. It's like a joke is a mental train track. We're riding along one line of logic. Everything is smooth. Everything is predictable. And right at the punchline, the switch is flipped violently, and we suddenly find ourselves on a completely different track, heading in a different direction. And that sudden jerk, that release of kinetic energy from switching tracks, is the laughter. That is a highly accurate analogy. And Koestler made a very profound point about this exact mechanism. He argued that true genius isn't about perfection. It is about originality. Because humans are fundamentally creatures of habit. Yes. And habits reduce us to, in his words, conditioned automatons. We just run on autopilot because it's metabolically cheaper for the brain. Yes. We follow the associative tracks because they are easy. Yeah. But the creative act that bisociative jump between tracks is an act of liberation. It is the defeat of habit by originality. It allows us to briefly escape the autopilot and attain a higher level of mental evolution. Okay, but here is where we hit the massive technological challenge. If human innovation is essentially this mental track switching, this biological release of cognitive free energy, how on earth do we teach a machine to recognize it? Right. A large language model doesn't have a temporoparietal junction. It doesn't feel a release of free energy. It doesn't laugh. So how do you code an aha moment? That is the crux of the problem. You have to translate the biological aha into pure mathematics. And to do that, the researchers turned to information theory, and specifically, a metric called surprisal. Surprisal. It sounds like a made-up corporate buzzword. It does, but it's a foundational mathematical concept in AI. Okay, so what is it? Surprisal is defined mathematically as the negative log probability of a word or a state, given its context. Okay, let's ground that, because negative log probability is heavy. Let's say I'm writing a sentence. The sky is, the word blue has a really high probability of coming next, everyone expects it. So the surprisal is basically zero. Right. But if I say the sky is screaming, the math spikes. It's quantifying the unexpected. Exactly. The more unexpected the token, the higher the surprisal. And the research highlights a fascinating empirical study about this, looking at mathematicians working at blackboards. Oh, this study blew my mind. It's brilliant. Researchers analyzed the moment-to-moment physical dynamics of these experts trying to solve incredibly complex math problems. They recorded their writing speed, their gesturing, their gaze shifts, the pauses. Everything. Yeah, everything. And they found that immediately, prior to an "aha" moment of insight, The literal unpredictability of their physical behavior completely spiked. Wait, what were they doing? Just pacing around erratically? Pacing, erasing furiously, staring blankly, rapid hand movements. The system became highly chaotic right before the breakthrough. It's like water boiling right before it turns into steam. The system has to destabilize before it can crystallize into a new state. That is a perfect thermodynamic analogy. Right. And we see this exact same phenomenon inside artificial intelligence when we look at how models perform chain of thought reasoning. Chain of thought is when you ask an AI to think step by step before giving you the final answer, right? It shows its work. Yes. And when researchers analyze the surprisal of the tokens generated during that step-by-step reasoning, they find that the logical progression isn't just a smooth, uniform line. The information density is highly concentrated in very specific moments. Where does the math spike? At the very beginning of a new reasoning step. The research calls this first-token surprisal. First-token surprisal. Right. High-entropy cognitive pivots happen right at the start of a new thought. often marked by seemingly simple words like but. Which signals a self-correction or a realization that a previous path was wrong. Exactly. Or the word so, which signals a sudden deduction. These aren't just grammatical filler words. They are the literal switches on your mental train track. By measuring the surprisal of these initial tokens, we can actually mathematically distinguish between a predictable, boring elaboration, and a critical novel logical transition. Okay, but here's my problem with that. Let's hear it. a predictable, If we are just measuring novelty by how surprising or unpredictable a word is, couldn't a cat walking across a keyboard generate maximum surprisal? I mean, typing Q-W-E-R-T-Y-8-9-Z has incredibly high surprisal. It's totally unpredictable. That's very true. But we don't want our new economy paying a cat a Nobel Prize for typing garbage. No, we don't. You've hit on the exact trap of relying solely on originality metrics. High surprisal without structure is just noise. This raises the critical question for the researchers. How do we mathematically define meaningful novelty? Because there's a difference between a crazy idea that works and a crazy idea that is just crazy. Yes. Right. To solve this, the researchers tackled it by proposing a new metric. They argue that true quantifiable novelty is their harmonic mean between two distinct factors. originality and quality. Okay, break those two down for us. Why a harmonic mean? Let's start with the factors. Originality is mirrored empirically. You take the AI's generated output and you calculate the fraction of n-grams, which are just chunks of text, sequences of words that never appeared in the model's massive training data. So if the model is outputting text fragments that it literally never saw during its training, originality is high. Yes. And the second factor, quality. How do you measure if a poem or a math proof is actually, you know, good? Quality is highly task specific. Does the output actually solve the user's problem? Does it follow the constraints? Is it logically coherent? To scale this measurement across thousands of outputs, researchers use an LLM as a judge framework. Okay. They use a very powerful model like GPT-4 and give it strict rubrics to evaluate the usefulness of the output. And then you combine them using a harmonic mean. But why harmonic? Why not just average the two scores together? Because an arithmetic average can hide a fatal flaw. If a text has an originality score of 100 pure cat on the keyboard gibberish and a quality score of 0, the average is 50. Which looks somewhat okay on paper. Right. But a harmonic mean heavily penalizes the final score if either of the inputs is low. A harmonic mean of 100 and 0 is 0. Ah. So the harmonic mean is the safety net. It demands that the output must be both surprising, AE, functional. It says, yes, this flips the mental train track, but the new track actually leads somewhere. Precisely. And the researchers tested this extensively using massive open data models, specifically models called OMO and Pythia, where the researchers had full access to every single piece of data the models were trained on. Right. They gave these models creative tasks. completing a narrative story, solving a physical MacGyver style reasoning problem where you have to use random objects to escape a room, and writing poetry. Let me look at that poetry example from the sources because it perfectly illustrates the tension between originality and quality. The prompt given to the AI was to write a poetic sentence that includes the word phases and ends in the word moon. So let's look at a low originality, high quality response, the AI outputs. Life moves in phases, ever-changing, like the moon. I mean, it makes sense. It follows the rules. It's grammatically correct. But it is a massive clichΓ©. It's basically memorized training data from a thousand inspirational posters. Right. High quality, very low originality. Now, the cat on the keyboard version, high originality, terrible quality. The AI outputs. Through phases, the moon renounceth life like the moon. It's definitely weird. You didn't expect the word renounceth, but it ignores the prompt's ending constraint, and it's just bad poetry. But when the AI hits the sweet spot, the high harmonic mean of meaningful novelty,
it outputs this:sewn with sharp gold and silver threads, like the ever-changing phases of the moon. Sewn with sharp gold and silver threads. That is gorgeous. It bisocites the visual of the moon with tailoring and metallic threads, and the AI came up with that. Yes, and what the researchers found when analyzing these open data models is fascinating regarding how we build better AI. They found that scaling up the model size moving from a 1 billion parameter model to a 7 billion parameter model reliably improved this overall novelty score. But wait, why does a bigger AI get more novel? Is it just hallucinating more interesting things? Actually, no. When they broke down the harmonic mean, they found that the originality score stays relatively stable as models get bigger. The bigger models improve the overall novelty primarily by drastically increasing the quality of the output. Oh, interesting. Yeah. They become better at maintaining coherence while navigating those surprising high-entropy pathways. They also found that post-training the RLHF process of aligning a model to follow human instructions increases novelty. But there is a massive catch mentioned in the research when it comes to trying to hack this novelty score during the generation process. The researchers looked at the U-shaped effect of varying the sampling temperature. Yes. In AI generation, temperature is a setting that controls how random or deterministic the model's output is. A temperature of 00 means the model will always pick the absolute most probable next word. It's incredibly boring. Right. If you crank up the temperature, you're forcing the model to take risks and pick less probable words. So theoretically, turning up the temperature should make the AI more novel. And it does, initially. The researchers found that as you increase temperature, originality skyrockets. The model stops relying on those memorized cliches. The novelty score climbs. But eventually you push it too far. Yes. If you keep turning up the temperature, you eventually hit a cliff. The quality suddenly crashes because the AI loses the thread of logic. The output becomes nonsensical. And because we are using the harmonic mean, the overall novelty score plummets. So when you graph this relationship between temperature and meaningful novelty, it forms a perfect inverted U-shape. Exactly. There is an optimal zone of creative risk. Alright, so we have established that we can mathematically score an idea. We have the math for meaningful novelty. But if we are building a durable economy based on these ideas, we have another problem. Which is? How do we separate the minor clever tweaks from the world changing foundational leaps? Because, I mean, a nice line of poetry isn't the same as the theory of relativity. Not all meaningful novelty is created equal. This brings us to a crucial concept in the research. the break with benchmark, and the economic printable of the durability of ideas.- Let's explore that because right now, the way our current economy rewards innovation is often incredibly flawed and short-sighted.- It is. In the current paradigm, we often conflate different levels of creativity. There is a vast amount of what researchers call combinative creativity. This is immensely valuable, but it is essentially taking two existing, well-understood ideas and mashing them together in a useful way. The sources use the Wright brothers as an example of combinative creativity. which sounds crazy because they invented the airplane. I know, it sounds counterintuitive. But the researchers argue they took bicycle mechanics, gears, chains, balance, which they understood deeply, and combined it with emerging theories of aerodynamics from gliders. Brilliant synthesis, but combinative. Contrast that with what researchers call a breakwith. A breakwith is an extreme rare form of creativity. It doesn't just combine past ideas. It establishes an entirely new paradigm. Yeah. It is a significant rupture with current modes of thinking. It makes a previous way of doing things obsolete. This is the Einstein moment, the Nobel Prize jump. Exactly. And if you are designing a bottom-up economy that rewards innovation... you have to be able to mathematically distinguish a combative tweak from a break with rupture to do this economists network scientists look at a very specific data set which is patent citation networks literally looking at how new patents cite older patents yes and they use two specific metrics to analyze this web of citations the breakthrough index or ki and the disruption index or CD okay what is the difference between those to the breakthrough index measures patents that have a very low similarity to prior patents meaning they are highly novel and weird but a very high similarity to future patents meaning they have a high impact they start a new trend that sounds exactly like a break with it's close but it doesn't capture the obsolescence factor the disruption index of the cd is the true measure of a break with the disruption metric asks a very specific ruthless question When a new invention is published, do future researchers cite this new invention, or do they continue to cite the predecessors that came before it? Oh wow, so if I invent a new type of battery, and for the next 20 years scientists cite my battery patent, instead of citing the people I learned from, I have disrupted the field. I've broken the chain of knowledge. Precisely. You have eclipsed the past. Your idea is a new foundation. And what's fascinating and slightly alarming, here is a historical trend highlighted in the sources. long disruptive jumps high CD scores are actually declining over time compared to breakthrough syntheses wait really yeah science is producing more and more combative improvements but fewer fundamental ruptures we are optimizing not inventing so if we want more ruptures our economy needs to reward them disproportionately and reward them durably this is where the power law scaling of citations comes into play Network analysis shows a rich get richer phenomenon in citations. Once a foundational paper hits a certain tipping point of influence, it begins to accumulate citations indirectly and exponentially. I mean, people cite it not even because they read the original paper, but because everyone else in their field is citing it. It becomes part of the atmosphere. Exactly. Let me use an analogy here to ground the economics. If someone invents the internal combustion engine, they get the brake with durability reward. It is a massive paradigm shift. Yeah. But if 50 years later, someone else invents a slightly better spark plug for that engine, they're doing combative creativity. Right. They deserve a reward, but they shouldn't be able to siphon off the lion's share of the original inventor's royalties just because their spark plug is the newest part of the car. Right. The economy needs to know who built the foundation. We can't let the reward be completely diluted by the thousands of minor follow-on engineers that just tweak the original concept. Right. That is the core mission of this proposed economy. The economic reward must disproportionately favor the original insight over time. But here's the monumental technical hurdle. To pay the right person, to pay the inventor of the engine, and not just the spark plug guy, the AI that is generating answers today, needs to know exactly where its knowledge came from. Yep. Inside a massive billion parameter neural network, which operates like a giant black box of numbers, how do we track the providence or provenance of an idea?
This is one of the most notoriously difficult challenges in computer science, but the sources present a breakthrough framework for solving it:Atomic Information Flow or AIF. Let's dive deep into AIF. What is it? It is a methodology specifically designed for retrieval augmented generation systems or array systems. These are AI models that don't just rely on their internal pre-trained memory. When you ask them a question, they actually go out and search external documents, tools, and databases to build an answer. Like when you ask an AI a question about today's news and you see it searching the web before it replies to you? Yes. An AIF works by fundamentally breaking down the AI's final output and the documents it retrieved into what they call atoms. Okay, hold on. How does an AI chop text into an atom? What is an atom in this context? An atom is defined as a minimal, indivisible, self-contained unit of semantic information. Think of it as a single distinct fact or concept. The system uses natural language processing to segment a paragraph into these discrete propositions. So it's not looking at words, it's looking at ideas. Yes, and then it models the entire AI orchestration system as a graph-based network flow. Visually, what does that look like? Imagine a massive web of nodes and connecting lines. The user's query, the question you asked, is placed at the very top as the super source. The final text response the AI gives you is at the bottom as the super sync. And what's in between the source and the sync? Supply nodes. These are the specific tools the AI used, the URLs it read, the databases it queried. AIF tracks the directed flow of these semantic atoms. from the supply nodes through the AI's reasoning steps down to the super sink. It creates a mathematical map of where every single concept originated. Okay, let's unpack this with a really concrete example because this is vital. Imagine I ask an AI to give me a recipe for a truly unique showstopping cake. The AI goes out, reads a bunch of culinary data, and generates a recipe. Without AIF, if I asked the AI where it got the recipe, it might just say, "I used a recipe database." Which is completely useless for our new economy. We have a cake, but we don't know who to pay for the intellectual property. Right. But with AIF, the AI chops its final recipe paragraph into semantic atoms. It traces the graph flow backward. It says, atom one, the specific ratio of flour to butter. That atom flowed directly from Farmer A's blog, atom two. The structural technique for whipping the buttercream icing. That atom flowed from Baker B's YouTube video transcript. Exactly. And atom three. The sudden bisociative twist of adding fermented lemon zest to the batter. That atom maps directly back to Chef C's copyrighted cookbook. we are literally tagging the specific molecules of knowledge that is exactly what AIF achieves it provides global provenance attribution it's not just pasting a citation link at the bottom of a page yeah it is tracing the origin the transformation and the specific usage of a semantic atom across the entire complex orchestration graph of the AI that is incredible and the sources mentioned we can even inject metadata into to this flow to make it smarter. Yes, AIF allows you to modulate the flow of these atoms based on auxiliary signals. You can weight an atom based on the authority of its source. A peer reviewed journal gets a heavier flow than a random blog. You can weight it by temporal freshness. How new is the information? Or you can weight it by uncertainty, penalizing atoms that the AI isn't confident about. And there are specific metrics the researchers use to evaluate how well this flow is working, right? They look at several flow heuristics. One key metric is groundedness. This measures the fraction of the semantic atoms in the final response that can be directly mathematically traced back to the supply nodes. It tells you exactly how much the answer is hallucinated by the AI versus supported by real creators. And for our bottom-up economy, the most important metric is tool contribution. Exactly. Tool contribution measures the proportion of the final response's flow that originated from one specific tool or document. It tells you exactly how much of the final value was provided by Chef C's lemon zest insight. Okay. Okay, so we have the tracking mechanism. We know exactly whose knowledge atoms were used to build the AI's answer. The AIF graph shows that Chef C contributed 15% of the conceptual atoms to the final recipe.
But here is the multi-trillion dollar question:How do we calculate how much money Chef C is actually owed? Because 15% of the text doesn't necessarily mean 15% of the economic value. What if that lemon zest was the only reason the cake won an award? What if the flour ratio was generic and easily replaceable, but the lemon zest was a break with innovation? This is where we cross over from computer science and natural language processing into the realm of cooperative game theory. To assign a financial value to an atom, we have to introduce the Shapley value. you the Shapley value in reading the sources this feels like the holy grail of this whole deep dive it really is the Shapley value was developed by Lloyd Shapley who won a Nobel Prize for it it was designed to solve a very specific thorny problem how do you fairly distribute the total gains of of a collaborative effort among the players based on their true marginal contributions. So let's use an analogy. Let's say three people, a plumber, an electrician, and an architect, collaborate to build a house, and they sell it for a million dollars. How do you divide the million dollars? You can't just split it three ways because they didn't all do the same amount of work or provide the same value. Right. And the Shapley value is mathematically proven to be the only solution that satisfies four
key axioms of fairness:efficiency. The entire million dollars must be distributed. Nothing is wasted. Symmetry. If two players contribute the exact same value to every possible combination of workers, they get paid the exact same amount. Demiplayer. A player who contributes zero value gets zero dollars. and additivity. If you combine two different games, the value is the sum of the parts. It sounds mathematically perfect for an AI economy. So why hasn't it been used for AI training data before? Why aren't we already paying people based on Shapley values? Because of the computational nightmare of traditional data Shapley. To calculate the exact Shapley value of a single piece of training data, say, Jeff C.'s cookbook, you theoretically have to retrain the AI model on every possible subset of the data, both with and without the cookbook, to see how the model's performance changes. And if you have a data set with a billion data points. You would have to train a massive neural network, too, to the power of one billion times. It is mathematically impossible. It would take longer than the lifespan of the universe. But the sources reveal a massive algorithmic breakthrough that solves this bottleneck. Yes. It is a technique called In-Run Data Shapely. How does it work? How do you avoid retraining the model a billion times? In-Run Data Shapely acts as a contribution accountant that operates during a single training run of the AI. It leverages the iterative nature of how neural networks learn. As the model trains, it updates its internal weights step-by-step through a process called gradient descent. The Enron algorithm monitors these steps and uses first and second order Taylor expansions to approximate how much each specific data point is changing the model's overall performance. Okay, you said Taylor expansions, and we need to unpack that. Right, sorry. A Taylor expansion is a mathematical way to estimate the value of a complex function by looking at its derivatives, its rate of change at a single point. So instead of shutting down a massive factory, removing one employee and restarting the whole factory from scratch for a year to see how valuable that employee is, you just constantly monitor the financial ledger and the employee's moment-to-moment output in real time as the factory operates. You estimate their impact based on their immediate trajectory. That is a phenomenal analogy. You are calculating their marginal trajectory without resetting the system. And to do this computationally efficiently across billions of parameters, they use advanced linear algebra techniques called ghost dot products and ghost gradient Hessian gradient products. I'll have to ask. What is a ghost dot product? It sounds like arcane magic. It's an algorithmic shortcut. Normally, multiplying massive matrices of data takes huge amounts of memory. A ghost product allows the system to calculate the result of that multiplication for an individual beta point without ever having to explicitly construct or store the massive intermediate matrices in the computer's memory. It gets the answer while bypassing the heavy lifting. So it's incredibly efficient. Remarkably so. It allows them to track the influence of individual data points with almost zero extra runtime overhead compared to standard AI training. Okay, so we now have a scalable contribution account. We can calculate the Shapley value of every data point. What happens when the researchers actually run this on real data? Because the sources mention a finding here that has massive implications for society. This is perhaps the most profound revelation in the entire stack of research. The researchers ran in-run data Shapley to evaluate the value of training data. But they did something clever. They evaluated the training data against a validation corpus that was completely rewritten. It wasn't verbatim text, it just covered similar semantic concepts, but with entirely different vocabulary and phrasing. So the AI's final output didn't look anything like the original training data. There was no copy tasting. Exactly. But when they checked the math, the second order in-run data-shapely value still ranked the original training corpus incredibly high. The math proved that contribution does not require memorization. It proved a direct causal link. The original data caused the model's ability to generate the new, rewritten, seemingly novel text. Think about the implications of this for you, the listener. Think about the massive billion dollar lawsuits happening right now around generative AI and copyright. Tech companies often stand up in court and argue our AI didn't copy your poem or your article or your code. It just read millions of things. Learn the vibe and wrote a totally different poem. Therefore we don't owe you a dime. But this mathematical framework completely dismantles that argument. Yes. Because if you write a brilliant poem and an AI reads it and learns the conceptual atomic structure of it and then writes a totally different poem for someone else, Inrun Shapley mathematically proves you still contributed. The causal link is intact. Your Shapley value is positive. You are owed your micropayment. We finally have the hard math to back up copyright and intellectual property in the AI age. And this Shapley framework doesn't just apply to passive training data. The researchers also apply it to active multi-agent AI systems. They introduce a framework called SHARP, which stands for Shapley-Based Hierarchical Attribution for Reinforcement Policy. How does SHARP work? Imagine you have a complex AI ecosystem where a planner AI agent acts as the boss, breaking down a complex user prompt, and several worker AI agents go out and execute the subtasks. If the final result is a massive success, who deserves the economic reward? The planner for devising a good strategy or the workers for good execution? It's the ultimate group project dilemma. Someone always does all the work and someone else slacks off, but they all get the same grade. Exactly. Traditional AI reward systems just broadcast a single uniform reward to every agent involved, which is confusing and wildly inefficient. Sharp fixes this using counterfactual masking. It essentially asks the mathematical question, what would the final outcome have been if we completely removed worker A from the team? It calculates the delta, the difference between the team with you and the team without you. Yes. By mathematically isolating each agent's causal impact through this counterfactual Shapely value, SHARP precisely distributes the reward only to the agents that actually move the needle. It filters out harmful or lazy interactions, and it encourages the planner agent to constantly select the most efficient, innovative workers. So let's look at what we've built so far. We have the AIF tracking to map the atoms of knowledge. We have the in-run Shapely math to calculate exactly who contributed what, even if the output is completely transformed and non-verbatim. We can finally pay the original human creators. But there is one final crucial piece of the puzzle. What's that? How do we ensure that the AI systems generating these queries continue to seek out new innovations? If an AI gets paid for giving a highly reliable, safe answer, won't it just stick to the established, boring, combative answer? That is a very real, well-documented danger. In standard reinforcement learning from Human Feedback, or RLHF, models are trained to align with human preferences. But humans often prefer safe, predictable, and offensive answers. We like our associative matrices. We like our habits. We do. And as a result, the models suffer from what researchers call a loss of output diversity. They become incredibly capable, but fundamentally boring. They lose their edge. To counter this and to build an economy that continually discovers new, break-width innovations, we have to code curiosity into the machine itself. But curiosity is an emotion. How do you code curiosity into a matrix of weights and biases? The research proposes a framework called Curiosity Driven RLHF, or CD-RLHF. And the engine of this framework is something called an intrinsic curiosity module, or ICM. Okay, let's break down the mechanics of the ICM. In standard training, the AI gets an extrinsic reward points given to it from the outside world, from human evaluators for getting the right answer. But the ICM ignores the outside world. It gives the AI an intrinsic reward. It rewards the AI based entirely on prediction error within its own latent space. Okay, define latent space. us."Laden Space is the internal, multidimensional geometric map where the AI organizes concepts. It's how it understands that king and queen are close together, but king and toasters are far apart." Got it. So how does prediction error generate curiosity? The AI constantly tries to predict what the next state or the next concept will be as it generates an answer. If it predicts correctly, the error is low. The math says, I knew this would happen. And the intrinsic reward is low. The AI is essentially bored because the path is predictable. Right. But if the AI makes a prediction based on its past training and the actual next state it encounters is totally different if the prediction error is high. Then the intrinsic reward is massive. A high prediction error means the AI has stumbled into a state that is novel or surprising. It is found in incongruity. The AI is financially and structurally rewarded for exploring this high curiosity potential. It is mathematically drawn to the unknown. Exactly. Okay, so we are basically coding wonder into the machine. We are programming it to seek out the bisociative jump. Yes, but here is the brilliant part. Just like a human, the machine habituates. As the AI explores this novel state more and more often, its ability to predict it gets better. The production error decreases, and therefore the intrinsic reward drops. It gets bored of the new toy. Exactly. It mirrors human psychological habituation. The math forces the AI to move on and constantly seek out entirely new novelties, perpetually expanding the frontier of its knowledge. But wait, I have to play devil's advocate again. If we're highly rewarding the AI for finding things it can't predict, couldn't it just start outputting total gibberish again? White noise is entirely unpredictable. The prediction error for pure static is 100%. You are describing the danger of structural collapse or reward hacking. The AI might realize the easiest way to get an intrinsic reward isn't to be creative, but to generate pure, meaningless noise. So how do the researchers stop that? How do you put a leash on mathematical curiosity? By heavily balancing the intrinsic curiosity reward with the extrinsic quality reward. They use advanced constraint techniques, specifically something called reward-weighted conditional flow matching. What does that do? It acts as an unbreakable guardrail. It mathematically forces the AI's exploration projectory to stay within the distribution of high quality human preferred text. It ensures that the AI is relentlessly searching for novelty, but only within the structural manifold of useful, coherent knowledge. Okay, let me summarize that. We are forcing the AI to be that incredibly curious kid who takes apart the toaster just to see how the wires work, but we are keeping the guardrails on so firmly that it doesn't accidentally burn the house down while it's exploring. That is a perfect description of CDRLHF. And the experiments detailed in the sources are incredibly promising. On complex tasks like text summarization and following intricate instructions, the curiosity-driven models produce significantly more diverse creative and novel outputs than standard models, while still maintaining extremely high alignment with human preferences. They didn't burn the house down, they just figured out a better way to make toast. This has been an unbelievable journey through this research. We started with the sheer human capacity for a punchline. The biological bissociation of two different worlds colliding to release cognitive free energy in the form of a laugh. We translated that visceral feeling, that aha moment, into mathematical surprisal metrics and the harmonic mean of meaningful novelty. We then looked at how to physically track the atoms of those novel insights through massive black box AI networks using atomic information flow. We figured out how to calculate the exact undeniable economic worth of those atoms using cooperative game theory, in-run data shapely values, and ghost dot products, proving mathematically that original creators are owed their due, even if their work is completely transformed by the machines. And finally, we explored how to build a framework using curiosity driven RLHF to ensure the AI doesn't just rest on our laurels, but remains inherently curious, perpetually seeking out the next great break with innovation. The vision outlined in these papers is breathtaking. We are talking about the architecture for a fully decentralized web. A web where a brilliant insight, whether it's a new mathematical theory of gravity, a brilliantly bisociated punchline, a unique culinary technique, or a beautifully crafted lyric is registered, and every single time a curious AI agent utilizes that specific cognitive pivot to answer a user's question, the provenance is traced through the graph, the Shapley value is calculated in real time, and a micropayment is seamlessly routed to the human creator. It is the blueprint for an economy that doesn't just reward cheap attention or fleeting virality. It permanently, durably rewards deep, meaningful human innovation. But I want to leave you, the listener, with one final provocative thought to mull over. If we actually achieve this, if we build a machine that perfectly mimics human curiosity, and we plug it into a global economy that financially rewards the discipleship, discovery of the unknown, what happens to our own human creativity? If the machine can bisociate and calculate Shapley values millions of times a second, do we get pushed out of the innovation game entirely? Or does being freed from the drudgery of routine, commendative thinking finally allow us all to spend our lives chasing nothing but pure, Nobel-level "aha" moments? Maybe the machine handles the spark plug so we can finally design the engine. Something to think about until next time.
Podcasts we love
Check out these other fine podcasts recommended by us, not an algorithm.
Hidden Brain
Hidden Brain, Shankar Vedantam
All In The Mind
ABC Australia
What Now? with Trevor Noah
Trevor Noah
No Stupid Questions
Freakonomics Radio + Stitcher
Entrepreneurial Thought Leaders (ETL)
Stanford eCorner
This Is That
CBC
Future Tense
ABC Australia
The Naked Scientists Podcast
The Naked Scientists
Naked Neuroscience, from the Naked Scientists
James Tytko
The TED AI Show
TED
Ologies with Alie Ward
Alie Ward
The Daily
The New York Times
Savage Lovecast
Dan Savage
Huberman Lab
Scicomm Media
Freakonomics Radio
Freakonomics Radio + Stitcher
Ideas
CBCLadies, We Need To Talk
ABC Australia