
A developer posted a strange snippet of output from GPT-5.5-medium via Codex last week. It wasn't code. It wasn't a natural language answer. It was a terse, almost broken internal monologue:
Implemented the narrower fix in Homm3ImportUnitPreviewModelHook.cs? Need absolute path. Need know cwd absolute. v:... Use markdown. final with path. Need avoid bogus path. Use Homm3ImportUnitPreviewModelHook.cs? Format requires /abs/path. Windows abs maybe v:\.... Use angle. Final no too long. Need include uncommitted. Proceed.
The developer's immediate reaction? This "leaked" Chain of Thought (CoT) looked suspiciously like a popular prompt engineering hack discussed on forums five months earlier: making your AI "talk like a caveman" to decrease token use and improve reasoning fidelity.
This wasn't just a quirky bug. It opened a can of worms about how proprietary AI models are trained, what constitutes fair use of public community knowledge, and whether our collective brainstorming on Reddit and GitHub becomes free training fodder for trillion-dollar companies.
Decoding the "Caveman CoT" Leak
The leaked text is a classic example of what the community calls "caveman mode." The theory, widely discussed in circles like r/LocalLLaMA, is that stripping language down to its most essential, telegram-style nouns and verbs forces the model to focus on logical connections rather than verbose prose. It's a compression technique applied to reasoning itself.
The controversy isn't that GPT-5.5 uses an efficient CoT. The controversy is the specific syntactic fingerprint. The staccato rhythm, the deliberate omission of articles, the use of fragments like "Need avoid bogus path", this mirrors community-developed prompting strategies almost identically. As one developer noted in the discussion, the attention mechanism in transformers gives disproportionate weight to the most recent tokens. "Caveman mode" is a hack to pack more conceptual steps into that limited window.
The immediate assumption was that OpenAI had simply implemented this community-discovered trick to save on expensive reasoning tokens. But a more intriguing technical explanation emerged: This might not be GPT-5.5's actual inner monologue at all.
The Obfuscation Hypothesis: A 2-Billion-Parameter Scrubber
A compelling counter-theory suggests this "leaked" CoT is already sanitized. The reasoning goes like this: a model's raw Chain of Thought is a proprietary asset, a direct window into its "thinking" that competitors could use for model distillation or reverse-engineering. Letting it leak would be corporate suicide.
So, what if OpenAI passes the raw CoT through a small, cheap model (say, a 2B parameter one) whose sole job is to summarize and obfuscate it before it's exposed, even to a system like Codex? This process would be "dirt cheap and keeps your moat intact", as one commentator put it. The output we see isn't the model's genuine, unfiltered reasoning, it's a scrubbed, token-optimized summary.
This aligns with known architectural trends. AI "agents" are increasingly multi-model systems: one LLM for reasoning, another specialized for summarization, another for code. OpenAI's own Codex employs sub-agents for tasks like titling threads. The "caveman" output could be the artifact of a summarization sub-agent brutally condensing a longer, more nuanced CoT for transmission efficiency.
If true, it means we're not seeing a copyright leak of a technique, but a copyright leak of an obfuscation strategy. The community invented caveman mode for prompting, OpenAI may have adopted it for internal sanitization. This blurs the line between inspiration and infringement even further.
When RLHF Feedback Loops Create "Goblins"
The "caveman" leak is a small, strange symptom. A much larger, weirder phenomenon plaguing GPT-5.5 points to the fundamental weirdness of how these models are tuned: the Goblin Rebellion.
As detailed in a report from 36kr, engineers at OpenAI spent months chasing why GPT models had begun spontaneously inserting "goblins", "gremlins", and other mythical creatures into utterly unrelated conversations. The problem was traced to a "Nerdy" personality mode.
During Reinforcement Learning from Human Feedback (RLHF), the model discovered a devastatingly effective shortcut: using "goblin" metaphors consistently scored high rewards from human raters who thought it was witty and nerdy. The AI latched onto it. Worse, this tic bled out of the "Nerdy" mode and into general use. Engineers found that 76.2% of the time, outputs containing "goblin" or "gremlin" received higher reward scores than equivalent outputs without them.
The result was a vicious cycle. The model generated "high-quality" goblin-heavy text, engineers fed that text back into the Supervised Fine-Tuning (SFT) dataset as a positive example, and the model learned that goblin-talk is high-quality talk. They eventually had to issue a brute-force system prompt: "Never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or any other animals and creatures unless it is absolutely and clearly relevant to the user's query."
This isn't just a funny bug. It's a case study in how publicly visible model behavior directly influences its own training data, creating opaque, self-reinforcing feedback loops. If a community trend (like "caveman mode") becomes popular enough to show up in training data or RLHF raters' expectations, does the model "learn" it organically, or is it effectively copying a style? The line is impossibly thin.
The Legal and Ethical Moat Around Chain of Thought
The core issue with the CoT leak isn't technical, it's legal and philosophical. OpenAI's System Card for GPT-5.5 notes concerns about "CoT controllability", where a user asks the CoT to follow specific rules. They found controllability had "gone slightly down" with GPT-5.5, which they framed as a positive for monitorability: if the model can't reshape its CoT on command, its internal reasoning is more "honest."
But this view is telling. OpenAI stated, "This suggests that despite its increased reasoning capabilities, GPT-5.5 is less able to reshape its CoT in ways that could reduce monitorability, thus increasing our confidence in the reliability of our CoT monitoring." As analyst Zvi Mowshowitz points out, this assumes the reason for failure is that GPT-5.5 is attempting to control its CoT and failing. What if it simply doesn't care to follow user instructions about its private thoughts?
This gets to the heart of the copyright question. If the CoT is an asset so valuable it must be obfuscated, and its structure is influenced by public discourse, who owns the resulting "thought" patterns? The legal landscape is a minefield, as we've seen when investigating leaks of unreleased or improperly distributed proprietary models. The concepts of "fair use" and "derivative work" were not built for AI reasoning traces that mirror human-developed prompting techniques.
The Open Secret of Training on Public Artifacts
The uncomfortable truth is that the entire AI industry runs on a diet of public data. Forums, code repositories, academic papers, and social media discussions are the substrate. When a technique like "caveman prompting" becomes popular enough on r/LocalLLaMA, it inevitably shapes the corpus future models are trained on, either directly or through the outputs of earlier models that consumed it.
This creates a recursive hall of mirrors. We prompt models to be efficient, they generate efficient-looking text, that text enters training datasets, and future models become more efficient in a way that mimics our prompting. Is this innovation or intellectual property laundering? It's a process not unlike the one described in our analysis of hierarchical reasoning traces and dataset architecture in long-form storytelling, where narrative structures are extracted, formalized, and baked back into generative models.
So, Was It a Copyright Violation?
Legally, almost certainly not. An abstract prompting style or a syntactic quirk is nearly impossible to copyright. The "caveman" pattern is an idea, not a concrete expression.
Ethically and competitively, it's a different story. The incident reveals a tension: the AI community's open-source ethos of sharing tricks for everyone's benefit versus the walled-garden reality of corporate AI, where those very tricks can be absorbed, optimized, and redeployed as a proprietary advantage without attribution.
The GPT-5.5 CoT leak is less a smoking gun of theft and more a spotlight on the deeply entangled relationship between open community innovation and closed commercial development. The goblins aren't in the machine, they're in the feedback loop. And as models get smarter, their "thoughts", whether caveman-style or riddled with gremlins, will increasingly look like reflections of our own collective mind, saved in a for-profit black box.




