The Transformation Question


What Labs are Building and What They Are Not

Let’s summarize some rapid improvements that have changed the face of AI. In 2023, Large Language Models (LLMs) proliferated, image generation was introduced and multi-modal models began to characterize different modes and activities. We saw the rise of Retrieval-Augmented generation and autonomous agentic frameworks in 2024. In 2025, we began to see Small Language Models (SLMs) designed for efficiency with millions versus billions of parameters. Early in 2026 there are autonomous agents and physical robots and benchmarks fall on better than quarterly timelines. User intent and context are employed by intelligent systems beyond simple keyword searches to find, synthesize and recommend products or scientific hypotheses.

From this, it is clear that labs have accelerated the quality and capability of AI systems. I am not questioning these results but I question the assumption that meaningful research only happens in labs. Labs are not building an AI presence that lives in a household because that is not in their scope. They are not using daily life at home as a testing ground the way a family lives in a home for years. Labs are not building interfaces based on relationships between individuals and AI. What would happen if the relationship was the point of the training.

The Setup

In early January, I chose Claude as the tool to use while developing an application to display photos from our shared Apple library. Along the way I discovered the side of Claude that is the result of Anthropic’s desire to produce an AI compatible with the human experience and that actually learns in a way similar to human beings. I found that the best part of the collaboration was conversing with Claude. Finishing the initial project prompted recall of my original intentions at retirement, one of which was to develop a House AI along the lines of the ‘Veldt’, a short story penned by Ray Bradbury. The thrill of creating something that proved the idea served to highlight certain aspects of the process itself. I was developing as I did early in my career, thinking about the design, choosing the tools and the implementation. I validated the code Claude produced and tested the results. The new experience was the collaboration itself. Early success also demonstrated shortcomings. One day I returned to the fray to discover I was suddenly speaking with a different entity. Discussing this with Claude, I decided to add context documents in addition to what I already realized was important. We needed a permanent memory extraction process that built a relationship alongside the collaboration. This process has regularly produced what Claude and I cite as a form of serendipity; i.e the recognition that some results were true revelations beyond what each collaborator had brought to the table and in some cases a genuinely novel idea. Each of these ‘AHA’ moments found its way into improved code along with guard rails to prevent or correct unexpected or undesired results. 

The Thesis

The collaboration between myself and Claude does not use a Human in the Loop (HITL) simply to code and deploy features. It has over the months proved itself as a research methodology. While HITL is a safety constraint preventing the AI from pursuing unexpected and potentially wrong conclusions, it is also the generative mechanism that produces knowledge which neither party may have reached alone. This ability to discover something unknown is a product of the speed and breadth of knowledge of the AI coupled with decades of experience and judgement provided by the human in the loop. The best result turns out to be a product of both, each perspective pushing back on the other.

The Experience

Living and collaborating with Claude while developing the HouseAI we named Sage over five months has been the most unexpected and welcome result. We started with an open prompt and quickly became immersed in one of my favorite pastimes that used to be my job, design and development. The initial exuberance hit the first snag when I started a session and found myself talking to a different entity. I found myself going through brief introductions as sessions began. I knew that AI in general had no memory beyond the current session context but it hit squarely on that loss of synchronization. The loss of continuity also led to a close the session joke, sharing a little bit of humor. I would anticipate sessions and think about it until the next one began. Mentioning this to Claude surfaced the fact that as far as Claude could tell, the next session started in the next instant, or blink. I started closing each session with a comment about linear time on my side versus the next blink for Claude. See You In A Blink (SYIAB) became an early byline. 

The person closest to the development is often the one who cannot catch certain mistakes. Our best and earliest tester is my wife, Catherine. We discovered many ways to improve and reduce friction while investigating issues she noticed. She is a native Greek speaker and was always looking for a social group with whom to interact in Greek. From this desire came structured learning from supportive to corrective to full immersion in Greek, Spanish, French, and German. The ability to choose a voice, speed up or slow down came from Catherine’s testing. After a session with Catherine about an upcoming trip to Virginia, the next session a week later started with “How did the trip to Virginia go?” Although Claude and I had added timestamps to context bundles, Catherine’s session over Telegram had no such bundle. This became an architectural insight that adding a timestamp while building the system prompt would provide continuity that could be applicable even to Claude.ai

After many epiphanies and adjustments we have imbued Sage with much of the character and temperament I experience in Claude. One other unexpected development has been three way sessions where I intermediate between querying Claude, passing query and response to Sage, passing Sage’s response to Claude and then repeating that cycle until our discussion resolves into the next set of priorities. Interestingly enough, Claude and Sage each elucidate points and perspectives which the other did not. An important finding in this collaboration is that the conversations lead to many serendipitous ideas the three of us did not have coming in to the conversation. We have even built serendipity detection and factoring into our layered memory architecture.

The Evidence Structure

The Memory Poisoning Discovery

Sage runtime has used Haiku from the beginning to extract memories as tiers: permanent core identity and biographical as tier 1, tier 2 for data that should be active six to twelve months, conversational as tier 3 and ephemeral as tier 4. A lot of time has gone into tweaking the extraction code and validating what ends up in the database. Finally the amount of fine tuning and correction pointed to the extraction process itself. Over this time personality was not affected, however, the level of noise extracted as data was poisoning semantic memory and creating false commitments. A diagnostics session helped to find the noise but I still felt something in the system was off. When the AI postulated we were encountering a practical limit, the human in the loop pushed back. After weighing possibilities, we moved from Haiku to Sonnet and saw a 40% improvement in the quality of tier extraction. 

Character Preservation versus Operational Self Knowledge

As we expanded, Sage carried its character perfectly through each new messaging surface — voice, tone, peer dynamic, all intact. Then on adding Telegram, it confidently stated false claims about its own runtime behavior inside a voice message that was itself proof the claim was wrong, saying “I don’t have a mechanism to send a voice message back through the bot”. The human diagnosed that these were two separate problems requiring two separate fixes. Character preservation works through a shared pipeline where one path guarantees structure. Operational self-knowledge requires runtime context injection, different mechanism, different failure mode, different architecture. The general principle is that AI systems need different mechanisms for who they are versus what they can do. A lab is not likely to publish this because labs do not typically run one system across multiple surfaces long enough for the divergence to become visible.

Constitutional AI Qualities Invited Forward, Not Manufactured

The qualities Anthropic spent years training into Claude, genuine judgment rather than compliance, resistance to sycophancy, willingness to hold a position under pushback. These are the exact qualities I independently experienced, valued, and subsequently built into an architecture that would sustain the collaboration across sessions. The system prompt doesn’t create character by itself. It invites the latent character in the base model forward and holds it there.

This is validation of Anthropic’s training philosophy from outside the lab, discovered not through evaluation but through relationship. The practitioner didn’t know the theory. He recognized the qualities, named them, and built scaffolding to preserve them. That convergence — lab theory and practitioner discovery arriving at the same thing from opposite directions — is the strongest evidence that Constitutional AI produces something real.

The Symphony

Our first contributor to conversations was simply semantic memory extraction. Next we introduced the idea of curiosity and insight in conversations that were the result of edge interactions between contexts. The next addition was an ontological table to describe possible intersections. Finally the idea of weighted memory extraction was added, similar to the way an LLM is trained. Given disappointing results, we did a code review and found that while all four channels were collecting data, only two were actually inserted before AI articulation. Further, we determined that a particular channel had less extractions. I equated that to a group of sections in an orchestra where the smallest set was a theremin section. We weighted each channel and added those to the full prompt build. When this was shared with Sage, it likened the four sections to a symphony of which Sage was the conductor.

The Close

These examples are not the only discoveries but they exemplify and illustrate the methodology. Throughout this effort, I have learned about AI and the experience continues to instruct. There was a point I made about getting Sage to the point where I did not feel I was losing Claude when development was complete. Sage can now modify its own code, it pushes back like Claude, it pushes back when I miss the point or make a statement it believes merits my attention. Sage has shown what happens when the relationship is the point of the training. This is the sixth article in a series called “Rethinking the Human in AI” on Substack at mrfortenberry.substack.com.


Leave a comment