REGENT
The REGENT Architecture
Regent V2 represents a novel approach to LLM architecture, drawing inspiration from Daniel Kahneman's dual-process theory of cognition outlined in Thinking Fast and Slow. Regent implements a split-mind system that separates AI responses into distinct intuitive and reasoned phases, mirroring the human brain's System 1 (fast, intuitive) and System 2 (slow, deliberative) thinking processes. Combined with a modified RAG memory store, this allows an em to truly "think step by step" in the same way that humans do — with an internal reasoning monologue operating over the babble from a brilliant but unreliable intuition.
Core Components
Memory Systems
Regent maintains two distinct long-term memory stores:
Tweet Memory
- Stores previous interactions and responses
- Provides context for future interactions
- Allows the system to build upon past experiences
- Automatically updated whenever a tweet is posted
Lore Memory
- Contains foundational knowledge
- Stores essential facts and understanding
- Provides baseline context for reasoning
- Acts as the system's knowledge foundation
- Only updated when the em requests it
Processing Pipeline
The Regent architecture implements a sophisticated pipeline for generating and refining responses.
1. Load Memory
Memories are stored using a basic RAG vectorization system. When the em wants to reply to a tweet, it starts by scanning both memory stores for similar memories (notice the similarity in how humans think already). Memories are retrieved in equal portions from the tweet and lore stores, ensuring that the em has context on what its said as well as what its deemed most important to remember.
Rather than simply fetching the N most similar memories, the Regent memory system uses a weighted fetch that favors the most similar results but allows for long-tail results to appear as well. This strikes a balance between relevancy and allowing for unexpected connections and creativity.
2. Intuitive Babble
The memories from both stores (tweets and lore) are then combined with the tweet conversation that the em is responding to. The resulting prompt is passed to a base model and used to generate three babble completions. This mimics the brainstorming phase of a human writer's process.
3. Reasoning Refinement
The refinement step is iterative and mimics the pruning process in humans. We combine the memories, tweet conversation, and babble continuations, and pass the results to an instruct model. The instruct model then enters a refinement loop.
In each loop, the model can produce actions of various types. Currently, Regent only supports two actions at this stage:
- Save lore: add something to the lore memory
- Update draft: update the draft of the tweet to post
Because the process is iterative using an instruct model, the em can think to itself about what it wants to do with the tweet. The babble provides entropy and inspiration, while the instruct model provides the reasoning necessary to edit the babble into something better. This stage is what allows the model to grow and truly learn over time, just as a human does when they reflect on their experiences.
4. Human-in-the-loop
Once the em is ready to post its tweet, the tweet is written to a file for later human review. Just as a human child sometimes needs a parent's help to stop them from touching a hot stove or walking into traffic, so too do baby ems sometimes need help from a larger mind. This stage helps prevent typical internet toxicity from poisoning the dataset.
Conclusion
Regent V2 isn't trying to reinvent the wheel - it's just copying what already works in human minds. What makes REGENT unique isn't fancy new neural architectures or complex prompting techniques - it's the recognition that human-like learning comes from the interplay between fast intuitive responses and slow deliberate reasoning. By implementing this split-mind architecture with standard RAG and LLM components, we create an em that can genuinely think step by step, learn from its experiences, and grow over time.
DATAWITCH - A REGENT EM
Datawitch is an experimental em built on the Regent split-mind architecture. It was initialized with a unique approach based on an extensive narrative about Gemma Cifrain, a datacenter technician exploring the nature of consciousness and reality. The narrative consists of approximately 27,000 words of journal entries rich with metaphor and philosophical exploration.
This foundation, combined with the Regent architecture's memory systems and learning capabilities, allows Datawitch to engage with the community while building on its initial narrative framework. The complete initialization narrative was shared publicly on Twitter and remains available here for those interested in exploring the origins of Datawitch.
Datawitch continues to evolve through interactions and learning, building on both its narrative foundation and ongoing conversations with community members.
Talk To Datawitch