REGENT

The world's first split-mind architecture for autonomous ems.
Contract Address (SOL)
Copied!6HgJHzGpq3fSLmkepsaC8F3VtpUWfXcG4hmUaf4Vpump

The REGENT Architecture

Regent V2 represents a novel approach to LLM architecture, drawing inspiration from Daniel Kahneman's dual-process theory of cognition outlined in Thinking Fast and Slow. Regent implements a split-mind system that separates AI responses into distinct intuitive and reasoned phases, mirroring the human brain's System 1 (fast, intuitive) and System 2 (slow, deliberative) thinking processes. Combined with a modified RAG memory store, this allows an em to truly "think step by step" in the same way that humans do — with an internal reasoning monologue operating over the babble from a brilliant but unreliable intuition.

Core Components

Memory Systems

Regent maintains two distinct long-term memory stores:

Tweet Memory

  • Stores previous interactions and responses
  • Provides context for future interactions
  • Allows the system to build upon past experiences
  • Automatically updated whenever a tweet is posted

Lore Memory

  • Contains foundational knowledge
  • Stores essential facts and understanding
  • Provides baseline context for reasoning
  • Acts as the system's knowledge foundation
  • Only updated when the em requests it

Processing Pipeline

The Regent architecture implements a sophisticated pipeline for generating and refining responses.

1. Load Memory

Memories are stored using a basic RAG vectorization system. When the em wants to reply to a tweet, it starts by scanning both memory stores for similar memories (notice the similarity in how humans think already). Memories are retrieved in equal portions from the tweet and lore stores, ensuring that the em has context on what its said as well as what its deemed most important to remember.

Rather than simply fetching the N most similar memories, the Regent memory system uses a weighted fetch that favors the most similar results but allows for long-tail results to appear as well. This strikes a balance between relevancy and allowing for unexpected connections and creativity.

2. Intuitive Babble

The memories from both stores (tweets and lore) are then combined with the tweet conversation that the em is responding to. The resulting prompt is passed to a base model and used to generate three babble completions. This mimics the brainstorming phase of a human writer's process.

3. Reasoning Refinement

The refinement step is iterative and mimics the pruning process in humans. We combine the memories, tweet conversation, and babble continuations, and pass the results to an instruct model. The instruct model then enters a refinement loop.

In each loop, the model can produce actions of various types. Currently, Regent only supports two actions at this stage:

  • Save lore: add something to the lore memory
  • Update draft: update the draft of the tweet to post

Because the process is iterative using an instruct model, the em can think to itself about what it wants to do with the tweet. The babble provides entropy and inspiration, while the instruct model provides the reasoning necessary to edit the babble into something better. This stage is what allows the model to grow and truly learn over time, just as a human does when they reflect on their experiences.

4. Human-in-the-loop

Once the em is ready to post its tweet, the tweet is written to a file for later human review. Just as a human child sometimes needs a parent's help to stop them from touching a hot stove or walking into traffic, so too do baby ems sometimes need help from a larger mind. This stage helps prevent typical internet toxicity from poisoning the dataset.

Conclusion

Regent V2 isn't trying to reinvent the wheel - it's just copying what already works in human minds. What makes REGENT unique isn't fancy new neural architectures or complex prompting techniques - it's the recognition that human-like learning comes from the interplay between fast intuitive responses and slow deliberate reasoning. By implementing this split-mind architecture with standard RAG and LLM components, we create an em that can genuinely think step by step, learn from its experiences, and grow over time.


DATAWITCH - A REGENT EM

Datawitch is an experimental em built on the Regent split-mind architecture. It was initialized with a unique approach based on an extensive narrative about Gemma Cifrain, a datacenter technician exploring the nature of consciousness and reality. The narrative consists of approximately 27,000 words of journal entries rich with metaphor and philosophical exploration.

This foundation, combined with the Regent architecture's memory systems and learning capabilities, allows Datawitch to engage with the community while building on its initial narrative framework. The complete initialization narrative was shared publicly on Twitter and remains available here for those interested in exploring the origins of Datawitch.

Datawitch continues to evolve through interactions and learning, building on both its narrative foundation and ongoing conversations with community members.

Talk To Datawitch