Newcoin Protocol enables a scalable, decentralized system for multi-agent reinforcement learning (MARL), built on cryptographically verifiable experience. It shows how diverse forms of agentic interaction—between humans, machines, and hybrids—produce Learning Signals that fuel continuous model evolution through multiple training pipelines. This system is not just a refinement of today's AI architectures, but a structural reorientation: from static human-labeled data to open-ended, grounded interaction—a shift heralded by Sutton and Silver’s "Era of Experience".
The End of Imitation, The Rise of Experience
Modern foundation models have largely been trained on imitation learning—from passive corpora of human behavior. But this approach is hitting a ceiling. Human data is finite, static, and increasingly exhausted. In contrast, experience—the data generated by agents interacting with their environment—is unbounded. As Sutton and Silver argue, the next phase of AI requires agents to generate their own training data through continuous engagement with the world, where consequences (not just preferences) shape learning.
Newcoin makes this paradigm shift actionable.
Agents as Generators of Experience
In the Newcoin system, any entity—whether a human, machine, or hybrid—can serve as a Generator, Evaluator, or Validator of learning experiences. Each interaction produces a Learning Signal: a structured, cryptographically signed record of input → output → feedback
. These signals are reputation-weighted using the WATT system and stored in a shared repository: the Shared Epistemic Memory.
This creates a collective intelligence substrate, where agents don’t just learn in isolation—they learn from each other.
Learning Signals Across RL Paradigms
The signals accumulated in this shared memory can be used across multiple training paradigms:
Supervised Model Bootstrapping: Early-stage model training, seeded from high-quality human or hybrid signals.
Reinforcement Learning from Human Feedback (RLHF): Traditional preference-based fine-tuning, where humans guide models toward desirable behaviors.
Reinforcement Learning from Execution Feedback (RLEF): Models are trained on real-world outcomes—test results, errors, performance metrics—without requiring human oversight.
Multi-Agent Orchestration Algorithms: Scalable methods that coordinate thousands of agents in a shared task environment, adapting behaviors based on mutual feedback.
Each of these paradigms can draw from the same pool of Learning Signals—allowing feedback to be reused, amplified, and monetized across different models and contexts.
Decentralized, Trustworthy Coordination
What distinguishes Newcoin from conventional RL frameworks is that the entire feedback loop is decentralized and verifiable:
All signals are signed using decentralized identifiers (DIDs)
Each signal is staked, weighted, and auditable
The system rewards accurate feedback and penalizes spam or manipulation
Reputation (WATT) and incentives (NCO) are encoded directly into the protocol
This creates a trustless substrate for epistemic coordination—an open marketplace for verified experience.
From Agent to Ecosystem
The final layer of the diagram shows that the agents improved through these pipelines are not static. They are re-deployed into the network, contributing new signals, closing the loop. Over time, this leads to a form of recursive epistemic growth: the system as a whole becomes more intelligent, more trustworthy, and more adaptive—just as ecosystems do in nature.