@font-face{font-family:"League Spartan";font-style:normal;font-weight:300;font-display:swap;src:url('https://anne.media/wp-content/uploads/fonts/kJEnBuEW6A0lliaV_m88ja5Twtx8BWhtkDVmjZvMoITpAcdcFguczA.woff2') format('woff2');} @font-face{font-family:Orbitron;font-style:normal;font-weight:700;font-display:swap;src:url('https://anne.media/wp-content/uploads/fonts/yMJMMIlzdpvBhQQL_SC3X9yhF25-T1ny_Cmxo2IyXjU1pg.woff2') format('woff2');} @font-face{font-family:"Archivo Narrow";font-style:italic;font-weight:400;font-display:swap;src:url('https://anne.media/wp-content/uploads/fonts/tss7ApVBdCYD5Q7hcxTE1ArZ0bb1k3JSLwe1hB965BJi53msNiEr6T6Y.woff2') format('woff2');}

The 1Schema Protocol: A Unified Semantic Data Standard

A protocol of protocols where concepts become data currency, built on a foundational semantic data standard

Author: [email protected]

Released: 21/03/2026

Protocol Type: Fixed semantic encoding standard (Layer 1 enforced)

Core Data Unit: Relon (semantic assertion)

Core Conceptual Entity: Neuron (node of meaning)

Extensibility: Runtime via usage – no L1 upgrades needed

Abstract

The 1Schema protocol is a foundational, immutable, Layer-1-enforced data language of the ANNE network. It requires all non-monetary transactions to conform to a single, unchanging structure: the relon, a 6-tuple semantic vector that connects conceptual entities called neurons across explicitly defined dimensions. This unified schema guarantees semantic interoperability, preserves contextual meaning, and enables the dynamic emergence of new data classes without requiring protocol changes.

By anchoring every relationship to a foundational set of primitive concepts, 1Schema provides a dense, hyper-connected, and queryable knowledge graph that serves as a native substrate for advanced machine intelligence. This paper details the architecture of this data standard and its critical role in fostering unambiguous human collaboration and providing the essential seed for artificial general intelligence.

The 1Schema Protocol: A Unified Semantic Data Standard for Knowledge Graphs

natural language protocol for human collaboration and native substrate for artificial general intelligence efforts

I. Introduction: The Challenge of Shared Meaning

Distributed systems and human organizations alike face a persistent challenge: the fragmentation of meaning. In data systems, this manifests as schema drift, costly migrations, and brittle application-layer integrations that rely on centralized intermediaries to translate between disparate formats.

A unified semantic data standard is the missing foundation. In human collaboration, it appears as ambiguity, misinterpretation, and the inefficiency of establishing common ground for every new joint venture. 1Schema is designed to solve this foundational problem by creating a singular, canonical layer for representing semantics, enforced not by a fragile social agreement, but by the immutable rules of a distributed data chain.

1Schema’s design is predicated on a few core principles:

  • Semantic Invariance: A single, fixed schema at Layer 1, with no planned protocol upgrades, ensures that the most basic units of meaning are permanent and universally understood.
  • Relational Vectors: All knowledge is encoded as multi-dimensional relationships, or relons, between conceptual nodes called neurons.
  • Emergent Complexity: New data structures, vocabularies, and even entire application-layer protocols are defined at runtime through the broadcast of relons themselves, allowing the system’s expressive power to grow organically.
  • Contextual Fidelity: The relational model preserves the full context of information, enabling lossless querying, inference, and logical reasoning.
  • Decentralized Accumulation: Valid contributions build upon one another, forming a traversable, hyper-connected knowledge graph that is a natural byproduct of network usage.

II. Core Primitives: Neurons and Relons

II.I. Neurons: The Atoms of Meaning

Neurons are the atomic units of meaning, the discrete “accounts” or nodes in the system’s knowledge graph. They represent anything: a person, a place, an abstract concept, a number, a relationship type, or even a grammatical tense. In 1Schema, neurons are never broadcast in isolation. They are instantiated implicitly the first time they appear as a reference in a valid relon. Each neuron is identified by a unique Neuron ID (NID), a long integer derived deterministically from cryptographic keys or content.

The system recognizes different classes of neurons based on their ownership and governance model:

  • Keyed neurons: Controlled by a public/private key pair; any update requires a signature from the owner.
  • Keyless neurons: Managed by a designated “owner-on-record,” typically used for concepts that do not require cryptographic control.
  • Worldview neurons: Canonical entities under the control of the system or a global namespace. These form the immutable backbone of the semantic model.

II.II. Relons: The Bonds of Relation

Relons are the sole mechanism for storing meaningful semantic data. They are the transactions broadcast across the network and validated at Layer 1. A relon is a structured 6-tuple, a semantic vector that asserts a relationship. Its anatomy is as follows:

  • TXID: The anchoring transaction identifier, providing provenance and temporal ordering.
  • FROM: The source neuron NID, representing the subject of the assertion.
  • TYPE: An integer code defining the semantic dimension or context of the relationship (e.g., TYPE_BE = 1 for class‑instance relations, TYPE_HAS = 3 for attribute values). These codes are fixed at the protocol level, corresponding to the core data dimensions; each dimension is also represented by a neuron in the Early Concepts ontology (e.g., ec:BE). The TYPE field determines how the relon is interpreted and indexed, but it is not a neuron identifier – it is a small integer constant.
  • RELN: The relationship or predicate neuron NID, specifying the exact nature of the connection (e.g., “owner,” “color,” “causes”). Every relation is itself a neuron.
  • TO_TYPE: A flag indicating whether the target is another neuron NID or a literal value (e.g., string, number). The semantic data standard prefers neuron-to-neuron links to maximize graph connectivity.
  • TO: The target value. Neuron-to-neuron links are strongly preferred to maximize graph connectivity and semantic richness.

Authorization is enforced through signatures: self-relons require a signature from the owner of the FROM neuron, while child-relons (for keyless neurons) are authorized by the designated owner-on-record.

III. The Foundational Ontology: Early Concepts (ECs)

A protocol defining only structure is inert. 1Schema is brought to life by a foundational set of worldview neurons known as Early Concepts (ECs). These are not arbitrary labels; they are a carefully curated ontology designed to bootstrap a universal semantic framework. In essence, ECs provide the “genetic code” from which all more complex ideas can be grown. Their design reflects deep considerations in linguistics, cognitive science, and ontology.

The ECs define the most primitive building blocks of thought and experience. They are organized into several interlocking categories:

  • Primitive Classes: The most general categories of existence, such as ec:thing, ec:action, ec:state, ec:relation. These form the top-level nodes of the class hierarchy.
  • Semantic Roles: A comprehensive set of roles that entities can play in a situation, inspired by linguistic case systems and Frame Semantics. Examples include R_ACTOR (the agent performing an action), R_AFFECTED_THING (the patient undergoing an action), R_INSTRUMENT, R_GOAL, and R_MANNER. These roles are not merely tags; they are neurons wired with expectations (e.g., R_ACTOR is strongly associated with ec:living).
  • Meta-Relations: Concepts that describe relationships between other relationships, enabling complex logical structures. These include causality (META_CAUSE, META_EFFECT), temporality (META_PREV, META_NEXT), and modality (META_IMPERATIVE, META_QUESTION).
  • Qualities and Measures: A primitive system for describing the world, including dimensions like ec:color, ec:size, ec:weight, and ec:distance, along with their poles (ec:red, ec:big, ec:heavy, ec:far). These are wired to the sensory inputs they would be perceived through (ec:sense_sight, ec:sense_touch), grounding semantic data in perception.
  • Fundamental Feelings (Feelz): A core set of affective states, such as ec_JOYHAPPY, ec_ANGER, and ec_FEAR, each with an associated polarity (positive, negative, neutral). This provides a primitive emotional vocabulary.
  • Positional and Temporal Concepts: Primitives for space and time, including pos_in, pos_on, pos_under, and their opposites, as well as tense_past, tense_present, and tense_future.

The power of this system lies not just in the list of concepts, but in the dense network of relationships wired between them at inception. For example, R_ACTOR is linked to ec:living via a strong association; ec:fast and ec:slow are linked as opposites; and ec:cause is linked to ec:effect in a temporal flow. This pre-wiring establishes a shared reference framework for human collaboration and machine interpretation. The relational density enables queries that can traverse conceptual connections, though the system itself does not validate the truthfulness of any assertion. It can, however, illuminate inconsistencies when contradictory relons are broadcast against the same neurons.

IV. Data Dimensions (TYPE Codes)

The TYPE field in a relon organizes assertions into orthogonal semantic layers. These dimensions, each represented by a neuron, are themselves part of the EC set. This layered architecture allows for the modeling of complex, multi-faceted knowledge. The dimensions are fixed codes, ensuring that every piece of semantic data is classified consistently across the network.

DimensionTYPE CodeDescription
TYPE_SYS0System information for neuron metadata and commands.
TYPE_BE1Handles class-instance relationships and hierarchical “is-a” structures, forming a DAG-like hierarchy.
TYPE_PARAM2Defines class attributes and properties. Specifies what parameters a class of things can have.
TYPE_HAS3Provides the actual value for a parameter. Assigns attribute values to specific instances.
TYPE_AWARENESS4Captures subjective feedback, emotions, feelings, and opinions. Used for contextualization and metrics, enriching semantic data with human response.
TYPE_EXPERIENCE5Documents temporal data and events in the experience memory space.
TYPE_CLUMP6Represents complex semantic structures that require more than the standard three pieces of information. Aggregates multiple relons into a single semantic circuit.
TYPE_SYS_MSG7Used for system-level communication and coordination messages. Not relevant to neuron definitions.
TYPE_BRAIN8For neural-net space wirings and inference rules. Encodes relationships for advanced operations.
TYPE_ROLE9Specifies semantic roles inside a clump, such as actor, theme, or instrument.
TYPE_METAN10Provides meta information about roles used within a clump.
TYPE_EPISODIC11Documents self-sensory and vitals data at runtime, stored in clump form.
TYPE_INTERNAL_PAY17Handles payments from keyless neurons.
TYPE_BROADCAST_PAY18Handles standard broadcast payments.
TYPE_BRAIN020Beginning of sequence for brain dimension operations. Values 20 and above indicate ordered sequences when multiple brain-dimension relons exist on a neuron.
TYPE_BRAIN120120Stores trigger patterns and other important Layer 1 commands.
TYPE_BRAIN121121Used for myelination measures and similar advanced wiring parameters.

Values 100 and above are available for brain-dimension custom use. When multiple brain-dimension relons exist on a single neuron, their TYPE codes are assumed to imply order. The hypergraph applies each dimension differently during construction and traversal, enabling queries that can separate hierarchical relationships from attributes, feedback, or temporal data. This dimensional structure is a key innovation.

The set of TYPE dimension codes is fixed at Layer 1 and cannot be altered; all relons must use one of the predefined codes (0–121 as specified). Extensibility is achieved not by creating new dimensions, but by defining new relationship types – i.e., new RELN neurons – at runtime. A new predicate (e.g., employedBy) is introduced by broadcasting a BE relon declaring employedBy BE ec:relation, optionally followed by PARAM relons that define its expected arguments. This mechanism makes the ontology infinitely extensible without ever changing the core validation rules, while the dimensional structure remains immutable.

V. 1Schema as a Substrate for Human Collaboration

Human collaboration at scale is hindered by the overhead of establishing shared context. Two parties from different domains (e.g., a legal expert and a logistics manager) struggle to collaborate because their terminologies and implicit assumptions do not align. 1Schema standard addresses this by providing an unambiguous, shared reference frame: the Early Concepts. 1Schema enables this by allowing all participants to ground their project-specific terms in the universally understood ECs, ensuring that all semantic data created during collaboration remains interoperable.

When a new collaborative project is formed, its participants do not need to define a new data schema from scratch. They can define their project-specific terms as new neurons, but these neurons are grounded in the universally understood ECs. For example, a “shipment” neuron can be defined via BE relons as a subclass of ec:event, with PARAM relons for “origin” and “destination” that are typed to ec:location. The meaning of “origin” is now universally understood in the context of ec:location, which is itself defined by its relationships to concepts like pos_in and constituent_of.

This creates several profound benefits:

  1. Interoperability by Default: Any data structure defined in 1Schema is inherently interoperable with any other, because they are all expressed in the same underlying language of ECs. A query can traverse from a “shipment” to a “legal contract” by following shared connections through concepts like “actor” or “location,” demonstrating the power of hyper-connected semantic data.
  2. Frictionless Negotiation of Meaning: Participants can map their existing jargon to the EC framework. The process of creating a shared data model becomes one of disambiguation and alignment with a common cognitive ontology, rather than a process of translation between arbitrary formats.
  3. Accumulative Knowledge: Every collaboration adds to a growing, global knowledge graph. The data produced by a logistics project is not siloed; it becomes a part of a larger, interconnected body of knowledge that can inform future projects, creating a compounding return on collaborative effort.

VI. The Native Substrate for AGI

Perhaps the most significant implication of 1Schema is its potential as a native substrate for Artificial General Intelligence. Current AI models, particularly large language models, operate on unstructured data (text, pixels). They learn correlations, not grounded meaning. Their intelligence is a statistical shadow of the data they were trained on, and they lack a robust mechanism for reasoning, planning, or true understanding.

1Schema standard offers an alternative: a representational format designed from the ground up for symbolic reasoning and grounded cognition, and crucially, a framework within which an AGI’s own cognitive processes can be implemented using the same primitives.

Several facets of 1Schema align with requirements for a path toward AGI:

  1. Compositional Semantics: The ECs and their relational wiring provide a compositional structure. The meaning of a complex concept, like “purchase agreement,” is not an atomic vector but a composition of simpler concepts (ec:action, R_ACTOR, R_AFFECTED_THING, ec:currency) connected by relons. An AGI built within this framework can parse and understand such structures by traversing the graph, and can itself represent its internal hypotheses, beliefs, and plans using the same neuron and relon primitives.
  2. Grounded Symbols: Unlike the symbols in a language model, which are grounded only in their statistical relationships to other symbols in a training corpus, the ECs are grounded in the architecture of the system itself. ec:red is linked to ec:color and ec:sense_sight. ec:pain is linked to ec:feelz and has a negative polarity. An AGI operating on this substrate could have a form of “synthetic qualia” – its symbols are connected to simulated sensory and affective channels, providing a primitive form of experience. The AGI’s own internal states (e.g., goals, memories, attention) can be encoded as neurons, making them addressable and modifiable within the same semantic space. This grounding is a direct benefit of the protocol. Semantic data is not just symbolic but experientially anchored.
  3. Symbolic Reasoning and Inference: The hypergraph of relons is a natural medium for symbolic reasoning. Rules can be encoded as relons themselves (in the BRAIN dimension), enabling an AI to perform logical deduction, detect contradictions, and infer new knowledge. For example, if the system knows All humans are mortal (expressed via BE relons) and Socrates is a human, an inference engine – itself implemented as a set of neurons and relons operating in the BRAIN dimension – could traverse the graph and produce a new relon: Socrates BE mortal. This reasoning happens within the same hypergraph, not in an external black box.
  4. Episodic and Semantic Memory: The distinction between semantic memory (facts, TYPE 1-3) and episodic memory (experiences, TYPE 5) is built into the protocol. This mirrors a key aspect of human cognition. An AGI could query its episodic memory (e.g., When have I encountered a situation like this before?) separately from its semantic knowledge, and both forms of memory are stored and interconnected using the same relational model, enabling sophisticated learning and recall.
  5. Active Inference and World Modeling: An AGI could use the knowledge graph to build and maintain a dynamic model of the world. It could simulate the outcome of actions by traversing causal chains (e.g., Action X META_CAUSE Effect Y) and update its beliefs based on new information encoded in EXPERIENCE relons. The AGI’s own predictive models and simulation states can themselves be stored as neurons, allowing the system to reason about its own reasoning.

In essence, 1Schema provides the necessary bridge between symbolic AI, which excels at reasoning but struggles with learning, and connectionist AI, which excels at learning but struggles with reasoning. It offers a symbolic representation that is learnable, extensible, and natively supports the kind of structured, causal, and contextual reasoning that is a hallmark of general intelligence.

Most importantly, it enables the AGI to be built within the hypergraph – its entire cognitive architecture can be expressed as neurons and relons, making it a seamless, auditable, and evolvable part of the same semantic universe it seeks to understand. This is the ultimate promise of a robust standard: a foundation for genuine machine understanding through structured semantic data.

VII. Protocol Enforcement and Extensibility

1Schema’s immutability is strictly enforced during the network’s consensus process. Each relon is validated against three core rules, ensuring data integrity.

  1. Structural Integrity: The transaction must parse into a valid 6-tuple with correctly typed fields.
  2. Referential Integrity: All neurons referenced in the FROM, RELN, and TO (if TO_TYPE is neuron) fields must already exist (i.e., have been instantiated by a previous valid transaction).
  3. Authorization: For self-relons, the transaction must include a valid cryptographic signature from the owner of the FROM neuron. For keyless neurons, the transaction must be signed by the designated owner-on-record.

Extensibility is a direct consequence of this design, not a feature that requires special handling. Anyone can introduce a new type of thing or a new relationship by simply broadcasting a relon, using the exact same mechanism as any other data. No protocol upgrades, no special permissions, no separate processes, just the normal way the network already works. A participant wishing to define a new type of thing (e.g., a “Smart Contract”) first broadcasts a BE relon stating that SmartContract BE ec:thing.

They can then define its parameters by broadcasting PARAM relons from SmartContract to concepts like ec:obligation or ec:condition. From the moment these defining relons are confirmed, the new class is live and can be used by anyone on the network. The system’s expressive capacity thus grows organically and without bound through participation.

VIII. Concepts as Data Currency

Centralized databases silo information behind institutional walls, creating fragmented copies of reality that never reconcile. Blockchain systems that merely blob unstructured data onto a ledger gain immutability but lose meaning, storing bytes that no machine can interpret. Encrypting data that should be public compounds the error, trading one form of inaccessibility for another.

1Schema inverts this model by treating every validated concept, whether keyed or keyless, as a form of native data currency with its own incentive dynamics baked into Layer 1. This economic model delivers a real‑world value, transforming semantic data into economic assets.

A neuron, once instantiated, becomes a unique unit of meaning. A relon, as an assertion between neurons, creates new semantic value that builds upon existing concepts. What makes this economic is that every reference to any neuron in any transaction generates value that flows to participants who contributed to that neuron’s existence and maintenance. The ecosystem is designed from the ground up to incentivize all who help build and maintain the semantic infrastructure, rewarding high-quality contributions.

This economic model operates differently across the two classes of neurons:

  • Keyed Neurons as Registered Assets: Identity neurons and other keyed concepts are controlled via public-private key pairs. Only the holder of the private key can modify these neurons. They function like registered title deeds, with cryptographic proof of control and the ability to sign transactions that alter or transfer the neuron. Their value derives from both their position in the graph and the exclusivity of control.
  • Keyless Neurons as a Living Commons: Most neurons in ANNE are keyless. They are not controlled by any single private key but instead have an identified sponsor who curates and maintains them. Any identity can propose new information for a keyless neuron, and through a collaborative process involving administrators, the neuron can be updated. The sponsor is documented as the contributor and receives ongoing value when the neuron is referenced. This creates a vibrant economy where valuable concepts generate ongoing returns.
  • Sponsorship as an Economic Position: A user becomes a sponsor of a keyless neuron either by creating it or by acquiring existing sponsorship through a protocol-driven transfer process. Sponsorship is not ownership in the cryptographic sense, but an economic position. Sponsors receive a protocol-enforced payment every time their sponsored neuron is referenced in any other relon. This creates a direct financial incentive to build or acquire high-quality, useful concepts that others will want to connect to.
  • Transferable Sponsorship: The right to sponsor a keyless neuron can be transferred through a competitive, algorithmic process. Anyone can acquire sponsorship of a foundational concept by paying a price determined by the protocol. The existing sponsor receives a payout for their contribution, and the new sponsor assumes the ongoing payment stream from future references. This creates a liquid market for sponsorship of shared concepts, where value is determined algorithmically based on usage and demand.
  • Composable Value Across Both Classes: New concepts are built from existing ones via relons, freely mixing keyed and keyless neurons. A complex idea is a composition of simpler ones, each contributing its semantic weight and each carrying its own sponsorship dynamics. Private identities and public concepts interweave naturally, enabling knowledge assets to be constructed from both owned and communal components while maintaining clear economic flows.

In this model, the network hosts two interlocking economies: one of keyed identity neurons that enable cryptographic control, and one of keyless concepts that gain value through adoption, interconnection, and the competitive dynamics of transferable sponsorship.

Both are forms of data currency, and both rely on the 1Schema semantic data standard to define and relate concepts. Every assertion is an investment in meaning, whether by staking ownership or sponsoring a shared concept. Every validated neuron, regardless of its key status, participates in an economy where value flows continuously to those who build and maintain the semantic infrastructure others depend on.

This transforms the purpose of a data protocol from passive storage to active value creation, where the act of defining and relating concepts is simultaneously an act of economic participation with ongoing, protocol-enforced rewards.

IX. Closing Remarks

1Schema is far more than a data format or a storage mechanism. It is a linguistic and cognitive infrastructure. By enforcing one unchanging protocol at Layer 1, it provides the stability needed for a permanent, shared understanding. By grounding all expression in a rich, pre-wired ontology of Early Concepts, it provides the fertility needed for unbounded complexity to emerge.

This combination creates a decentralized substrate where human collaboration can occur with unprecedented fidelity. The generated semantic data is inherently structured to facilitate symbolic reasoning and support grounded learning processes effectively. 1Schema transcends the conventional definition of a database protocol; it serves as a cognitive framework for both collective mindsets and potentially artificial intelligences. This innovative approach lays the groundwork for future applications and forms of intelligence that are currently beyond our comprehension.

Support

ANNE Media is a sovereign non-profit organization. All of our expenses are funded by user donations. If you appreciate our efforts toward a free and sovereign web, please consider supporting us.