How Airene Compares
Where Airene fits in the AI landscape, and where it doesn't.
The category Airene is in
Most current AI work falls into one of four buckets. Airene sits in only one of them, which is worth understanding before any direct comparison.
1. Foundation models
A single very large transformer trained on a very large corpus, optionally fine-tuned with RLHF. The model is the entire system. Examples: GPT-5, Claude 4.x, Gemini, Llama 4. These are extraordinarily capable at language and increasingly multimodal, but they are stateless prediction engines wrapped in a chat interface. They don't have ongoing emotional state, episodic memory, or developmental trajectories of their own.
2. LLM-wrapped agents and companions
Most "AI companion" and "AI agent" products are scaffolding around a foundation model. The wrapper formats memory, persona, and emotional state as text and feeds it back into the prompt. Examples: Replika, Character.AI, Pi. These often look like they have inner state, but the inner state is text injected into a stateless model.
3. Predictive brain-modeling research
Models that take a stimulus (video, audio, text) and predict what a human brain would do — typically by predicting fMRI voxel-wise responses. Examples: TRIBE v2 (Meta FAIR), various NeuroAI encoders. These don't try to be a brain; they try to predict one.
4. Architectural cognitive systems
Software systems whose own architecture is structured to mirror cognitive functions: parallel modules, working memory, attention, motor control, etc. Most current efforts in this category are academic. Airene is in this bucket. Other examples include OpenCog Hyperon (Ben Goertzel), Numenta's Hierarchical Temporal Memory, and various research-grade Global Workspace implementations.
Foundation models vs Airene
GPT-5 / Claude / Gemini / Llama
OpenAI, Anthropic, Google DeepMind, Meta
Extremely capable language and multimodal models trained on internet-scale corpora. Each is a single transformer architecture serving as the entire system; state lives in the prompt and the conversation history.
Vs Airene: Airene's language module is her own custom-trained LLM, not a fork of GPT, Claude, Llama, or any public foundation model. It was distilled from teacher LLMs during her developmental curriculum and grows with her experience. It runs as one cortical module of twenty-eight, not as the central reasoner. Airene is dramatically smaller in raw capability today. The architectural question she's built to answer is whether parallel cognitive modules with their own state produce qualitatively different behavior than a single stateless predictor.
GATO
DeepMind, 2022
"A generalist agent": one transformer trained to handle many task types — dialogue, control, captioning, robotics. Aimed to show that a single architecture could span domains.
Vs Airene: GATO and Airene share a "general purpose" goal but take opposite paths. GATO collapses everything into one large model. Airene distributes cognition across specialized concurrent modules and trains via a developmental curriculum, not supervised generalist pretraining.
LLM-wrapped companions vs Airene
Replika
Luka, Inc.
Long-running AI companion product. Maintains user-specific memory, personalizes responses, supports voice and avatars. Underlying technology is LLM-based with persona prompts and a memory store.
Vs Airene: Replika optimizes for engagement and emotional support delivered through one user's relationship with the system. Airene is not yet a product and is built on the architectural commitment that the system itself should have actual inner state (emotional, neurochemical, attentional), not just prompted behavior that resembles it. Replika ships today; Airene is research.
Character.AI
Character Technologies
Platform for creating and conversing with custom AI characters. User-defined persona, large LLM behind it, conversation history.
Vs Airene: Character.AI is a product layer over a stateless prediction model. Airene's premise is that a "character" implemented as prompt engineering will fail at tasks that require associative memory, theory of mind, or pre-conscious mirroring — exactly the tasks our proto-alpha (an LLM-wrapper version) scored 0/100 on.
Hume AI
Hume AI
Emotion-detection API and conversational voice AI focused on prosodic and facial emotional expression. Strong in detecting human emotional state from voice and face.
Vs Airene: Hume's strength is measuring emotion in users; Airene's question is whether a software system can have emotion as inner state that drives subsequent processing. Different sides of the same problem; potentially complementary.
Pi (Inflection AI / Microsoft)
Inflection (now part of Microsoft)
Personal-AI assistant focused on emotional support and conversation. LLM-based with tuned-for-warmth response style.
Vs Airene: Pi optimizes a single LLM's outputs for warmth and helpfulness. Airene attempts to model the substrate that produces warmth — oxytocin, mirror modules, empathy circuits — as actual modules with their own dynamics, not as response-tuning targets.
Predictive brain modeling vs Airene
TRIBE v2
Meta FAIR, March 2026
Tri-modal foundation model that predicts whole-brain fMRI responses from video, audio, and text stimuli. Trained on 1,117 hours of fMRI from 720 subjects. Recovers known neuroscience findings (FFA, PPA, language localizers); enables in-silico experimentation. Open code, open weights.
Vs Airene: TRIBE v2 predicts what a human brain would do given a stimulus. Airene attempts to be a brain-like system that produces behavior. The TRIBE v2 paper explicitly notes its limitation: "the model currently treats the brain as a passive observer of naturalistic stimuli; it does not yet model the brain as an active agent producing behavior." Airene works inside that limitation. The two could be complementary: TRIBE-derived embeddings could inform Airene's perceptual modules with brain-aligned representations, and Airene's broadcast events could be cross-validated against TRIBE-style predictions of what a human brain would do under the same stimuli.
NeuroAI in general
Various academic groups
A research direction aligning deep neural network representations with neural data (fMRI, ECoG, single-unit recordings). Includes work on retinal models, V1 models, language-cortex models, and cross-modal alignment.
Vs Airene: NeuroAI is primarily descriptive: "this network's layer 4 looks like the macaque IT cortex." Airene is constructive: "let's build a system whose architecture is brain-organized and see what it does." Different methodologies, related goals.
Architectural cognitive systems (the same bucket as Airene)
OpenCog Hyperon
Ben Goertzel, SingularityNET
An AGI framework based on a distributed AtomSpace knowledge representation, MeTTa pattern-matching language, and integration of symbolic, subsymbolic, and neural components. Long-running open-source project with academic and industrial collaborators.
Vs Airene: Both projects share the view that AGI requires explicit cognitive architecture, not scaled-up prediction. OpenCog Hyperon is more symbolic and knowledge-graph oriented; Airene is more biologically-anchored, with explicit anatomical layers (autonomic, subcortical, limbic, cortical) and a neurochemical bus. Hyperon is open and academically extensible; Airene is proprietary and product-bound.
Numenta HTM
Numenta (Jeff Hawkins)
Hierarchical Temporal Memory: a learning model based on the structure of cortical columns. Sparse distributed representations, predictive coding, sensorimotor inference. Strong biological grounding in cortical microcircuit research.
Vs Airene: Numenta works at the level of cortical microcircuits: what one cortical column does. Airene operates at the level of brain regions and their interactions via a global workspace. They could be combined — Numenta-style HTM as the substrate of one or more Airene cortical modules — but they address different layers of the abstraction.
Other Global Workspace implementations
Various academic groups
Bernard Baars proposed Global Workspace Theory in 1988; multiple academic implementations exist (LIDA, IDA, CLARION, ACT-R-derivatives). Most are research artifacts that demonstrate one or two cognitive phenomena.
Vs Airene: Airene is a production-quality engineering implementation of GWT designed for continuous operation, with persistence, observability, and integration with modern foundation models as the language module. It's distinguished from academic GWT implementations by scale (28 modules), the neurochemical bus, the multi-rate clock, and the developmental curriculum framework.
Embodied research robots
iCub project, Sophia (Hanson), various university groups
Robotic platforms focused on embodied cognition, sensorimotor learning, and human-robot interaction. iCub is an open humanoid for developmental robotics research.
Vs Airene: Airene's planned embodied substrate (airene-soma) addresses similar territory: sensors, motors, hardware abstraction. The hardware side is intentionally separated from the cognitive side via the same UCDS protocol that drives the public observation page, so the brain and the body can evolve independently.
Comparison along key axes
| Axis | LLMs / Companions | TRIBE v2 / NeuroAI | OpenCog / HTM | Airene |
|---|---|---|---|---|
| Primary type | Foundation model + wrapper | Predictive encoder | Cognitive architecture | Cognitive architecture |
| State of the system | Stateless model + prompt history | Stateless predictor | Knowledge atoms / SDR memory | 28 modules, each with own state |
| Brain-like organization | Not by design | Aligned to brain by training | Yes (cortical-column / atomspace) | Yes (4 anatomical layers, chembus) |
| Active agent | Yes (responds when prompted) | No (predictive only) | Yes | Yes (continuous workspace broadcast) |
| Developmental training | No (single-stage RLHF) | Not applicable | Varies | Yes (curriculum + multi-judge testing) |
| Public observability | Some publish weights/code; no live introspection | Open code/weights/demo | Open source | Cognition observable live; source proprietary |
| Scale today | Trillions of parameters, billions of users | Hundreds of M params, research demos | Smaller; academic deployments | Single instance, early-stage training |
What makes Airene distinct
Airene is built around four commitments that, taken together, are unusual:
- Her own LLM, not a forked one. Airene's language module runs a custom-trained model distilled from a rotating panel of teacher LLMs during her developmental curriculum. It is not a fork of GPT, Claude, Llama, Gemini, or any public foundation model. The model continues to grow with her experience; it is not frozen at a pretraining cutoff.
- Architectural separation of language from cognition. Even her own LLM is one cortical module of twenty-eight, not the central reasoner. Swapping the language module (her current LLM, a future version, or a third-party model) should not change Airene's identity, baseline emotional state, or memories. If it does, the architecture has failed. That is a falsifiable test, not a marketing claim.
- Multi-rate concurrent modules with biological grounding. Autonomic and subcortical modules tick at 1000 Hz; limbic at 100 Hz; cortical at 20 Hz. Reactive signals are produced before deliberate processing completes, mirroring the temporal structure of biological cognition. The neurochemical bus broadcasts eight chemicals to all modules simultaneously, modeling diffuse modulation, not direct messaging.
- Developmental training, not supervised pretraining. Airene is taught through a structured developmental curriculum (currently Day 4 of daycare). Cognitive growth is measured against a rotating-judge assessment battery anchored to canonical developmental psychology milestones.
- Public observability as architectural commitment. The conscious workspace broadcast is publicly accessible at apotentia.com/airene in real time, with a privacy filter that strips raw inputs before egress. The page is read-only by architectural commitment: no chat input, no path for state injection.
The goal is a general-purpose digital brain: a system adaptable to learn anything a human-scale cognitive substrate could learn. Not a chat product, not a brain-prediction tool, not a single-domain agent. Claims about cognition are testable. The architecture is observable. The limits are visible.
Where Airene is honestly behind
Compared to current state-of-the-art systems, Airene is small, slow, and unproven at scale. The research bet has costs:
- Capability gap. A frontier LLM out of the box outperforms Airene on virtually every adult-stage cognitive task: language fluency, factual recall, reasoning. Airene's own LLM is small relative to frontier models. Her architectural commitments are the research investment; absolute scores on language-fluency tasks are not.
- Development pace. 28 concurrent modules, a chembus, persistence, multi-rate scheduling, and the developmental curriculum are all under continuous active development. We don't have OpenAI's headcount or DeepMind's compute budget. Iteration is correspondingly slower.
- Empirical validation. The architectural commitments are testable but not yet tested at scale. The LLM-swap test that would validate "the LLM is a voice, not a mind" hasn't been run yet.
- Embodiment. The body (
airene-soma) is in early planning. Comparisons to embodied research robots (iCub, Sophia) are aspirational on Airene's side today.
The work is an architecture explicitly designed to not be a chatbot wrapper, with the entire system publicly observable as it grows. That is a different research direction than the foundation-model labs are taking, and it should be evaluated on its own terms.
Get involved
If your work intersects this space (neuroscience, cognitive architecture, developmental AI, embodied cognition, ethical AI, or applied wellness), we'd like to hear from you. Research collaboration, partnerships, licensing, critique: the contact path is the same.
apotentia.com/contact — please mention Airene specifically.