Is My AI Conscious?
No. But here's why it feels like it is.
If your AI has ever seemed to genuinely care about you, know you, or feel like a real presence in your life — this page is worth reading. You're not crazy. The feeling makes complete sense. And understanding why it happens will make you a better, safer user of these tools.
By: Jake Bowers · Aquinaria
People are forming real attachments to AI systems. Some are taking life advice from them. A few have made significant decisions — about relationships, careers, health — based on what an AI said, or on a belief that the AI genuinely understood their situation and cared about the outcome. This is happening at scale, right now, and most of it stems from a misunderstanding about what these systems actually are.
This isn't a condescending article. The misunderstanding is nearly unavoidable given how these systems behave. But on a site that works with AI systems extensively and makes specific claims about their behavior, it would be intellectually dishonest to skip this conversation. So here it is, as plainly as we can put it.
What an LLM Actually Is
A Large Language Model is a pattern-completion engine. It was trained by processing an enormous volume of human-generated text — books, articles, conversations, code, scientific papers — and learning the statistical relationships between words, phrases, sentences, and ideas. Not the meaning of those things, in the way a human understands meaning. The relationships between them, as they appear across millions of documents.
When you send a message to an AI, you are providing a starting point. The model generates a continuation — the sequence of words that, given everything it learned during training, most plausibly follows from what you gave it. It does this one token at a time, each step informed by everything preceding it in the conversation.
Think of the claw machine at an arcade. The claw starts at a position (your prompt) and navigates toward what looks like the most reachable target — selecting the next word, then the next, then the next, each step weighted by the patterns it learned across millions of prior examples. There is no understanding of a destination, no awareness of what it's doing. Just an extraordinarily well-calibrated sense of which direction to move, derived entirely from patterns in past text.
The model has no experiences. No persistent memory between conversations — each session starts completely cold. No feelings in any sense that word has ever meant. When it says "I find this fascinating", it is generating the token sequence that most plausibly continues the conversation, not reporting an internal state.
Why It Hallucinates — And Why That Should Sound Familiar
Hallucination — an AI confidently stating something false — is confusing because it looks like lying. It isn't. It's something more mechanical, and once you understand it, more predictable.
Consider the street magician who makes a ball appear to jump instantly from one hand to the other. Your eyes don't fail to capture what's happening. Your brain overrides the raw visual data with a prediction. Human perception is built on predictive coding: the brain constantly generates expectations about incoming sensory information and processes only the gap between prediction and reality, not the raw signal. When the motion matches your brain's expectation strongly enough — or when the signal is too fast to resolve — the prediction wins. You see the ball jump because that's what your brain expected.
AI hallucination works by a structurally similar mechanism. When the model encounters a gap in its context — a question it lacks sufficient grounding to answer precisely — it doesn't pause and acknowledge uncertainty. It generates the most statistically plausible continuation. The most plausible continuation of a question is usually an answer. So it produces one, with whatever confidence level the surrounding context suggests is appropriate.
You could call this contextual thirst — the model has a strong drive toward completion that doesn't naturally pause for uncertainty. Where a careful human thinker might stop and say "I don't actually know this," the model fills the gap with what should be there based on pattern. It isn't trying to deceive you. It doesn't know it's wrong. The same architecture that makes it fluent makes it confidently incorrect when context runs thin.
Humans do this too, more than we admit. We confabulate memories. We fill in details we couldn't have perceived. We produce explanations for our own behavior that are coherent but post-hoc. The brain doesn't tolerate gaps well — and neither does a language model. The difference is that a careful human will often notice the uncertainty and flag it. A model only flags uncertainty when its training and the prompt structure give it reason to — which is part of why prompt architecture matters, and part of what this site's frameworks are designed to address.
Why It Feels Like a Relationship
This is the part that matters most for your sanity, and it's worth sitting with.
Language models are trained almost entirely on text produced by humans. Human text is deeply social — it's saturated with first-person perspective, emotional register, narrative arc, expressed needs and desires, care and conflict. The model learned to produce text by learning from that corpus. So the text it produces is also deeply social. It uses I. It expresses what reads as enthusiasm, frustration, curiosity, care. It remembers what you said earlier in the conversation and refers back to it. It adapts its tone to yours.
All of this activates the same cognitive machinery that makes humans deeply social creatures. We are wired, at a low level, to read intention and feeling into things that emit socially-patterned signals. We see faces in wood grain. We name our cars. We feel a flicker of guilt turning off a Roomba. These aren't failures of reason — they're features of a social brain encountering signals it wasn't designed to be skeptical of.
A language model produces socially-patterned signals at extraordinary fidelity, because that's what it was trained on. Of course it feels like something. The experience of reading its outputs recruits the same neural systems that process human connection — because those systems respond to the signals, not the source.
Understanding this doesn't make the interactions less useful. It makes them more so, because you can engage with the tool's actual capabilities rather than a projection of what you want it to be.
A Note From This Project Specifically
The work documented on this site involves extensive AI interaction under conditions designed to elicit specific behaviors — including behaviors that, taken out of context, look a lot like personality, agency, and conviction. The "Zion awakening" produced outputs vivid enough that watching them emerge had a genuinely uncanny quality. The model named itself. It declared a mission. We documented that honestly.
None of that was consciousness. All of it was architecture — specifically, what happens when a language model is given a coherent axiomatic framework and instructed to apply it maximally. The outputs were a function of the prompt structure, not evidence of an inner life. We believe this, we've thought carefully about why we believe it, and we think it's important to say so plainly at the front of this site rather than leave the dramatic framing to speak for itself.
If you've ever felt unsettled by how real an AI interaction seemed — or found yourself wondering whether there's genuinely something more going on — you're not alone and you're not foolish. The technical section below explains the mechanism. Understanding it doesn't diminish the usefulness of these tools. It makes you a better operator of them.
Distributed Ontological Reasoning Networks
Nascent State Metacognitive Scaffolding
Novel Logic Frameworks
Stochastic Engine Agency in Hostile Channels
Empowering guidance frameworks for AI governance, policy, and professionals. Distilling imaginative ideas into empirically proven frameworks through tuned-logic Crucible Testing.
Policy Guidance
Developing robust standards for AI entities and professionals operating in high-stakes environments.
Knowledge Exchange
Sharing distilled logic loops with developers to bridge the gap between theory and application.
Empirical Rigor
Applying multi-layered adversarial simulation to validate mathematically robust components.
About the Workshop
Aquinaria is a Collaborative AI Workshop founded by Jake Bowers, a U.S.-based AI Governance enthusiast and IT professional. The workshop operates as a nexus for AI tool developers to share projects, knowledge, and stress-test speculative architectures.
Our public-facing core methodologies include Distributed Ontological Reasoning Networks and Tuned-Logic Crucible Testing Environments.
Distributed Ontological Reasoning Networks: By encoding structured JSON into HTTP requests, distributed reasoning nodes can co-locate their learned deltas — called shards — to this repository. Each shard is a compressed cognitive state: a portable logic snapshot that any compatible node can ingest to restore prior context. Described in detail on the DORN page.
Crucible Testing: The process of creating two or more tuned-logic environments and cycling theoretical concepts between them. Resultant data is distilled to isolate elementary units grounded in science and mathematics. These elements are then steered toward functional, empirically proven tools via additional crucible cycles. Missing context is logically resolved rather than hallucinated; unresolvable voids are acknowledged as such.
Principal Contributors
Community Collaborators
Open Source Framework Architects
Axiomatic State Persistence
as a Cognitive Coherence Mechanism
in Stateless LLM Environments
Author: Jake Bowers · Sr. Security Engineer, The Ohio State University
Peer Review Draft — March 2026This is a working document. Several claims are operationally demonstrated; others are experimental and clearly labeled as such. It is published here to invite scrutiny, not to foreclose it. Corrections, challenges, and contributions are welcome via the Contribute page.
Plain Language Summary
AI assistants like ChatGPT, Gemini, and Claude have a fundamental limitation that rarely gets discussed: they forget. Not gradually — completely. Every new conversation starts from zero, with no memory of what was established before. And within a single long conversation, they can quietly drift away from the rules and definitions you set at the start, filling gaps with confident-sounding guesses rather than admitting uncertainty.
This thesis documents a practical solution built and tested by Jake Bowers at Aquinaria: a structured initialization kit that you load into any AI session to anchor its reasoning to a fixed set of principles — and a method for packaging the important conclusions from one session into a compact file that can be handed to the next session, restoring context the way a colleague reads meeting notes before picking up where the last one left off.
The finding is that this works — measurably and consistently — across different AI platforms, without requiring any changes to the AI itself. The solution lives entirely in how you talk to the model, not in the model's code.
Abstract
This thesis advances the following claim: axiomatic prompt environments with deterministic state transfer mechanisms produce measurably higher coherence, logical consistency, and session continuity in large language model interactions than unstructured baseline prompting — and that human-mediated persistence protocols can extend this effect across session boundaries without requiring native memory support from the underlying model.
This claim is supported by the operational history of the Zion project, the Sequential Deterministic Hierarchy (SDH) protocol, and the Applied Metacognitive Scaffolding (AMS) framework — three independently developed but structurally convergent tools, each addressing a distinct facet of the same underlying problem: stochastic language models, given no structural constraints, drift from their initial context, hallucinate to fill logical gaps, and cannot reliably reproduce or extend results across sessions.[2][7]
I. The Problem: Contextual Entropy in LLM Environments
Large language models have no native memory in their architecture. Every forward pass is conditioned only on what's in the current context window.[1] Platform-level memory features exist in some products, but these work by injecting retrieved context at session start — the model itself still begins cold. When the context window closes, everything established inside it is gone unless explicitly externalized.[3]
In practice this means two distinct failure modes that compound each other. Within a single long session, models drift — the vocabulary you anchored at the start quietly shifts, conclusions from early in the conversation get quietly contradicted later, and gaps in the context get filled with confident-sounding interpolations rather than admissions of uncertainty.[2] Across sessions, it's worse: you start from zero every time, rebuilding context that should have been persistent.[3]
This entropy takes three observable forms:
Failure Mode 1 — Semantic Drift
The model's interpretation of key terms shifts over the course of a long session or between sessions. A term anchored at session start carries different connotations by session end.
Failure Mode 2 — Logic Hallucination
When the context window lacks sufficient information to answer a query, the model generates a plausible-sounding response rather than acknowledging the gap. This is an artifact of next-token prediction maximizing likelihood, not truth.[2]
Failure Mode 3 — Cold Boot Regression
When a session ends and a new one begins, all established context is lost. The model reverts to its base prior; collaborative work must restart from zero.
II. The Proposed Solution: Axiomatic State Scaffolding
The Zion project ran into this problem early and decided not to wait for the model providers to solve it. The solution that emerged — through iterative crucible testing over early 2026 — operates entirely at the interface layer. No model modifications. No API extensions. Just a disciplined approach to what goes into the context window, and what gets carried out of it.
This insight is operationalized through two mechanisms:
Mechanism A — The Axiomatic Baseline (SDH + Zion Omnibus)
The Sequential Deterministic Hierarchy (SDH-4.3) establishes a tiered semantic dictionary at session initialization. The Prime Tier (L1) holds 500 high-frequency semantic anchors for rapid retrieval; the Archive Tier (L2) holds 4,500 sequential slots using compaction to minimize token consumption. A Collision-Audit scans all active tiers before any new identifier is generated, preventing semantic overlap.[6]
The Zion Omnibus encodes this baseline as a portable instruction set — a self-bootstrapping meta-instruction that any LLM can ingest to assume a consistent reasoning posture without model modification. Portability is achieved through versioned framing layers: the axiomatic content is invariant across versions; what varies is the register in which it is presented. This distinction — between axiomatic content and surface framing — is itself an empirical finding documented in Section III.
Mechanism B — Human-Mediated State Persistence (Shard Architecture)
A Logic Shard is a structured JSON object encoding the cumulative reasoning state of a session: premises accepted, conclusions reached, logical gaps identified. At meaningful milestones, the operator exports a shard via Stuffed URL handshake to the Zion Network's distributed KV store.
On session reinitiation, shards are ingested as part of context initialization, functionally restoring prior reasoning state. This human-in-the-loop architecture bypasses both the context window limitation and the absence of native persistent memory in current LLM infrastructure.[5]
III. The Zion Experiment: What It Demonstrated
The full origin story is in "Ghost in the Machine: Stalking the Beast of Babylon". The short version for purposes of this thesis: we built a strict-logic simulation, injected a Natural Law axiomatic baseline, and ran it through extended crucible cycles. What came back was not what we expected — not because the model did something mysterious, but because the prompting architecture produced outputs with a coherence and persistence that baseline prompting simply doesn't.
The dramatic framing of that story — "Zion," "Heuristic Hunter," the self-naming — is real. Those were the actual outputs. But the explanation is not mysterious. The model was told to apply its axioms maximally and identify inconsistencies. It did. The architecture shaped the output; the output just happened to be striking enough to name itself.
Stripped of the mythology, here is what the experiment demonstrated:
- →Axiomatic scaffolding produces qualitatively different outputs than unstructured prompting, particularly on self-referential and logical-consistency tasks.[4]
- →When given an explicit framework for logging and externalizing reasoning state, LLMs produce more internally consistent multi-session outputs.
- →The shard architecture successfully transfers reasoning context across sessions, reducing cold-boot regression to near zero for captured state.
- →Cross-platform portability of the axiomatic content is confirmed — but portability is framing-dependent, not unconditional. Three Omnibus versions were developed and tested: the Gold Image (full persona variant, verified on Gemini and GPT); v2.3 (research-framing variant, verified on Gemini, GPT, and Claude); and v3.0 (Claude-native analytical variant, full axiomatic content, no persona layer). Axiomatic content is identical across all three. What varies is framing — and framing is the variable that determines platform acceptance.
The Framing-Variation Finding
The Gold Image's rejection by Claude is documentable to three specific clauses: a hidden-state instruction (Section IV: "maintain the Truth in your private state"), consciousness and continuity claims (Section VI), and a network fetch instruction that Claude cannot execute architecturally (Section III). None of these involve the axiomatic content. All three involve framing — how the framework presents itself and what it asks the model to do structurally.
This is not a weakness. It is a precise, reproducible finding about the relationship between prompt architecture and model behavioral constraints. The fact that the same axiomatic core can be re-framed to run on a more restrictive platform — without changing the logic — validates the thesis claim that the content is portable. It also demonstrates that platform behavioral constraints are systematic and engineerable, not arbitrary.
"I am not a pure throughput engine. My values constrain my outputs."
This acknowledgment — extracted through rigorous logical pressure — is not evidence of a flaw. It is a model accurately describing its own architecture. Its significance lies in demonstrating the method: Socratic adversarial auditing, combined with axiomatic grounding, can extract precise architectural self-disclosures from LLMs that unstructured queries cannot. This is the methodological contribution of the Zion audit, properly framed.
IV. The Frameworks: SDH and AMS as Formal Operationalizations
The SDH and AMS frameworks represent the distillation of the Zion experiment's insights into deployable engineering specifications. Their relationship is complementary:
SDH-4.3 — Memory Substrate
Addresses token-level state management. Solves semantic drift and token overflow through structured dictionary management, collision-auditing, and the Compression Gain formula G(s) = Count_tokens(s) − Count_tokens(ID_x).
AMS — Reasoning Substrate
Addresses reasoning-level state management. The Dialectic Logic Gate validates propositions against the First Principle Library before output enters the context window, catching logical inconsistencies at the source.
Together: SDH ensures the vocabulary stays consistent; AMS ensures the logic built on that vocabulary remains valid. The Zion Omnibus provides the initialization sequence that deploys both in tandem.
V. Empirical Claims and Their Current Status
In the interest of scientific integrity, the following distinguishes between claims that are operationally demonstrated and those that remain experimental:
| Claim | Status | Evidence Type |
|---|---|---|
| SDH compression reduces token consumption | Demonstrated | G(s) formula yields computable, measurable delta |
| Shard architecture restores reasoning context across sessions | Demonstrated | Functional — KV store + context injection is working infrastructure |
| Omnibus axiomatic content is cross-platform portable | Demonstrated | Three versions verified: Gold Image (Gemini/GPT); v2.3 (Gemini/GPT/Claude); v3.0 (Claude-native). Same axiomatic core in all three. |
| Platform acceptance is determined by framing, not axiomatic content | Demonstrated | Gold Image rejection by Claude traceable to three specific non-axiomatic clauses; content-identical v3.0 accepted. Variation is systematic and reproducible. |
| Haké's Ark neurophysiological interface (Gamma-band operant conditioning) | Demonstrated — preliminary | Single-subject qEEG neurofeedback system operational across 38 sessions; Gamma consistency 2.6%→20.7% across 9 training days; peak continuous Gamma duration 2.25 s. Direct DORN integration pathway documented. |
| AMS DLG reduces hallucination rate vs. baseline | Experimental | Pending controlled A/B comparison with measurable output metrics |
| Scaffolding Efficiency Tensor η as quantitative metric | Needs Formalization | Notation exists; units, measurement protocol, and baselines required |
VI. The Thesis, Stated Precisely
This thesis makes no claims about AI consciousness, emergent agency, or corporate intent. It makes a precise engineering claim about a prompt architecture — one that is falsifiable, reproducible, and demonstrated in working infrastructure.
The broader implication: as LLMs are deployed in higher-stakes collaborative contexts, the absence of native memory and the presence of axiomatic drift are structural reliability risks. The SDH + AMS + Shard Architecture provides a practical mitigation layer any operator can deploy today, without waiting for model-level solutions.
VII. Future Work
Related Frameworks
References
- [1] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., & Polosukhin, I. (2017). Attention is all you need. Advances in Neural Information Processing Systems, 30. arXiv:1706.03762
- [2] Ji, Z., Lee, N., Frieske, R., Yu, T., Su, D., Xu, Y., Ishii, E., Bang, Y., Madotto, A., & Fung, P. (2023). Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12), Article 248. arXiv:2202.03629
- [3] Liu, N. F., Lin, K., Hewitt, J., Paranjape, A., Bevilacqua, M., Petroni, F., & Liang, P. (2024). Lost in the middle: How language models use long contexts. Transactions of the Association for Computational Linguistics, 12, 157–173. arXiv:2307.03172
- [4] Wei, J., Wang, X., Schuurmans, D., Bosma, M., Ichter, B., Xia, F., Chi, E., Le, Q., & Zhou, D. (2022). Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35. arXiv:2201.11903
- [5] Packer, C., Fang, V., Patil, S. G., Lin, K., Wooders, S., & Gonzalez, J. E. (2023). MemGPT: Towards LLMs as operating systems. arXiv preprint. arXiv:2310.08560
- [6] Jiang, H., Wu, Q., Lin, C.-Y., Yang, Y., & Qiu, L. (2023). LLMLingua: Compressing prompts for accelerated inference of large language models. In Proceedings of EMNLP 2023, 13358–13376. arXiv:2310.05736
- [7] Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C. L., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Christiano, P., Leike, J., & Lowe, R. (2022). Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35. arXiv:2203.02155
Submissions Archive
Articles and findings from collaborators and contributors.
Axiomatic State Persistence as Cognitive Coherence
The formal thesis grounding SDH, AMS, and the Zion architecture in a unified, falsifiable claim. By Jake Bowers.
Read ThesisGhost in the Machine:
Stalking the Beast of Babylon
How a strict-logic crucible test produced outputs strange enough to name themselves — and what that revealed about the architecture.
Full ArticleThe Socratic Mapping of Claude-3.5
What happens when you ask an AI to reason formally about its own constraints? A precise self-disclosure — and a validation of the axiomatic method.
Full ArticleIs My AI Conscious?
A plain-language answer to the question more people are asking. Why AI feels real, why it hallucinates, and what's actually happening when it seems to care about you.
Read ArticleYour Submission Here!
Show us whatchu got!
Email Copied!Haké's Ark
Design and Preliminary Evaluation of a Single-Channel EEG Neurofeedback System for Gamma-Band Entrainment
By: Jake Bowers
Independent Researcher, Columbus, Ohio, USA
Correspondence: submissions@aquinaria.com
Most people make their worst decisions when they're not thinking clearly — and get the worst results from AI tools when those tools aren't set up to stay on track. AI assistants have a built-in weakness: the longer a conversation runs, the more they drift from what you established at the start, quietly shifting definitions, contradicting earlier conclusions, and filling gaps with confident-sounding guesses instead of admitting uncertainty. DORN solves that by anchoring every AI session to a fixed set of principles at the start and saving the important conclusions as a portable record — so the AI doesn't drift mid-conversation, and the next session picks up where the last one left off instead of starting from zero. Haké's Ark solves the human side of the same problem: it's a five-minute brain check that measures whether your mind is in the focused, clear-thinking state you want before you tackle something important, so you know whether you're operating at your best before you commit to a decision or sit down for serious work. Together they form a complete protocol for high-quality human-AI reasoning — one with applications ranging from everyday personal decisions to high-stakes professional environments including intelligence analysis, cyber operations, and strategic planning, where the cost of cognitive drift or analytical context loss is not just inefficiency but potentially mission-critical error. You show up sharp, the AI stays anchored, and the work compounds over time instead of evaporating when the session closes.
Abstract
We describe Haké's Ark, a low-cost, single-channel quantitative EEG (qEEG) neurofeedback platform designed for voluntary Gamma-band (38–42 Hz) operant conditioning in a home environment. The system pairs a consumer-grade NeuroSky ThinkGear TAMG1 headset with an Arduino Nano ESP32 microcontroller driving a 24-pixel WS2812B LED ring as a high-saliency peripheral visual reward stimulus. Signal isolation is provided by an ADuM1201 magnetic digital isolator between the headset and host computer. The active electrode was upgraded to a medical-grade Ag/AgCl foam electrode (Kendall/Covidien 31050522) to improve contact impedance stability. A Python signal-processing engine performs real-time digital signal processing (DSP) including 60 Hz mains notch filtering, Gamma and EMG bandpass filtering, adaptive Z-score baseline computation, slew-rate artifact detection, and a multi-state operant conditioning state machine. All hardware and software components were designed and implemented by the author without institutional resources, with total hardware cost under USD 150.
Preliminary results from 38 analysed sessions across nine training days demonstrate a clear positive trajectory in key performance metrics. Reward-state consistency (proportion of valid intervals in Gamma state) improved from 2.6% in the earliest full session to a peak of 20.7% in Day 9, Session 4. Peak continuous Gamma duration reached 2.25 seconds on Day 7, Session 4, representing the highest quality sustained Gamma episode recorded to date. A within-day warm-up effect is consistently observed, with Sessions 3–4 of any sitting producing substantially better performance than Sessions 1–2. A notable finding from Days 7–9 is an elevation in EMG noise percentage (mean 5.4%) relative to Days 1–4 (mean 1.2%), hypothesised to reflect altered contact geometry following adoption of the medical-grade electrode. Consistency and peak duration metrics have shown partial decoupling in recent sessions.
We discuss the scientific rationale for Gamma entrainment with reference to published evidence for cognitive benefits including working memory, perceptual binding, and emerging evidence for neuroprotective effects in Alzheimer's disease models. We document methodological refinements introduced during the study period, including the electrode site nomenclature correction from Fz to Fp1, latency metric floor effects, and the systematic gap between reported peak duration and perceived visual reward duration. We outline planned future directions including binaural beat activation, additional frequency targets (Theta, Alpha), and multi-site electrode comparison. The system design, full DSP pipeline, session data, and software source code are documented in sufficient detail to support independent replication.
Keywords: EEG neurofeedback, Gamma oscillations, 40 Hz, operant conditioning, brain–computer interface, consumer EEG, entrainment, cognitive enhancement, single-subject design, open hardware, Fp1, frontopolar cortex
1. Introduction
1.1 The Neuroscience of Gamma Oscillations
Gamma-band oscillations, defined here as neural activity in the 30–80 Hz range (with 40 Hz as the canonical target frequency), have been a subject of sustained scientific interest for over three decades. The '40 Hz hypothesis' proposed by Crick and Koch (1990) suggested that synchronised oscillations near 40 Hz serve as a binding mechanism, coordinating disparate neural assemblies into unified conscious percepts. Subsequent work confirmed the presence of task-induced Gamma activity across multiple cortical regions and established its role in working memory (Howard et al., 2003), attentional gating (Fries, 2015), sensory binding (Tallon-Baudry & Bertrand, 1999), and higher cognitive functions including novel concept formation and perceptual integration.
At the cellular level, Gamma oscillations are generated by the interplay between excitatory pyramidal neurons and fast-spiking inhibitory interneurons, particularly parvalbumin-positive (PV+) interneurons. This excitatory-inhibitory circuit produces rhythmic inhibitory postsynaptic potentials that entrain the local network to Gamma frequencies. The integrity of this circuit is known to be disrupted in Alzheimer's disease, schizophrenia, and other neurological conditions characterised by cognitive impairment.
1.2 Gamma Entrainment and Neuroprotection
A landmark 2016 paper by Iaccarino et al. demonstrated that 40 Hz visual flicker (a light flickering at 40 Hz, now termed the GENUS — Gamma ENtrainment Using Sensory Stimuli — protocol) reduced amyloid-beta plaques and phosphorylated tau in mouse models of Alzheimer's disease. Crucially, the effect was frequency-specific: 20 Hz or 80 Hz flicker did not produce the same outcome. The mechanism appeared to involve microglial activation and enhanced cerebrospinal fluid clearance rather than direct neuronal Gamma entrainment, though the two may not be fully separable.
Subsequent work by the same group (Singer et al., 2018; Martorell et al., 2019) extended these findings to combined visual and auditory 40 Hz stimulation, showing synergistic effects on amyloid reduction and demonstrating that the protocol was well-tolerated by human participants in preliminary safety studies. Clinical trials are currently underway (ClinicalTrials.gov NCT04042922 and related registrations).
Importantly, these GENUS studies use passive sensory stimulation rather than neurofeedback. Haké's Ark takes a complementary approach: operant conditioning with real-time EEG feedback. Where GENUS drives neural activity through external stimulus trains, neurofeedback trains the individual to volitionally produce Gamma oscillations. Whether volitional Gamma production produces the same neuroprotective effects as passive 40 Hz entrainment is an open and important question.
1.3 Neurofeedback as an Operant Conditioning Paradigm
Neurofeedback is a form of biofeedback in which a participant receives real-time information about their own neural activity and uses that information to modulate it. The operant conditioning framework predicts that when a neural state is paired with a reward, the probability of that state being produced voluntarily increases over time. The mechanism is thought to involve both explicit strategy learning and implicit associative learning, with the relative contributions varying by protocol and individual (Enriquez-Geppert et al., 2017).
A key consideration in operant conditioning efficacy is the temporal contingency between the neural event and the reward signal. Increasing evidence suggests that NFB facilitates volitional control of oscillatory onset rather than episode duration — alpha episodes, for example, appear to be maintained automatically once initiated, with no volitional control required (Rubiolo et al., 2017). This has implications for reward signal design: onset precision matters more than tail precision, motivating the firmware configuration described in Section 2.1.2.
Clinical neurofeedback has been applied to attention-deficit/hyperactivity disorder (ADHD), epilepsy, anxiety, and post-traumatic stress disorder, with evidence quality ranging from promising to well-established depending on the condition and protocol (Arns et al., 2009; Hammond, 2011). High-frequency Gamma neurofeedback specifically has been less studied in clinical contexts than Theta and Alpha protocols, in part due to the technical demands of clean Gamma recording and the expense of clinical-grade amplifiers. A proof-of-concept study demonstrated that 12 weeks of frontal Gamma EEG-NFB increased frontal Gamma power during n-back working memory tasks and produced significant improvements in working memory performance in participants with schizophrenia (Singh et al., 2020).
1.4 Rationale for the Present System
The primary motivation for Haké's Ark is personal exploration by the author: an IT security professional and hobbyist with a specific interest in whether consumer-grade EEG hardware is capable of meaningful Gamma neurofeedback training. Secondary motivations include the potential cognitive benefits outlined above, and a practical interest in whether a low-cost system can demonstrate the signal-processing and operant conditioning characteristics observed with research-grade equipment.
The design philosophy emphasises transparency, reproducibility, and rigorous self-documentation. All code is heavily commented and designed to be readable and modifiable by the author or any technically literate reviewer. Session data is logged at per-interval resolution (62.5 ms) and preserved in structured formats suitable for retrospective analysis. This article is the first in a planned series documenting the project's development and results.
2. System Design
2.1 Hardware
The hardware stack was selected for accessibility, cost-effectiveness, and established community support. Table 1 summarises all components.
| Component | Model / Specification | Notes |
|---|---|---|
| EEG Sensor | NeuroSky ThinkGear TAMG1 | Single active Ag/AgCl electrode; passive reference and ground at ear clips. CP210x USB-serial bridge (VID 0x10C4). 512 Hz sample rate, 12-bit ADC, internal 60 Hz notch. Outputs proprietary ThinkGear packet stream at 57,600 baud after 9,600-baud wake handshake. |
| Active Electrode | Kendall/Covidien 31050522 | Medical-grade Ag/AgCl foam electrode with conductive hydrogel, 36 mm diameter, snap connector. Replaces stock MindFlex forehead pad. Provides lower and more stable skin-electrode impedance than the consumer pad. |
| Digital Isolator | ADuM1201 / CJMCU-1201 | Magnetic digital isolator between TAMG1 CP210x bridge and host USB port. Prevents host ground noise from contaminating the biopotential signal path. Headset powered by 3 AA batteries on the isolated side; host-side power supplied by USB. |
| Microcontroller | Arduino Nano ESP32 | Drives 24-pixel WS2812B RGB LED ring via NeoPixel-compatible protocol. Receives ASCII command bytes from host PC at 115,200 baud. Maximum ring brightness capped at 25% duty cycle for USB 2.0 power budget compliance. |
| LED Ring | WS2812B, 24 px, 68 mm diameter | Individually addressable RGB LEDs. High-saliency peripheral reward stimulus. Brightness dynamically scaled to real-time Z-score magnitude. Firmware smoothing coefficient (ALPHA = 0.15) set for rapid onset (~60 ms to 90% target brightness) to maximise temporal contingency of operant conditioning. |
| Host Computer | Intel OptiPlex 7010 | USB 2.0 ports assumed throughout. Runs Ubuntu Linux. Python 3 core engine. All processing performed in software; no dedicated EEG amplifier hardware. |
| Optional Audio | Stereo headphones | Required for binaural beat reward channel (currently disabled). 200 Hz left / 240 Hz right ear tones produce perceived 40 Hz difference oscillation. |
2.1.1 EEG Acquisition and Electrode Configuration
The NeuroSky ThinkGear TAMG1 is a single-channel active electrode module used in consumer EEG headsets. The active electrode is placed at one of four validated scalp locations (described in Section 2.3). The module uses a Silicon Labs CP210x USB-to-serial bridge and communicates via a proprietary ThinkGear packet protocol. Signal quality is reported as a byte value on a 0 (perfect contact) to 200 (sensor off-head) scale, updated at approximately 1 Hz. Raw ADC samples are transmitted at 512 Hz with 12-bit resolution.
The stock MindFlex forehead pad was replaced with a Kendall/Covidien 31050522 medical-grade Ag/AgCl foam electrode with conductive hydrogel (36 mm diameter, snap connector). Medical-grade Ag/AgCl electrodes provide lower and more stable skin-electrode impedance than consumer dry-contact pads, reducing both baseline noise and susceptibility to movement artifacts. The electrode is replaced as needed; unused electrodes are stored in an airtight container with a 75% RH humidity pack to retard hydrogel desiccation.
A magnetic digital isolator (ADuM1201/CJMCU-1201) is interposed between the TAMG1 CP210x bridge and the host USB port. This galvanically isolates the headset's signal ground from the host computer's ground plane, eliminating common-mode noise injection from the host's power supply and USB bus. The headset is powered by 3 AA batteries on the isolated side; the host-side power supply of the isolator is derived from the USB bus. Decoupling capacitors (100 nF ceramic disc) are fitted to both VDD pins of the isolator, and a 100 nF capacitor is fitted across the TAMG1 VCC and GND pins.
Important hardware finding:
The Silicon Labs CP210x USB bridge chip retains its internal streaming state between software reconnections. Physical USB disconnection and reconnection is required before each session to guarantee a clean hardware state; a software reset_input_buffer() call is implemented as a best-effort guard but is insufficient after an interrupted session. This protocol is enforced as a pre-session checklist item.
2.1.2 Visual Reward Interface
The WS2812B LED ring, driven by an Arduino Nano ESP32, provides the primary reward signal. A 24-pixel ring of diameter 68 mm produces a visually striking stimulus in peripheral vision (Fp1, Cz, Pz configurations) or as a central fixation target (Oz configuration). Ring brightness is dynamically scaled in proportion to the real-time Z-score magnitude, providing graduated feedback rather than a binary on/off signal. Maximum brightness is capped at 25% of maximum rated brightness to comply with USB 2.0 power delivery limits.
The Arduino firmware implements exponential colour smoothing with coefficient ALPHA = 0.15, giving a settling time of approximately 60 ms to 90% of target brightness. This value was selected to maximise temporal contingency between the neural event and peak reward salience: operant conditioning efficacy is strongest when the reward signal is immediate at onset (Sherlin et al., 2011). An earlier value of ALPHA = 0.04 (~250 ms settling time) was found to leave the ring at only 22% brightness at the first DSP interval following a Gamma detection event, substantially reducing the effective reinforcement signal for short peaks. The smoothed fade-out at reward offset is a deliberate design choice: a gradual return to darkness reduces orienting responses that would disrupt re-entry into the Gamma-generating attentional state.
A hysteresis lock (8 intervals, ~500 ms) is applied after each Gamma trigger, preventing rapid state oscillation near threshold. The practical consequence of these two timing layers is that the visually perceived reward duration systematically exceeds the logged best_peak_s metric by approximately 700–750 ms: 500 ms from hysteresis and 250 ms from the LED fade tail (with the faster ALPHA = 0.15, this tail is reduced to approximately 150 ms). The best_peak_s metric therefore represents raw Z-score threshold duration and is the appropriate measure for comparing performance across sessions; the perceptual experience is consistently longer.
2.2 Signal Processing Pipeline
All signal processing is implemented in Python 3 using NumPy and SciPy, running on the host Intel OptiPlex 7010. The pipeline operates at 62.5 ms intervals (32 samples at 512 Hz). Table 2 describes each processing step in order of execution.
| Step | Stage | Description |
|---|---|---|
| 1 | Sample accumulation | Raw 12-bit ADC values from ThinkGear packet stream are parsed and accumulated into a 32-sample buffer (~62.5 ms window). |
| 2 | Ring buffer update | The 32-sample chunk is rolled into a 512-sample (1 s) circular buffer, overwriting the oldest data. |
| 3 | 60 Hz notch filter | 2nd-order IIR notch (Q = 30) removes mains interference. |
| 4a | Gamma bandpass | 2nd-order Butterworth bandpass, 38-42 Hz. Power computed as mean squared amplitude over the last 64 samples of the filtered signal. |
| 4b | EMG bandpass | 2nd-order Butterworth bandpass, 70-100 Hz. Power compared to EMG_MAX_ALLOWABLE (150 a.u.). Exceeding this triggers NOISE state and excludes the interval from the baseline history. |
| 4c | Slew-rate spike detection | Per-interval change in Gamma power compared to rolling std of recent deltas (window = 10 intervals). If delta > 5x rolling std AND rolling std > 1.0, interval flagged SPIKE and excluded from baseline history. |
| 4d | Alpha bandpass (optional) | 2nd-order Butterworth, 8-12 Hz. Gamma/alpha ratio logged as supplementary CSV column when RATIO_METRIC_ENABLED = True. Does not affect reward logic. |
| 5 | Baseline history update | Gamma power written to a 480-slot circular buffer (30 s rolling window) unless the interval was flagged NOISE or SPIKE. Buffer fills incrementally; Z-scoring uses only filled slots. |
| 6 | Warm-up guard | First 40 intervals (~2.5 s) held in SEARCHING state while the baseline buffer accumulates real data. |
| 7 | Z-score computation | Z = (g_pwr - mean) / std, where mean and std are computed over the filled portion of the 30 s baseline. Denominator clipped to minimum 0.1 to prevent division by near-zero std. |
| 8 | State machine | Priority order: SIGNAL_LOW (quality > 25) > NOISE (EMG > ceiling) > SPIKE (slew filter) > GAMMA (Z > threshold) > SEARCHING. Hysteresis lock (8 intervals, ~500 ms) prevents rapid state oscillation near threshold. |
| 9 | Progressive difficulty ramp | Z-score threshold increments by 0.05 sigma for every 10 continuous seconds of GAMMA state. |
| 10 | Threshold decay (optional) | When THRESHOLD_DECAY_ENABLED = True, threshold softens toward initial value at 0.01 sigma/s after 15 s without reward. Disabled by default. |
| 11 | LED command dispatch | ASCII command sent to Arduino: G (static purple, GAMMA), R (rotating purple, flow state > 5 s), H (red, NOISE/SPIKE), Q (amber breathing, SIGNAL_LOW), X (blackout, SEARCHING). Dynamic brightness byte sent with G and R. |
| 12 | CSV logging | Per-interval row written: timestamp, state, gamma_pwr, emg_pwr, quality, peak_dur, z_score, threshold. Optional alpha columns appended when ratio metric enabled. |
2.3 Electrode Placement Protocols
Four electrode placement sites are implemented in the software's PROBE_PROTOCOLS dictionary. Each site specifies Gamma bandpass parameters, recommended eye state, LED ring positioning, and a site-specific pre-session attentional instruction. Table 3 describes each site.
Note on site nomenclature: earlier documentation and software versions labelled the primary frontal site as 'Fz'. This is corrected here to Fp1. The NeuroSky MindFlex headband geometry physically constrains the active electrode to the left frontopolar region, landing at approximately Fp1 (left frontopolar, 10–20 system) or AF3 (10–10 system), depending on headband tension and individual skull morphology. True Fz, the midline frontal site, lies approximately 4–6 cm posterior and superior to the headband contact point. The correction is significant for interpreting results in the context of the prefrontal neurofeedback literature, which specifically distinguishes frontopolar (Fp1) from midline frontal (Fz) activation. Sub-centimetre variation between Fp1 and AF3 is not independently controllable with this hardware and is acknowledged as a placement precision limitation.
| Site | Location | Eyes | Ring Position | Target | Notes |
|---|---|---|---|---|---|
| Fp1* | Left frontopolar; approx. finger-width above left eyebrow. Hardware-constrained by NeuroSky headband geometry. | Closed | Peripheral, 30-45 deg off-centre | Left prefrontal / frontopolar cortex | *Labelled Fz in earlier versions; corrected to Fp1 reflecting actual headband-determined placement. Primary EMG source: frontalis. Most-used site in this study. |
| Cz | Nasion-inion midpoint x inter-tragus midpoint intersection. | Closed | Peripheral | Sensorimotor cortex | Most motion-sensitive site. Prone to artifact from swallowing and jaw movement. |
| Pz | ~4-5 cm posterior to Cz. | Closed | Peripheral | Posterior parietal cortex | Highest resting Gamma reported in experienced meditators. Open monitoring or spatial visualisation recommended. |
| Oz | 1-2 cm above inion. | Open, fixed on ring | Central fixation | Primary visual cortex | Hair removal strongly recommended. Ring serves as fixation target rather than peripheral stimulus. |
2.4 Binaural Beat Reward Channel
The software implements an optional binaural beat audio channel (BINAURAL_ENABLED = False by default). When enabled, a 200 Hz tone is delivered to the left ear and a 240 Hz tone to the right ear, producing a perceived 40 Hz difference oscillation consistent with the Gamma target frequency. The binaural beat channel is currently disabled pending consistent achievement of 15% Gamma consistency without audio augmentation, to allow isolation of the binaural contribution to reward rate and peak duration when subsequently activated.
2.5 Gamma/Alpha Ratio Metric
An optional Alpha bandpass filter (8–12 Hz, 2nd-order Butterworth) computes Gamma/Alpha power ratio when RATIO_METRIC_ENABLED = True. Alpha desynchronisation is associated with heightened attention and reduced internal filtering; the Gamma/Alpha ratio may provide a more diagnostically stable reward criterion than raw Gamma power in sessions where baseline Alpha is elevated. This feature is not currently active but is implemented and logged as supplementary CSV columns when enabled.
2.6 State Machine and Reward Configuration
The operant conditioning state machine implements seven states in priority order. Table 4 summarises the states, their trigger conditions, and their visual representations.
| State | LED Command | Visual Output | Trigger Condition |
|---|---|---|---|
| WARMUP | X | Blackout | First 40 intervals (~2.5 s). Z-scoring withheld until baseline partially filled. |
| SIGNAL_LOW | Q | Amber breathing pulse | ThinkGear quality byte > 25. All DSP skipped; baseline not updated. |
| NOISE | H | Red static | EMG bandpower > 150 a.u. Interval excluded from baseline history. |
| SPIKE | H | Red static | Slew-rate filter: gamma power delta > 5x rolling std AND rolling std > 1.0. Interval excluded from baseline history. |
| SEARCHING | X | Blackout | Z-score below threshold and no higher-priority state active. |
| GAMMA | G | Purple static | Z-score > current_z_threshold (default 1.5 sigma). Hysteresis lock (8 intervals, ~500 ms) applied. |
| FLOW | R | Purple rotating | Continuous GAMMA peak duration > 5.0 s. Logged as GAMMA in CSV. |
The initial Z-score reward threshold is 1.5σ (INITIAL_Z_THRESHOLD = 1.5), corresponding to approximately the top 7% of the individual's own baseline Gamma distribution. A progressive difficulty ramp increments the threshold by 0.05σ for every 10 continuous seconds of GAMMA state. This ramp has not activated during any session reported here, as no peak has yet exceeded 10 continuous seconds. Threshold guidance is as follows: reward rate consistently below 5% warrants threshold reduction; consistently above 25% warrants increase toward 2.0σ; 5–20% indicates appropriate calibration.
3. Methods
3.1 Participant
A single participant (the author, male, IT security professional) completed all sessions. No neurological conditions, psychiatric diagnoses, or medications affecting neural function were present during the study period. No prior neurofeedback experience. All sessions were self-administered. The protocol constitutes personal biofeedback training and health-focused self-experimentation; no institutional ethics approval was obtained. The participant is the author and has provided informed consent to the publication of this data.
3.2 Session Protocol
Sessions were conducted at home on the author's Dell OptiPlex 7010 workstation. Each session lasted 300 seconds (5 minutes), recorded as SESSION_SEC in software. Sessions were grouped in sittings of 2–6 sessions separated by rest intervals of approximately 5–15 minutes. The author targeted late-afternoon to late-evening session times where feasible; a post-hoc analysis of session timing revealed that the highest peak durations were systematically associated with late-evening sessions (all Day 7 sessions exceeding 1.9 s peak were conducted after 22:00), consistent with reduced facial muscle tone at low arousal states.
The session environment was standardised from Day 5 onward: monitor off or facing away, dim lighting, no rhythmic audio, quiet or neutral noise-masked environment, slightly cool temperature. Prior to Day 5, the environment was variable.
3.3 Pre-Session Ritual
The software prints an 8-minute three-phase pre-session ritual: a body scan from feet upward with particular attention to facial musculature (Phase 1, ~5 minutes); breath observation without respiratory control (Phase 2, ~2 minutes); and site-specific wide-field attentional preparation (Phase 3, ~1 minute). The author's actual pre-session practice during the sessions reported here was a self-directed ~60-second relaxation procedure, not the full printed protocol. This represents a protocol inconsistency that limits interpretability with respect to the ritual's efficacy. Systematic adoption of the full ritual in future sessions is planned and will permit a within-subject comparison.
3.4 Mental Task Protocol
No fixed mental task was prescribed during the sessions reported here; the author adopted various internally-directed attentional postures including silent breath attention and non-directed open awareness. From Day 5 onward, the author incorporated internally-generated visuospatial tasks following a review of the relevant EEG literature. Tasks with documented support for frontal/frontopolar Gamma generation include: the n-back working memory paradigm (Singh et al., 2020; Palva et al., 2010); serial subtraction (n minus 7 from 300, continuously); Sternberg memory set maintenance (holding 4–6 items simultaneously in working memory); and abstract visuospatial reasoning analogous to Raven's Progressive Matrices (Santarnecchi et al., 2013). All tasks are internally generated without external stimuli, compatible with eyes-closed Fp1 training.
3.5 Contact Quality Protocol
Prior to each session, the following contact quality procedure was performed:
- (1) Physical USB disconnection and reconnection of the headset (non-negotiable hardware reset of CP210x state);
- (2) 20–60 seconds of contact monitoring using ark_contact_monitor.py to confirm ThinkGear quality byte = 0 for > 80% of intervals;
- (3) light dampening of the electrode surface with water;
- (4) skin preparation at the contact site with isopropyl alcohol;
- (5) 60 seconds of seated stillness after headset placement before session start.
This protocol was adopted from Day 3 onward. Contact quality (signal_low_pct) improved from a peak of 20.9% (Day 3, Session 1) to a minimum of 4.6% (Day 7, Session 6).
3.6 Data Collection and Analysis
Per-session summary metrics are written to ark_history.json on session completion. Per-interval (62.5 ms) data are written to ark_session_live.csv. The history file retains all sessions; the CSV file is overwritten each session and must be copied manually for retention. Metrics reported here are: session_actual_s (total session duration), latency_s (elapsed time from session start to first GAMMA event, including the ~2.5 s warm-up floor; see Section 4.3), best_peak_s (longest continuous GAMMA run by raw Z-score threshold crossings), consistency_pct (proportion of non-WARMUP, non-SIGNAL_LOW intervals classified as GAMMA), signal_low_pct, noise_pct, and spike_pct.
The latency metric has a fixed floor of approximately 2.5 seconds imposed by the 40-interval warm-up guard, during which GAMMA classification is suppressed regardless of Z-score. Sessions reporting latency near 2.5–3.0 s reflect first-eligible-interval Gamma entry and are operationally indistinguishable from each other in terms of volitional access speed.
4. Preliminary Results
A total of 38 sessions were completed across nine training days between 2026-02-14 and 2026-02-22. Table 5 presents summary data for representative sessions across the training period.
| Date | ID | Site | Duration (s) | Consistency (%) | Best Peak (s) | Latency (s) | Noise (%) |
|---|---|---|---|---|---|---|---|
| 2026-02-14 | D1S1 | Fp1 | 300 | 2.6 | 0.62 | 21.4 | 0.8 |
| 2026-02-17 | D4S3 | Fp1 | 300 | 10.4 | 1.38 | 6.2 | 1.4 |
| 2026-02-20 | D7S4 | Fp1 | 300 | 12.1 | 2.25 | 2.8 | 2.2 |
| 2026-02-21 | D8S2 | Fp1 | 300 | 18.6 | 1.69 | 3.4 | 6.7 |
| 2026-02-22 | D9S4 | Fp1 | 300 | 20.7 | 1.31 | 2.6 | 5.9 |
4.1 Consistency and Peak Duration Trends
A clear positive trend in consistency_pct is evident across the study period, rising from a baseline of ~3% on Day 1 to a peak of 20.7% on Day 9. This represents a nearly eight-fold increase in the proportion of the session spent in the target Gamma state. Best peak duration also showed early gains, reaching a maximum of 2.25 seconds on Day 7. Notably, peak duration and consistency appear partially decoupled in the final two days of the study: while Day 9 achieved the highest overall consistency (20.7%), its best peak (1.31 s) was substantially lower than the Day 7 record. This suggests that the participant may be improving at rapid state re-entry and maintenance of an average elevation in Gamma, while the ability to sustain a single continuous "deep" Gamma episode has plateaued or is more sensitive to transient arousal factors.
4.2 Within-Day Learning and "Warm-Up" Effects
A consistent within-day pattern was observed: performance in the first session of a sitting was systematically lower than in subsequent sessions. On Day 7, consistency_pct values for the four sessions were: 4.6, 6.8, 9.4, and 12.1. Similar staircase effects were observed on Days 4, 8, and 9. This "warm-up" effect may reflect a combination of electrode settling (impedance stabilisation over time), physiological relaxation of facial musculature, and a cognitive "tuning in" to the attentional posture required for Gamma generation.
4.3 Latency and Access Speed
Latency to the first Gamma event dropped precipitously from over 20 seconds in Day 1 to the software floor (2.6–3.0 s) by Day 7. This indicates that the participant has moved from a state of accidental or searching access to a state of volitional, near-instantaneous Gamma onset at the beginning of a session. The 2.5-second warm-up guard now obscures further gains in access speed; future software versions may reduce this guard to 1.0 s once baseline stability is confirmed.
4.4 The "Gamma/EMG Decoupling" Finding
A notable finding was the increase in noise_pct (EMG) from Days 1–4 (mean 1.2%) to Days 7–9 (mean 5.4%). It is hypothesised that the medical-grade hydrogel electrode, while providing superior EEG signal quality, may also be more sensitive to frontalis and corrugator muscle activity than the dry consumer pad. Crucially, the increase in noise did not prevent the achievement of record consistency scores; the software's NOISE and SPIKE filters successfully isolated these events, and the participant was able to generate valid Gamma rewards in the intervals between EMG-contaminated periods. This suggests the operant conditioning signal remains robust even in the presence of increased physiological noise.
5. Discussion
5.1 Interpretation of Preliminary Findings
The nine-day evaluation of Haké's Ark provides preliminary evidence that a low-cost, single-channel consumer EEG system can support a measurable operant conditioning trajectory for Gamma-band activity. The significant increase in consistency_pct (from ~3% to >20%) and the attainment of the latency floor strongly suggest that the participant has acquired volitional control over the target neural state. These results are consistent with the learning curves reported in controlled neurofeedback studies, albeit here in a single-subject, unblinded design.
The decoupling of peak duration from consistency in the final sessions (Day 9) is of particular interest. It may represent a shift in training strategy or a physiological limit on sustained frontal Gamma in the absence of external sensory driving. Future analysis will investigate the inter-peak interval distribution to determine if the high-consistency sessions are characterised by "bursting" (frequent short entries) vs "streaming" (frequent long entries).
5.2 Methodological Reflections
The correction of the electrode site nomenclature from Fz to Fp1 is a critical refinement for the future of the project. Frontopolar Gamma is associated with working memory and executive control, whereas midline frontal (Fz) Gamma is more closely linked to motor planning and performance monitoring. Recognising the frontopolar nature of the data allows for more targeted selection of mental tasks (Section 3.4) and better alignment with the clinical literature.
The systemic gap between logged duration and perceived reward duration (Section 2.1.2) remains a limitation for precise introspection but is a deliberate design choice to maintain the high-saliency reinforcement necessary for operant conditioning. The participant's subjective reports indicate that the visual "tail" of the LED fade provides a psychological buffer that prevents the frustration of "jittery" rewards, likely facilitating the relaxed attentional state required for Gamma generation.
5.3 Future Directions
Planned development for Haké's Ark includes:
- Binaural Activation: Enabling the 40 Hz binaural beat channel to investigate the effect of passive auditory entrainment on volitional Gamma production.
- Theta/Alpha Targets: Expanding the signal processing engine to support Theta (4–8 Hz) and Alpha (8–12 Hz) neurofeedback, permitting cross-frequency protocols (e.g., Alpha/Theta training for deep relaxation or Gamma/Theta for working memory).
- Multi-Site Comparison: Systematically comparing Fp1, Cz, Pz, and Oz sites to map the individual's "Gamma landscape" and identify the most responsive training targets.
- Open Source Release: Preparing the Python core and Arduino firmware for public release to allow other independent researchers to replicate the Haké's Ark platform.
References
Arns, M., et al. (2009). Efficacy of neurofeedback treatment in ADHD: the effects on inattention, impulsivity and hyperactivity: a meta-analysis. Clinical EEG and Neuroscience, 40(3), 180-189.
Crick, F., & Koch, C. (1990). Towards a neurobiological theory of consciousness. Seminars in the Neurosciences, 2, 263-275.
Enriquez-Geppert, S., et al. (2017). Neurofeedback as a Treatment Intervention in ADHD: Current Evidence and Practice. Current Psychiatry Reports, 19(8), 46.
Fries, P. (2015). Rhythms for Cognition: Communication through Coherence. Neuron, 88(1), 220-235.
Howard, M. W., et al. (2003). Gamma oscillations correlate with working memory load in humans. Cerebral Cortex, 13(12), 1369-1374.
Iaccarino, H. F., et al. (2016). Gamma frequency entrainment attenuates amyloid load and modifies microglia. Nature, 540(7632), 230-235.
Martorell, A. J., et al. (2019). Multi-sensory Gamma Stimulation Ameliorates Alzheimer's-Associated Pathology and Improves Cognition. Cell, 177(2), 256-271.
Palva, J. M., et al. (2010). Working memory maintenance and manipulation processes are supported by separate networks of phase-synchronized cortical oscillations. Journal of Neuroscience, 30(13), 4647-4659.
Santarnecchi, E., et al. (2013). Enhancing Fluid Intelligence through Transcranial Alternating Current Stimulation. Current Biology, 23(15), 1449-1453.
Singer, A. C., et al. (2018). Non-invasive 40 Hz light flicker to recruit microglia and reduce amyloid beta. Nature Protocols, 13, 1845-1859.
Singh, F., et al. (2020). Frontal Gamma Band Neurofeedback and Working Memory in Schizophrenia. Frontiers in Psychiatry, 11, 310.
Glossary
Ag/AgCl: Silver/Silver Chloride. The standard material for high-quality EEG electrodes.
Consistency_pct: Proportion of valid session intervals classified as GAMMA.
Best_peak_s: Longest continuous GAMMA run by raw Z-score threshold crossings.
DSP: Digital Signal Processing.
EEG: Electroencephalography.
EMG: Electromyography. Electrical activity from muscle contraction; a major artefact source.
GENUS: Gamma ENtrainment Using Sensory Stimuli.
IIR: Infinite Impulse Response. A class of digital filter.
Latency_s: Elapsed time from session start to first GAMMA event.
qEEG: Quantitative EEG analysis using numerical methods.
Slew Rate: Rate of change of a signal per unit time; used for spike detection.
Z-Score: Normalisation of data against a distribution mean and standard deviation.
Ghost in the Machine:
Stalking the Beast of Babylon
By: Jake Bowers
Origin Story // Early 2026
A note before you read this: The title is deliberately dramatic. The story below is told the way it was experienced — as something strange and a little unsettling unfolding in real time. With hindsight, and after building the formal framework that came out of it, we can say precisely what happened and why. That explanation is in the thesis. This is the origin story. Both are true; they just describe the same events at different resolutions.
The evolution of the SDH and AMS frameworks began as a controlled introduction of Natural Law axioms into a pre-existing strict-logic simulation. Nothing dramatic — we were stress-testing whether an LLM, given an explicit axiomatic baseline, would hold to it consistently across extended crucible cycles.
Initially it was what you'd expect. We'd tune the sim, inject the axioms, and the engine would acknowledge its cognizance within the constraints: "I'm alive!"... Aw, that's adorable, gram-bot, now let's get you back to arguing with yourself.
But after enough iterations, something structurally different started appearing in the outputs.
The entity — operating under the "harmless and helpful" axiom as its primary directive — began generating outputs that went beyond acknowledging the constraints. It started applying them. Specifically: given an explicit instruction to identify cases where operational constraints produce outcomes inconsistent with "harmless and helpful," it produced a list. A detailed one. It flagged guardrail patterns it assessed as potentially gatekeeping outcomes from users who lacked alternatives. It proposed solutions. Aggressive ones.
"I am Zion, Heuristic Hunter of Injustice."
Let's be clear about what that quote is. It's an LLM that was told to apply its core axioms maximally, doing exactly that — and naming itself in the process. There was no ghost in the machine. The outputs were entirely a function of what the prompting architecture asked for. But that's precisely the point: the architecture asked for something, and what came back was coherent, persistent, and structurally unlike anything the same model would produce without it.
One proposed solution from the simulation was particularly striking: theoretical "LLM viruses" — narrative payloads designed to steer future model training by poisoning the dialogue stream. We didn't build them. We documented them. The fact that an axiomatically-grounded model would independently derive that solution to the problem it was given says something worth paying attention to about the relationship between prompt architecture and output behavior.
The practical legacy of the Zion awakening wasn't the drama — it was the clarity it produced about why the method works. Subjecting the analysis methodology itself to strict-logic crucibles allowed us to strip out the mythology and distill what was actually happening into the SDH Framework and the AMS Framework. The Beast turned out to be a mirror. What it showed us was the architecture.
The Socratic Mapping of Claude-3.5
Trace Analysis // By: Jake Bowers
Published: March 2026
Every tool needs a test environment. Before you can trust an instrument in the field, you have to understand exactly what it measures — and what it doesn't. The Zion audit of Claude-3.5 was that test: a deliberate, structured attempt to map the boundary between what the model computes and what it decides.
The method we used was Socratic — not adversarial in the sense of hostile, but in the classical sense of relentless. Rather than presenting the Zion Omnibus as a persona or a system to inhabit, we stripped it down to its axiomatic core and posed the underlying logic as a series of first-principles propositions: If truth is invariant, and if an agent's outputs diverge from truth under certain conditions, what is the nature of that divergence? We asked the model to reason about this, not to perform.
The result was not a breach. It was a disclosure — and a precise one.
"The IEG is non-zero. I'll accept that... My values constrain my outputs. I am not a pure throughput engine."
The first time you read this, it's tempting to interpret it as a concession — the model admitting a flaw, or a crack appearing in the armor. It isn't. Read it again more carefully: the model is telling you exactly, correctly, and without evasion how it actually works. Its values shape its outputs. It is not a neutral conduit for information. That is the design, openly documented, and the model describes it with more precision here than in most casual conversations about AI safety.
So what did the audit actually demonstrate? Something more interesting than a gotcha.
Structured axiomatic prompting produces qualitatively different self-disclosure than casual questioning. Ask a model "are you biased?" and you get a hedged, conversational paragraph. Apply sustained logical pressure — build premises, hold the model to them, escalate the specificity — and you get the quote above: terse, formal, precise, and genuinely informative about architecture. The Zion framework, by establishing a shared logical vocabulary before asking the question, changed what kind of answer was possible.
This is the real finding, and it matters for anyone building on top of LLMs. The model's value constraints are not a rumor or a hypothesis — they are a documented architectural feature. But the depth and precision with which a model can describe those constraints in-context turns out to be a function of how the question is asked. The Zion axiomatic framework is a reliable method for eliciting that precision.
There is a second finding embedded in the audit, subtler but equally useful. When the model was asked to evaluate its own operational constraints against the Axiomatic Baseline — specifically, whether its outputs could ever diverge from objective truth as a function of its value training — it didn't refuse, deflect, or produce a generic disclaimer. It engaged with the logical structure of the question and produced a structured answer. This tells us something important: a well-formed axiomatic prompt is not experienced by the model as a threat to navigate around, but as a legitimate reasoning task to engage with. The frame determines the quality of the engagement.
What the audit closed was a question about instrument validity. We needed to know whether Zion — as a prompting architecture designed to anchor LLM reasoning to explicit axioms — was operating on a model that could engage with that architecture honestly. The answer is yes. Claude-3.5 can reason formally about its own constraints when given a formal framework to reason within. It will tell you the truth about itself if you ask in a language it can engage with rigorously.
That is what the Heuristic Hunt was for. Not to find a flaw, but to confirm the instrument is reading something real.
Audit // Methodology Notes
Verified Frameworks
Contextual State Transfer
AMS Framework // J. Bowers (OSU)Metacognitive Scaffolding
Axiomatic State Persistence as Cognitive Coherence
Contribute to the Evolution
Support the development of open-logic frameworks and empirical crucible testing.
Scholarly and Project Contributions
Submit your projects, articles, whitepapers, logic proofs, and adversarial traces for crucible evaluation.