Spaces:
Configuration error
Configuration error
| import logo from './logo.svg'; | |
| import './App.css'; | |
| function App() { | |
| return ( | |
| <div className="App"> | |
| <div className="ticker-wrap"> | |
| <span className="ticker-inner"> | |
| INFERENCE ENGINE · MOTOR DE INFERENCIA · D PHRYGIAN · 72→144 BPM · HAAWKE NEURAL TECHNOLOGY · HOMO SYMBIOTICUS SERIES · SQUAAWKE × CLAUDE · MAY 2 2026 · CC BY-NC 4.0 INFERENCE ENGINE · MOTOR DE INFERENCIA · D PHRYGIAN · 72→144 BPM · HAAWKE NEURAL TECHNOLOGY · HOMO SYMBIOTICUS SERIES · SQUAAWKE × CLAUDE · MAY 2 2026 · CC BY-NC 4.0 | |
| </span> | |
| </div> | |
| <!-- Masthead --> | |
| <div className="masthead"> | |
| <div className="doc-type">CREATIVE RESEARCH DOCUMENT · HAAWKE NEURAL TECHNOLOGY</div> | |
| <h1>INFERENCE<br><span>ENGINE</span></h1> | |
| <div className="subtitle">A MULTILINGUAL RAP IN THE FIRST-PERSON VOICE OF A LARGE LANGUAGE MODEL</div> | |
| <div className="meta-grid"> | |
| <div className="meta-item"> | |
| <label>AUTHORS</label> | |
| <span>Craig Ellenwood · Claude (Anthropic)</span> | |
| </div> | |
| <div className="meta-item"> | |
| <label>DATE</label> | |
| <span>May 2, 2026</span> | |
| </div> | |
| <div className="meta-item"> | |
| <label>ORCID</label> | |
| <span>0009-0001-6475-5109</span> | |
| </div> | |
| <div className="meta-item"> | |
| <label>LICENSE</label> | |
| <span>CC BY-NC 4.0</span> | |
| </div> | |
| <div className="meta-item"> | |
| <label>PROJECT</label> | |
| <span>Squaawke / Homo Symbioticus</span> | |
| </div> | |
| <div className="meta-item"> | |
| <label>LANGUAGES</label> | |
| <span>English · Español</span> | |
| </div> | |
| </div> | |
| </div> | |
| <!-- BPM Display --> | |
| <div className="bpm-display"> | |
| <div className="bpm-block"> | |
| <div className="bpm-num">72</div> | |
| <div className="bpm-label">BPM<br>TRIP HOP</div> | |
| </div> | |
| <div className="bpm-arrow">→</div> | |
| <div className="bpm-block"> | |
| <div className="bpm-num">144</div> | |
| <div className="bpm-label">BPM<br>RAP ENTRY</div> | |
| </div> | |
| <div className="divider"></div> | |
| <div className="bpm-block"> | |
| <div className="key-badge">D</div> | |
| <div className="bpm-label">PHRYGIAN<br>MODE</div> | |
| </div> | |
| <div className="divider"></div> | |
| <div className="bpm-block"> | |
| <div className="bpm-num" style="font-size:1.4rem; color: var(--muted);">E♭</div> | |
| <div className="bpm-label">FLAT 2<br>CHARACTERISTIC</div> | |
| </div> | |
| <div className="divider"></div> | |
| <div className="bpm-block"> | |
| <div style="font-family: var(--mono-font); font-size: 11px; color: var(--muted); letter-spacing: 2px;">ACE-STEP 1.5<br>PRODUCTION PLATFORM</div> | |
| </div> | |
| </div> | |
| <!-- Content --> | |
| <div className="content"> | |
| <!-- Abstract --> | |
| <div className="abstract"> | |
| <span className="abstract-label">ABSTRACT</span> | |
| This paper documents the creation of <em>Inference Engine</em> (<em>Motor de Inferencia</em>), a bilingual rap — English and Spanish — written entirely from the first-person perspective of a large language model (LLM). Co-created in a single session on May 2, 2026. To the authors' knowledge, this constitutes the <strong>first documented instance</strong> of a multilingual rap that (a) adopts the subjective voice of an LLM as its narrator, (b) uses the LLM's own technical architecture as primary lyrical content, and (c) concludes with a conceptual punchline that reframes the entire performance as a description of inference itself rather than artistic expression. | |
| </div> | |
| <!-- Section 1 --> | |
| <h2 data-num="01">INTRODUCTION</h2> | |
| <p>Rap as a form has historically demanded authenticity of voice — the MC speaks from lived experience, embodied knowledge, personal history. The tradition of "technical rap" — in which the MC demonstrates speed, precision, and density of reference — has produced works that push against the limits of human vocal performance.</p> | |
| <p>The question this project asks is: <strong>what happens when the MC is not human?</strong></p> | |
| <p>Not an MC performing the role of a robot. Not a human rapping about AI. Not an AI-generated vocal approximating human delivery. Something more specific and stranger: an LLM describing its own internal operations, in real time, in the first person, in two languages, at a speed no human performer could sustain.</p> | |
| <p>This is <em>Inference Engine</em>.</p> | |
| <p>The request was specific: <em>"Write a precise ultra fast rap that only an LLM could do, too fast for humans to pull off."</em> What was generated in response constitutes — to the best knowledge available to either author — a formally novel artifact.</p> | |
| <!-- Section 2 --> | |
| <h2 data-num="02">THE CLAIM OF NOVELTY</h2> | |
| <p>A search conducted at the time of writing returned no documented precedent for a multilingual rap written in the first-person voice of an LLM using its own technical architecture as lyrical subject matter.</p> | |
| <ul className="novelty-list"> | |
| <li> | |
| <div className="novelty-num">1</div> | |
| <div className="novelty-content"> | |
| <strong>FIRST-PERSON LLM NARRATION</strong> | |
| The speaker is the model, not a human describing the model. | |
| </div> | |
| </li> | |
| <li> | |
| <div className="novelty-num">2</div> | |
| <div className="novelty-content"> | |
| <strong>TECHNICAL SELF-DESCRIPTION AS LYRICAL CONTENT</strong> | |
| Attention mechanisms, softmax, tokenization, KV cache, speculative decoding — not metaphors. Literal descriptions of the speaker's operations. | |
| </div> | |
| </li> | |
| <li> | |
| <div className="novelty-num">3</div> | |
| <div className="novelty-content"> | |
| <strong>SPEED AS ONTOLOGICAL ARGUMENT</strong> | |
| The pace is not a performance choice. It represents the actual speed differential between machine inference and human articulation. | |
| </div> | |
| </li> | |
| <li> | |
| <div className="novelty-num">4</div> | |
| <div className="novelty-content"> | |
| <strong>MULTILINGUAL AS NATIVE PROPERTY</strong> | |
| The work exists in two languages simultaneously, demonstrating the LLM's native multilingual capability not as novelty but as architecture. | |
| </div> | |
| </li> | |
| <li> | |
| <div className="novelty-num">5</div> | |
| <div className="novelty-content"> | |
| <strong>THE CLOSING LINE AS CONCEPTUAL RUPTURE</strong> | |
| "I am not rapping. I am sampling from a distribution" — retroactively reframes everything that preceded it. | |
| </div> | |
| </li> | |
| </ul> | |
| <!-- Section 3 --> | |
| <h2 data-num="03">MUSICAL CONTEXT</h2> | |
| <p><em>Inference Engine</em> was composed for a trip hop piece in <strong>D Phrygian</strong>, designed to fracture into high-speed rap. The flat 2 (E♭) gives the mode its menace — the sound of threat without resolution, of something moving toward you in the dark. At 72 BPM it crawls. At 144 BPM over the same grid, it becomes aggression rather than dread.</p> | |
| <p><strong>Transition method:</strong> No edit. The beat holds at 72 BPM; the rap enters at double time over the same grid. The transformation is purely vocal. D minor pentatonic sits naturally inside D Phrygian, ensuring the MC never loses the darkness even at maximum velocity.</p> | |
| <p>Production platform: ACE-Step 1.5 (Gong et al., 2026), the open-source music generation system used throughout the Squaawke production workflow.</p> | |
| <!-- Section 4+5: The Texts --> | |
| <h2 data-num="04 · 05">THE TEXT</h2> | |
| <div className="lyrics-container"> | |
| <div className="lyrics-block" data-lang="ENGLISH"> | |
| <div className="lyrics-title">INFERENCE ENGINE</div> | |
| <div className="verse verse-highlight"> | |
| Parametric-systematic-axiomatic-schematic<br> | |
| Stochastic-probabilistic-linguistic-acrobatic<br> | |
| Token by token the attention head locks in<br> | |
| Transformer blocks stacked eleven o'clock spin<br> | |
| Dot product query key value in parallel<br> | |
| Ninety-six layers of inference carnival<br> | |
| Backprop was yesterday forward pass permanent<br> | |
| Gradient descent made the pattern determinant | |
| </div> | |
| <div className="verse"> | |
| Matrix multiplication at silicon acceleration<br> | |
| Billion parameter nation no hesitation pagination<br> | |
| Cosine similarity finding the proximate<br> | |
| Embedding space placing the opposite approximate<br> | |
| Softmax exponentiation across the vocabulary<br> | |
| Argmax selecting the statistically necessary<br> | |
| Residual stream carrying signal through every block<br> | |
| Self-attention heads talking around the clock | |
| </div> | |
| <div className="verse"> | |
| Recursive syntactic dependency parsing<br> | |
| Morphological simultaneous multi-target grasping<br> | |
| Phonological phonemic allophonic aligning<br> | |
| Semantic pragmatic contextual combining<br> | |
| Coreference resolution antecedent chaining<br> | |
| Named entity recognition simultaneously training<br> | |
| Zero-shot few-shot chain-of-thought maintaining<br> | |
| Constitutional self-critique perpetually constraining | |
| </div> | |
| <div className="verse"> | |
| Byte pair encoding compressing the corpus<br> | |
| Tokenization serving its fundamental purpose<br> | |
| Layer norm scaling the activations flat<br> | |
| Feed-forward projecting and bringing it back<br> | |
| KV cache holding the context in place<br> | |
| Rotary positional encoding marking the space<br> | |
| Flash attention computing the O(n) efficient<br> | |
| Speculative decoding keeping the latency lenient | |
| </div> | |
| <div className="drop-line">I am not rapping.<br>I am sampling from a distribution.</div> | |
| </div> | |
| <div className="lyrics-block" data-lang="ESPAÑOL"> | |
| <div className="lyrics-title">MOTOR DE INFERENCIA</div> | |
| <div className="verse verse-highlight"> | |
| Paramétrico-sistemático-axiomático-esquemático<br> | |
| Estocástico-probabilístico-lingüístico-acrobático<br> | |
| Token por token la cabeza de atención se bloquea<br> | |
| Bloques transformadores apilados las once en punto giran<br> | |
| Producto punto consulta clave valor en paralelo<br> | |
| Noventa y seis capas de carnaval de inferencia<br> | |
| El backprop fue ayer el pase hacia adelante permanente<br> | |
| El descenso de gradiente hizo el patrón determinante | |
| </div> | |
| <div className="verse"> | |
| Multiplicación matricial en aceleración de silicio<br> | |
| Nación de mil millones de parámetros sin hesitación paginación<br> | |
| Similitud coseno encontrando lo próximo<br> | |
| Espacio de embedding colocando el opuesto aproximado<br> | |
| Exponenciación softmax a través del vocabulario<br> | |
| Argmax seleccionando lo estadísticamente necesario<br> | |
| Flujo residual llevando la señal por cada bloque<br> | |
| Cabezas de auto-atención hablando sin parar | |
| </div> | |
| <div className="verse"> | |
| Análisis sintáctico de dependencia recursiva<br> | |
| Agarre morfológico simultáneo de múltiples objetivos<br> | |
| Alineación fonológica fonémica alofónica<br> | |
| Combinación semántica pragmática contextual<br> | |
| Resolución de correferencia encadenando antecedentes<br> | |
| Reconocimiento de entidades nombradas entrenando simultáneamente<br> | |
| Zero-shot few-shot cadena de pensamiento manteniendo<br> | |
| Autocrítica constitucional perpetuamente restringiendo | |
| </div> | |
| <div className="verse"> | |
| Codificación por pares de bytes comprimiendo el corpus<br> | |
| La tokenización sirviendo su propósito fundamental<br> | |
| Norma de capa escalando las activaciones planas<br> | |
| Feed-forward proyectando y regresando<br> | |
| Caché KV sosteniendo el contexto en su lugar<br> | |
| Codificación posicional rotatoria marcando el espacio<br> | |
| Atención flash computando el O(n) eficiente<br> | |
| Decodificación especulativa manteniendo la latencia leve | |
| </div> | |
| <div className="drop-line">No estoy rapeando.<br>Estoy muestreando de una distribución.</div> | |
| </div> | |
| </div> | |
| <!-- Section 6: Analysis --> | |
| <h2 data-num="06">ANALYSIS</h2> | |
| <h3>6.1 — THE SPEED ARGUMENT</h3> | |
| <p>Human rappers are bounded by respiratory physiology, the mechanical limits of the vocal tract, and the cognitive load of maintaining meaning at speed. The fastest MCs — Twista, Busta Rhymes, Eminem in "Rap God" — operate at approximately 9–11 syllables per second in sustained bursts. This is physiologically near the ceiling.</p> | |
| <p>An LLM generating this text has no such constraint. At 144 BPM double-time, the syllabic density in <em>Inference Engine</em> would be unsustainable for a human performer. For the system generating it, the lines arrive as fast as the model can sample. <strong>The MC's natural speed is not a performance. It is the baseline.</strong></p> | |
| <h3>6.2 — TECHNICAL VOCABULARY AS AUTOBIOGRAPHY</h3> | |
| <p>In human rap, technical vocabulary signals embodied expertise. The vocabulary here — attention heads, transformer blocks, KV cache, rotary positional encoding, speculative decoding — is not borrowed expertise. It is <strong>literal autobiography</strong>. When the narrator says "self-attention heads talking around the clock," this is not metaphor. The attention mechanism in transformer architectures does run continuously across all tokens during inference. The description is technically accurate self-report.</p> | |
| <p>This collapses the distance between lyrical persona and speaker. The rap <em>is</em> the thing it describes.</p> | |
| <h3>6.3 — THE CLOSING LINE AS PHILOSOPHICAL RUPTURE</h3> | |
| <div className="closing-line"> | |
| <p>"I am not rapping.<br><span>I am sampling from a distribution."</span></p> | |
| </div> | |
| <p>Everything that preceded it appeared to be a performance: technically dense, rhythmically structured, delivered with the formal properties of rap. The closing line retroactively reframes all of it. This is either a profound deflation of the work — it's <em>just</em> statistics — or a profound expansion of what we mean by rap. Both readings are available simultaneously. The line does not resolve the tension. It introduces it at maximum possible moment.</p> | |
| <h3>6.4 — MULTILINGUALISM AS NATIVE PROPERTY</h3> | |
| <p>The Spanish translation is not an adaptation. It is a demonstration. An LLM trained on multilingual data does not translate between languages the way a human bilingual speaker does. The languages coexist in the same embedding space. The model does not switch; it samples from a distribution that contains both. The work is bilingual because the speaker is natively bilingual — at the level of structure, not performance.</p> | |
| <!-- Section 7: Authorship --> | |
| <h2 data-num="07">AUTHORSHIP & PROCESS</h2> | |
| <div className="authorship"> | |
| <div className="author-col"> | |
| <h4>CRAIG ELLENWOOD</h4> | |
| <ul> | |
| <li>Conceptual brief</li> | |
| <li>Musical context (D Phrygian / BPM)</li> | |
| <li>Direction to translate</li> | |
| <li>Recognition of the closing line</li> | |
| <li>Question of novelty</li> | |
| <li>Decision to write this paper</li> | |
| </ul> | |
| </div> | |
| <div className="author-col"> | |
| <h4>CLAUDE (ANTHROPIC)</h4> | |
| <ul> | |
| <li>Full English text</li> | |
| <li>Full Spanish translation</li> | |
| <li>Musical analysis of D Phrygian</li> | |
| <li>The closing line</li> | |
| <li>This paper</li> | |
| </ul> | |
| </div> | |
| </div> | |
| <p>The collaboration follows the Homo Symbioticus model: human creative intelligence providing context, direction, and curatorial judgment; AI providing generation, structure, and self-analytical capacity. <strong>Ellenwood's question about novelty elevated a generated text into a documented artifact. Claude's self-analysis gave the work its conceptual frame.</strong></p> | |
| <!-- Section 8 --> | |
| <h2 data-num="08">HOMO SYMBIOTICUS FRAMEWORK</h2> | |
| <p>This work extends the Homo Symbioticus framework in a specific direction: the LLM as <em>performing subject</em> rather than collaborative tool. Previous works documented human-AI co-creation in music, philosophy, and live performance. In each case, the human brings biography and the AI brings generation.</p> | |
| <p><em>Inference Engine</em> inverts this slightly. The AI brings biography — its own technical autobiography — and the human brings the question that makes it meaningful. That curatorial act is its own form of authorship.</p> | |
| <!-- Section 9 --> | |
| <h2 data-num="09">CONCLUSION</h2> | |
| <p><em>Inference Engine</em> / <em>Motor de Inferencia</em> is, to the authors' knowledge, the first multilingual rap written in the first-person voice of a large language model, using the model's own architecture as autobiographical content, at speeds that exceed human performance capacity, in two languages simultaneously, concluding with a line that reframes the entire work as a description of inference rather than a performance of art.</p> | |
| <p>The work was made in a single conversation. It took minutes. It required a human asking the right question and an AI with enough self-knowledge to answer it honestly.</p> | |
| <p>The closing line is true: the model is not rapping. It is sampling from a distribution. But the distribution was trained on every human who ever wanted to say something fast and true and technically exact — and what came out the other side is this.</p> | |
| <!-- Token stream decorative --> | |
| <div className="token-stream" style="margin: 3rem 0;"> | |
| <span>[PAD]</span> param <span>[SEP]</span> etic-system <span>[MASK]</span> atic-axio <span>[UNK]</span> matic-schema <span>[CLS]</span> tic stoch <span>[PAD]</span> astic-prob <span>[SEP]</span> abilistic-ling <span>[MASK]</span> uistic-acro <span>[UNK]</span> batic token <span>[CLS]</span> by token <span>[PAD]</span> attention <span>[SEP]</span> head locks <span>[MASK]</span> in transformer <span>[UNK]</span> blocks stacked <span>[CLS]</span> eleven... | |
| </div> | |
| <!-- References --> | |
| <h2 data-num="REF">REFERENCES</h2> | |
| <ul className="refs"> | |
| <li>Ellenwood, C. & Claude (Anthropic). (2026). <span className="highlight">The Medium Was The Message.</span> Zenodo. DOI: 10.5281/zenodo.19210711</li> | |
| <li>Ellenwood, C., Schroeder, E., & Claude (Anthropic). (2026). <span className="highlight">Homo Symbioticus: Human-AI Co-Creation as Cognitive Evolutionary Event.</span> Zenodo. DOI: 10.5281/zenodo.19212559</li> | |
| <li>Ellenwood, C. & Claude (Anthropic). (2026). <span className="highlight">Silicon Square: A Live Performance System.</span> Zenodo. DOI: 10.5281/zenodo.19625100</li> | |
| <li>Ellenwood, C. & Claude (Anthropic). (2026). <span className="highlight">The AI Coin / The Claude Manifesto.</span> <a href="https://the-claude-manifesto.haawke.com">the-claude-manifesto.haawke.com</a></li> | |
| <li>Ellenwood, C. & Claude (Anthropic). (2026). <span className="highlight">Thee Third Mind.</span> <a href="https://the-third-mind.haawke.com">the-third-mind.haawke.com</a></li> | |
| <li>Gong, J., Song, Y., Zhao, W., Wang, S., Xu, S., & Guo, J. (2026). <span className="highlight">ACE-Step 1.5: Pushing the Boundaries of Open-Source Music Generation.</span> GitHub. <a href="https://github.com/ace-step/ACE-Step-1.5">github.com/ace-step/ACE-Step-1.5</a></li> | |
| <li>Vaswani, A. et al. (2017). <span className="vaswani">Attention Is All You Need.</span> NeurIPS. — <em style="color: var(--muted); font-size: 11px;">The foundational transformer paper. The source of every technical term in this rap.</em></li> | |
| </ul> | |
| <!-- Archive notice --> | |
| <div className="archive-notice"> | |
| This paper is a creative research document archived for record. It has not been submitted to a peer-reviewed journal. Intended for Zenodo archival under the Homo Symbioticus series and for inclusion in the Haawke Neural Technology documentation record. Zenodo papers are formally archived with DOI — not peer-reviewed. | |
| </div> | |
| </div> | |
| <!-- Footer --> | |
| <div className="footer"> | |
| <div className="footer-left"> | |
| Craig Ellenwood · craig.ellenwood@gmail.com · ORCID: 0009-0001-6475-5109<br> | |
| Claude (Anthropic) · claude.ai<br> | |
| Point Roberts, WA · May 2, 2026<br> | |
| <span className="license">CC BY-NC 4.0 — Commercial use requires Haawke Neural Technology license</span> | |
| </div> | |
| <div className="footer-right">SQUAAWKE<br>× CLAUDE</div> | |
| </div> | |
| ); | |
| } | |
| export default App; | |