A language model that emits raw VM opcodes instead of text

hackernews | | 📰 뉴스
#ai agents #llm #opcodes #virtual machine #하드웨어/반도체
원문 출처: hackernews · Genesis Park에서 요약 및 분석

요약

I've built a demo about this concept.I replaced the decoder head of a Qwen 1.5B with a small cross-attention head (38M params) that emits raw CHIP-8 opcodes. Every opcode executes on a real CHIP-8 emulator.Interesting failure: "3 plus 5" works and draws 8, but "two plus three" produces wrong operands.

본문

A few months ago I posted asking why AI agents control machines through human text instead of emitting machine instructions directly. I&#x27;ve built a demo about this concept.<p>I replaced the decoder head of a Qwen 1.5B with a small cross-attention head (38M params) that emits raw CHIP-8 opcodes. The LLM encodes the instruction once and never generates a token. The head outputs machine instructions by attending to the actual machine state.</p><p>It handles arithmetic with BCD, subroutine calls, timer wait loops, and conditional branching. 1-3ms per opcode. Every opcode executes on a real CHIP-8 emulator.</p><p>Interesting failure: &quot;3 plus 5&quot; works and draws 8, but &quot;two plus three&quot; produces wrong operands. The frozen LLM&#x27;s hidden states for &quot;two&quot; and &quot;2&quot; are nearly orthogonal (cosine sim 0.09) in this context. Removing the decoder apparently removes the path the LLM would use to bridge word-form and digit-form numbers.</p><p>Demo and code: https:&#x2F;&#x2F;github.com&#x2F;ilbertt&#x2F;reflex</p>

Genesis Park 편집팀이 AI를 활용하여 작성한 분석입니다. 원문은 출처 링크를 통해 확인할 수 있습니다.

공유

관련 저널 읽기

전체 보기 →