A language model that emits raw VM opcodes instead of text
hackernews
|
|
📰 뉴스
#ai agents
#llm
#opcodes
#virtual machine
#하드웨어/반도체
원문 출처: hackernews · Genesis Park에서 요약 및 분석
요약
I've built a demo about this concept.I replaced the decoder head of a Qwen 1.5B with a small cross-attention head (38M params) that emits raw CHIP-8 opcodes. Every opcode executes on a real CHIP-8 emulator.Interesting failure: "3 plus 5" works and draws 8, but "two plus three" produces wrong operands.
본문
A few months ago I posted asking why AI agents control machines through human text instead of emitting machine instructions directly. I've built a demo about this concept.<p>I replaced the decoder head of a Qwen 1.5B with a small cross-attention head (38M params) that emits raw CHIP-8 opcodes. The LLM encodes the instruction once and never generates a token. The head outputs machine instructions by attending to the actual machine state.</p><p>It handles arithmetic with BCD, subroutine calls, timer wait loops, and conditional branching. 1-3ms per opcode. Every opcode executes on a real CHIP-8 emulator.</p><p>Interesting failure: "3 plus 5" works and draws 8, but "two plus three" produces wrong operands. The frozen LLM's hidden states for "two" and "2" are nearly orthogonal (cosine sim 0.09) in this context. Removing the decoder apparently removes the path the LLM would use to bridge word-form and digit-form numbers.</p><p>Demo and code: https://github.com/ilbertt/reflex</p>
Genesis Park 편집팀이 AI를 활용하여 작성한 분석입니다. 원문은 출처 링크를 통해 확인할 수 있습니다.
공유