Externalization in LLM Agents: Unified Review of Memory and Harness Engineering

hackernews | | 🔬 연구
#ai 딜 #claude #머신러닝 #머신러닝/연구 #생산성 #연구 #프롬프트엔지니어링
원문 출처: hackernews · Genesis Park에서 요약 및 분석

요약

최신 LLM 에이전트는 모델 자체의 변경보다는 실행 시간 환경을 재구성하는 방식으로 발전하고 있습니다. 이 연구는 기억, 기술, 프로토콜을 외부화하여 모델의 인지적 부담을 줄이고 신뢰성을 높이는 하니스 엔지니어링의 중요성을 강조합니다. 또한 모델과 외부 인프라가 상호 작용하는 시스템 수준의 프레임워크를 통해 실용적인 에이전트 개발의 방향성을 제시합니다.

본문

Computer Science > Software Engineering Title:Externalization in LLM Agents: A Unified Review of Memory, Skills, Protocols and Harness Engineering View PDF HTML (experimental)Abstract:Large language model (LLM) agents are increasingly built less by changing model weights than by reorganizing the runtime around them. Capabilities that earlier systems expected the model to recover internally are now externalized into memory stores, reusable skills, interaction protocols, and the surrounding harness that makes these modules reliable in practice. This paper reviews that shift through the lens of externalization. Drawing on the idea of cognitive artifacts, we argue that agent infrastructure matters not merely because it adds auxiliary components, but because it transforms hard cognitive burdens into forms that the model can solve more reliably. Under this view, memory externalizes state across time, skills externalize procedural expertise, protocols externalize interaction structure, and harness engineering serves as the unification layer that coordinates them into governed execution. We trace a historical progression from weights to context to harness, analyze memory, skills, and protocols as three distinct but coupled forms of externalization, and examine how they interact inside a larger agent system. We further discuss the trade-off between parametric and externalized capability, identify emerging directions such as self-evolving harnesses and shared agent infrastructure, and discuss open challenges in evaluation, governance, and the long-term co-evolution of models and external infrastructure. The result is a systems-level framework for explaining why practical agent progress increasingly depends not only on stronger models, but on better external cognitive infrastructure. Bibliographic and Citation Tools Code, Data and Media Associated with this Article Demos Recommenders and Search Tools arXivLabs: experimental projects with community collaborators arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them. Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Genesis Park 편집팀이 AI를 활용하여 작성한 분석입니다. 원문은 출처 링크를 통해 확인할 수 있습니다.

공유

관련 저널 읽기

전체 보기 →