Show HN: Prawduct, a product development framework for Claude Code

hackernews | | 📰 뉴스
#ai 모델 #ai 소프트웨어 #claude #claude code #show hn #개발 프레임워크 #품질 관리
원문 출처: hackernews · Genesis Park에서 요약 및 분석

요약

Prawduct는 AI 기반 코드 생성 시 흔히 발생하는 요구사항 이탈과 기술 부채 누적 문제를 해결하기 위해 체계적인 계획 수립과 독립적인 품질 관리를 제공하는 클로드 코드(Claude Code)용 제품 개발 프레임워크입니다. 이 프레임워크는 가족용 스크래치패드 같은 소규모 작업에 5~8개, 금융 플랫폼 같은 고위험 작업에 15~25개의 질문을 던지는 등 프로젝트의 위험도에 따라 리스크 관리 및 발굴(Discory) 단계의 깊이를 자동으로 조절합니다. 특히 빌더 에이전트의 추론에 접근할 수 없는 완전히 분리된 '크리틱(Critic)' 에이전트가 코드를 독립적으로 검토하여 맹점을 보완하며, 완성된 제품에는 런타임 의존성 없이 자체적인 학습 기록과 관리 시스템만 독립적으로 유지되어 이식성이 뛰어납니다.

본문

Prawduct is a product development framework for Claude Code that adds structured planning, independent quality reviews, and continuous per-project learning to AI-assisted software development. Going from "I need an app that does X" straight to code skips the hard questions: Who are the users? What are the edge cases and failure modes? What does "done" look like? What needs to be tested, and how? Claude Code is fantastic at writing code, but without discipline it makes assumptions about product intent, produces code that drifts from requirements, skips edge cases, weakens tests to make them pass, and accumulates technical debt — all without telling you. You describe what you want to build, either a net-new product or enhancements to an existing one. Prawduct scales governance to match the work: Discovery — Asks about your users, workflows, edge cases, security, and scope. Scales question depth to risk: a family scratchpad gets 5-8 questions; a financial platform gets 15-25. Discovery is continuous — new features need their own discovery. Planning — Produces structured specifications in dependency order: product brief, data model, security model, test specifications, non-functional requirements, and a chunked build plan. Building — Implements the product in governed chunks. Governance depth scales with work size (trivial → large) and type (bugfix → feature → refactor). Each chunk follows a cycle: read spec, write tests alongside implementation, verify, then submit for independent Critic review. The Critic runs as a separate agent with no access to the builder's reasoning — it sees only the code and specs, catching things the builder's own context blinds it to. Reflection and Learning — After each significant action, captures what happened, whether it was expected, and what it teaches. Learnings follow a lifecycle (provisional → confirmed → incorporated) and accumulate across sessions. Learnings are checked when planning new work, closing the loop. Telling an LLM to "always do X" works until context gets large, and those instructions degrade with compaction. Prawduct enforces governance at four levels: - Session briefing — On session start, a staleness scan checks artifacts against code reality and delivers a structured briefing with project context, warnings, and relevant learnings - Critic review — A session hook blocks completion if code was modified against a build plan but no independent review happened. The Critic skill has structural tool restrictions preventing test/build execution - Session reflection — A session hook blocks completion if no reflection was captured (skipped for doc-only changes) - Compliance canary — At session end, informational checks flag common governance failures (code without tests, dependencies without rationale, broad exception handling) Everything else is governed by 22 principles and four methodology guides that stay in context via CLAUDE.md. The Critic runs as a Claude Code skill with context: fork (separate context) and allowed-tools that prevent running tests, builds, or executables. It has no access to the builder's reasoning or justifications — only the code, tests, and specifications. This structural separation catches blind spots that in-context review misses. The builder records test evidence (.prawduct/.test-evidence.json ) during verification; the Critic reads it instead of re-running the suite. All projects suffer drift over time. Each individual review can be executed perfectly, but accumulation over time means cruft appears, old code is not updated to new architectural patterns, tests go stale, documentation goes stale, Git accumulates dead branches, etc. The janitor skill is focused on periodic repo maintenance to catch these kinds of issues that are next to impossible for humans or LLMs to be perfect at during day-to-day work. Generated product repos carry everything they need: own CLAUDE.md, own hooks, own Critic instructions, own learning history. There is no runtime dependency on the Prawduct framework. Products work identically whether the framework repo exists or not. This means: - Products are portable and shareable - The framework is a generator, not a runtime dependency - Each product can evolve independently - Framework updates propagate via automatic, edit-preserving sync (requires having the framework cloned to a sibling dir) The framework detects structural characteristics (human interface, API, background automation, multi-party, sensitive data, distributed) and scales everything accordingly — discovery depth, artifact detail, test coverage, review intensity. A quick utility doesn't get exhaustive governance; a platform doesn't get hobby-grade review. Learnings are captured during development, and read at session start to inform decisions. A two-tier system keeps active rules concise ( let's run through tests/scenarios/home-environmental-monitor.md The coordinating Claude spawns a second session that does the work without seeing the full scenario

Genesis Park 편집팀이 AI를 활용하여 작성한 분석입니다. 원문은 출처 링크를 통해 확인할 수 있습니다.

공유

관련 저널 읽기

전체 보기 →