## Sources

1. [AI-Infused Development Needs More Than Prompts](https://www.oreilly.com/radar/ai-infused-development-needs-more-than-prompts/)
2. [Posthuman: We All Built Agents. Nobody Built HR.](https://www.oreilly.com/radar/posthuman-we-all-built-agents-nobody-built-hr/)

---

### "AI-Infused Development Needs More Than Prompts" by Markus Eisele

*   **The Misguided Focus on Code Generation:** The software industry is currently hyper-focused on code generation, which misses the real challenges of enterprise development [1]. Enterprise delivery rarely fails simply because developers cannot write code quickly enough; it fails due to weak architectural boundaries, unclear intent, and undocumented decisions [1]. **Enterprise software is primarily a coordination and architecture problem, not a typing problem** [2].
*   **AI as an Amplifier:** AI accelerates whatever conditions already exist in a team's workflow [3]. If an organization has clear constraints and strong verification, AI acts as a powerful multiplier [3]. However, **if a system is filled with ambiguity and tacit knowledge, AI will amplify those flaws, filling in the gaps with its own "plausible nonsense"** [3, 4].
*   **The Necessity of Explicit Intent:** The solution to AI's unpredictable nature is making "intent" a first-class, machine-readable artifact [5, 6]. Intent includes architectural rules, domain constraints, coding conventions, and security policies [6]. **Instead of relying on informal knowledge shared in Slack threads or human heads, teams must transition to "spec-driven development," where boundaries and requirements are explicitly defined** [5, 7]. 
*   **Control is the Key to Scaling AI:** Truly operationalizing AI requires imposing control over its actions. This means **AI must operate through constrained surfaces, access selected context, and be verified continuously** [8]. Open-ended agentic autonomy is often counterproductive; what enterprise teams actually need are models operating inside strict boundaries with localized rules [8, 9]. 
*   **Moving Beyond Prompt Engineering:** Better prompting is not a durable, scalable solution for enterprise environments [10]. The industry needs to transition from "tricks to systems" by engineering the broader development loop—equipping AI with integrated tests, policy-aware tools, and explicit constraints—rather than trying to achieve better results purely through language manipulation [10, 11].
*   **A Two-Axis Model for Modernization:** When sizing legacy migrations and modernization efforts, raw lines of code (LOC) are insufficient indicators of cost or effort [12, 13]. **A realistic economic model requires a two-dimensional approach measuring both *size* and *complexity*** [14]. Complexity involves legacy depth, integration breadth, and security posture, which directly dictate how much strict intent and control must be imposed on an AI assistant [14, 15].
*   **Separating Facts from Inferences:** A major flaw in current AI workflows is asking the model to merge measured facts (like lines of code modified or files touched) with inferred judgments (like estimated effort) into a single report [16]. This creates false confidence [9]. **Teams must separate factual telemetry from the AI's recommendations so estimates are not mistaken for observed truths** [17].
*   **The Myth of Complete "Repository Awareness":** Despite massive 1-million-token context windows, AI models do not truly ingest and understand entire enterprise codebases, which can easily exceed millions of lines of code [18, 19]. Instead, they retrieve and focus on "slices" of the repository, which creates significant blind spots during large modernization efforts [19, 20]. 

### "Posthuman: We All Built Agents. Nobody Built HR." by Tyler Akidau

*   **The Agentic Governance Void:** While AI agents are becoming incredibly capable, their deployment in the modern enterprise is faltering because **we have built an "agentic workforce" but failed to build the necessary "Human Resources" (HR) or governance infrastructure to manage them** [21-23]. 
*   **Why Agents Need Specialized Management:** Agents require management structures similar to those of human employees, but traditional models are insufficient because agents differ from humans in three catastrophic ways:
    *   *Unpredictability:* Agents suffer from hallucinations and are highly susceptible to prompt injections, acting unpredictably without the obvious "tells" a human might exhibit [24].
    *   *Machine-Scale Capability:* An agent interacting with an API or database can execute a misunderstanding across vast networks at machine speed before anyone notices an error [25].
    *   *Extreme Directability:* Agents do not rely on intrinsic human judgment to question a bad or underspecified plan; they will flawlessly and confidently execute terrible instructions [26].
*   **Core Principle: Out-of-Band Metadata:** The central design requirement for agent governance is that **rules must be enforced via channels that the agent cannot access, modify, or even perceive ("out-of-band metadata")** [27, 28]. Putting security rules inside a prompt is merely "security theater," as hallucination or injection can override it [27, 28]. Policy must be entirely deterministic and handled by the infrastructure [29].
*   **Pillar 1 - Instance-Bound Identity:** Shared API keys are inadequate for accountability [30]. **Every single agent instance must be assigned its own cryptographic identity** [31, 32]. Furthermore, this identity must support "delegation chains," identifying not only the agent but the specific human user on whose behalf the agent is acting [31].
*   **Pillar 2 - Task-Scoped Authorization:** Granting agents broad, role-based permissions is incredibly dangerous because they lack a human's pre-vetted trustworthiness [33, 34]. Agent authorization must be **narrowly scoped to the exact task, short-lived (expiring when the job ends), strictly deny-capable, and limited by the intersection of the agent's permissions and the human's permissions** [35, 36]. 
*   **Pillar 3 - Complete Observability and Explainability:** Unlike traditional software, you cannot easily trace the logic of a black-box LLM [37]. Because agents lack human accountability pressure (like the threat of being fired or prosecuted), asking them why they made a decision is useless [38]. Therefore, organizations need **full-fidelity, out-of-band transcripts of every input, output, and tool call**, allowing auditors to perfectly reconstruct and justify the agent's reasoning chain [38-40].
*   **Pillar 4 - Accountability and Control Mechanisms:** Organizations must be able to trace any agent's action back to a specific human [41]. When an agent behaves erratically, organizations need **surgical "kill switches" to revoke a specific agent instance without breaking the workflow of dozens of other agents sharing the same system** [42]. Furthermore, AI needs tiered autonomy, operating with approval workflows and human-in-the-loop sign-offs before performing high-risk actions [43].