====== Today in AI: May 12, 2026 · 4 min read ====== **PwC's agent research just killed a myth: goal clarification only matters in the first 10% of task execution.** [[https://news.smol.ai/issues/26-05-11-not-much/|PwC research on time-dependent goal clarification]] shows that asking agents to clarify objectives early—before they commit to a strategy—drives massive efficiency gains. But after ~10% of work is done, the intervention window closes. This reframes how you architect agent workflows: front-load your goal negotiation or get nothing. The takeaway for builders: design for early clarity, not mid-course corrections. 🤖 **Coding agents are officially fragmented into three tiers.** [[opus_4_7_vs_gpt_5_5_vs_open_weight|Anthropic's Opus 4.7, OpenAI's GPT-5.5, and open-weight alternatives]] now dominate the SWE-bench landscape with measurably different cost-to-performance profiles. Proprietary models still win on raw accuracy; open-weight alternatives cut inference costs by 70%. Pick your poison based on whether you're optimizing for benchmark rank or production margin. For builders: the open-weight tier is finally competitive enough to ship. 🛠️ **Simon Willison just shipped LLM Templates, and it's a quiet game-changer.** [[https://til.simonwillison.net/llms/llm-shebang|YAML-based LLM Templates]] let you package prompts, system instructions, model selection, and tool definitions into reusable, portable units. This is what CI/CD should've been for AI workflows from day one. Builders can now version control their agent configs like regular code. 🏗️ **Autonomous agents just got a 7-day maintenance heartbeat.** [[autonomous_skill_maintenance|Skill decay in agent systems]]—where API integrations and tool bindings degrade over time—gets solved by periodic refresh cycles. [[https://alphasignalai.substack.com/p/you-should-install-hermes-agent-this|AlphaSignal's write-up on Hermes Agent]] shows this isn't theoretical; teams are shipping it now. For builders: automation that maintains itself wins the reliability war. 🔬 **Byte-level modeling might finally dethrone tokenizers.** Operating directly on raw bytes (0-255) instead of subword tokens eliminates vocabulary ceilings and enables true multilingual flexibility. The cost is compute; the win is universality. Watch this space—it's the tokenizer debate restarted with better hardware. 💰 **Oracle Developers released 16 reasoning strategies in one framework.** [[https://www.therundown.ai/p/google-deepmind-powerful-ai-co-mathematician|The Oracle Developers Agent-Reasoning Framework]] ships open-source reasoning tactics that work with Ollama without retraining models. This is democratizing what cost $10M to develop two years ago. Builders can now layer sophisticated reasoning onto any LLM. **Silences:** Gemini 3.5 still nowhere. Llama 4 is radio silence. Meta's multimodal agenda remains opaque. That's the brief. Full pages linked above. See you tomorrow.