Eric Jang examines the evolution of AI agents beyond simple chatbots, highlighting improved reasoning capabilities alongside unsustainable inference compute requirements as agentic systems scale.

When Vannevar Bush envisioned the "memex" in 1945, he imagined an associative memory system that could augment human intellect. Eighty years later, Eric Jang argues we're witnessing the emergence of such systems through AI agents – but with computational demands that threaten their viability.
Beyond Prompt Engineering: The Agentic Shift Recent advances have moved beyond simple instruction-following models. Modern agents (ReAct paradigm, tree-of-thought architectures) demonstrate multi-step reasoning capabilities. These systems can decompose complex queries, execute research workflows, and self-correct using techniques like:
- Recursive self-improvement loops
- Tool augmentation (web search, code execution)
- Environment-aware action selection
Jang notes that Anthropic's Claude Code exemplifies this shift, reportedly authoring 4% of public GitHub commits and projected to reach 20% by year's end. Unlike earlier models that merely completed code snippets, Claude demonstrates contextual awareness across entire codebases.
The Compute Time Bomb Agentic systems consume orders of magnitude more compute than single-prompt models. Each reasoning step requires full inference passes, creating quadratic scaling costs. Current projections indicate:
- Agent tasks require 10-100x more compute than equivalent chatbot interactions
- Inference demands will outpace training costs by 2027
- Energy requirements could consume entire regional power grids
This aligns with Nvidia CEO Jensen Huang's recent defense of AI infrastructure spending, though Jang questions the ecological sustainability of current scaling trajectories.
Practical Limitations Despite impressive demos, fundamental constraints remain:
- Brittle reasoning: Agents frequently derail on long-horizon tasks (beyond 50 steps)
- Hallucination amplification: Errors compound in multi-step workflows
- Tool limitations: APIs constrain agents to predefined action spaces
- Security risks: Autonomous systems create novel attack surfaces
Automated research systems exemplify both promise and peril. While agents can rapidly synthesize papers (e.g., DeepMind's Genie 3), they struggle with novel concept integration and often reproduce training data biases.
The Path Forward Jang suggests three critical focus areas:
- Architecture innovations: Sparse activation models like Mixture-of-Experts to reduce inference costs
- Formal verification: Techniques to mathematically guarantee agent behavior boundaries
- Hybrid systems: Human-agent collaboration frameworks where humans provide high-level oversight
As agent capabilities evolve from specialized tools toward general assistants, the field must confront the uncomfortable reality that exponential capability growth requires exponential resources – a tradeoff demanding fundamental architectural breakthroughs beyond scaling existing paradigms.

Comments
Please log in or register to join the discussion