Back to The Times of Claw

The Future of AI Agents in Business

AI agents are moving from toys to infrastructure. Here's what that looks like in practice, and why local-first deployment is the architecture that survives.

Kumar Abhirup
Kumar Abhirup
·10 min read
The Future of AI Agents in Business

There's a pattern I keep noticing in conversations about AI in business. People are oscillating between two failure modes: either dramatically overstating what current AI agents can do, or dramatically understating where this is going in three to five years.

The reality is more interesting and more specific than either extreme. AI agents are becoming real infrastructure for businesses — not in the "AI will replace everyone" sense, but in the much more useful "AI will handle the tedious connective tissue work" sense. And the companies that figure this out are going to have a meaningful operational advantage over those that don't.

Let me tell you what I actually think is happening, based on building one of these systems.

The Current State Is Genuinely Useful#

Before speculating about the future, it's worth being precise about what AI agents can do reliably right now. Not hypothetically — right now, in production, with current models.

Structured data extraction. If you describe something in natural language — "I just had a call with Sarah Chen, she's the Head of Product at Stripe, interested in the enterprise tier, needs to come back to us in Q3 after their budget cycle" — a good agent can parse this and create a structured CRM record with zero friction. Name, company, role, status, follow-up date, notes. This is more reliable than you'd expect and saves meaningful time for anyone who takes a lot of meetings.

Information retrieval and synthesis. "Show me all my contacts at fintech companies who I haven't talked to in a month" is a question that requires querying a database and returning readable results. Agents do this well. "Summarize my relationship with this company before my call with them" — also reliable, if the agent has access to the relevant notes and history.

Draft generation in context. Writing a follow-up email that references specific conversation points, or drafting a proposal that uses the account's context, is something current models do well. Not perfectly. But well enough to be a useful first draft that saves 20-30 minutes.

Routine classification and routing. Categorizing inbound leads, tagging contact records, identifying duplicates — these are high-volume, low-judgment tasks that agents handle with high accuracy.

This is the current state. It's useful, not magic. The value is in consistency and speed, not in some kind of emergent intelligence.

Where Agents Break Down (Honestly)#

Before claiming a rosy future, I want to be honest about where current AI agents are unreliable.

Multi-step reasoning with real stakes. If I ask an agent to analyze which accounts are at churn risk and recommend interventions, I can't fully trust the output without reviewing it. The agent might weight the wrong signals. It might miss context I haven't explicitly provided. It might be confidently wrong.

Novel situations. Agents trained on general patterns apply those patterns even when the situation is unusual. If your business has some idiosyncratic characteristic that makes the general patterns wrong, the agent won't automatically know to adjust.

Long-horizon tasks. "Manage my pipeline this week" is too vague and too consequential to hand off entirely. You'll get something reasonable but not quite right. The agent doesn't know enough about the specific relationships, the specific context, the specific sensitivities.

This isn't unique to my system. It's the honest picture of AI agents in 2026.

The Architecture That Survives#

Here's what I've learned building DenchClaw: the architecture of AI deployment matters as much as the AI itself.

Most businesses right now are deploying AI via SaaS APIs. You send data to a cloud service, you get back a result. This works for many use cases, but it has fundamental limitations.

Latency. Every AI operation that goes through a cloud service adds 100-500ms of network latency on top of the model inference time. For interactive applications, this is the difference between a tool that feels instant and a tool that feels slow.

Data privacy. Every time your business data leaves your network, you're accepting some risk. Your contact notes, your deal terms, your customer relationships — all of this goes to a cloud service that has its own data practices.

Reliability. Cloud AI services have outages. They have rate limits. They have quota exhaustion. A local AI deployment doesn't.

Cost at scale. API pricing per token makes sense for occasional use. It becomes expensive at scale. A local model has zero marginal cost per inference.

The architecture that survives for business AI looks like this: a local agent that handles the majority of operations using local models (for speed, privacy, and cost), with selective calls to cloud models for tasks that genuinely require their capability (complex reasoning, advanced code generation, multimodal tasks).

DenchClaw is built this way. Most operations run locally. Embedding for semantic search, classification, entity extraction — all local. Complex generation tasks optionally call cloud APIs when the quality justifies it.

What Changes in the Next Three Years#

I want to make some specific predictions about what business AI agents will look like by 2029. These are based on trajectories I can see clearly, not speculation.

Agents will have persistent, deep context about your business. Right now, AI assistants are mostly stateless — they don't really remember last week's conversation. In three years, agents will maintain rich, structured knowledge about your business: your key accounts and the history of each relationship, your pipeline patterns, your competitive positioning, your team's strengths and weaknesses. Not as a chat transcript, but as a structured knowledge base that the agent actively maintains.

DenchClaw's memory system — MEMORY.md, daily logs, the DuckDB database — is an early version of this. In three years, this will be much more sophisticated.

Agents will take more initiative. Right now, agents are mostly reactive — you ask, they respond. The agents that survive will be proactive: "I noticed you haven't followed up with this account in three weeks and they had a meeting with your competitor last Tuesday based on their LinkedIn activity. Do you want me to draft a check-in message?" That's an agent that's watching, connecting dots, and surfacing insights you didn't ask for but needed.

The interface will mostly be conversation. Not exclusively — dashboards and tables are still useful for certain kinds of information. But the primary interface for most business data interactions will be conversational, because conversation is faster and more flexible than menus. The question is which conversations happen locally, which happen via phone, and which happen on whatever new interface emerges.

Local compute wins at the edge. As models get more efficient through quantization and architectural improvements, the compute required for business AI inference will continue to drop. The $1,200 Apple Silicon laptop today can run models that required a $50,000 server five years ago. That trajectory continues. In three years, your laptop runs better AI locally than your company's cloud deployment runs today.

The Skill Architecture Is the Key Abstraction#

One thing I'm increasingly convinced of: the way AI agents extend their capabilities matters more than the baseline capability of the model.

In DenchClaw, we use a skills architecture — capabilities are defined in markdown SKILL.md files that describe what a skill does, what inputs it needs, and what operations it performs. The agent reads these files to understand its capabilities and composes them to handle complex requests.

This is different from the approach most people take, which is to hardcode capabilities as function signatures or API endpoints. The skills approach has a key property: it's human-readable and community-maintainable. A developer can write a SKILL.md file in an afternoon that extends the agent's capabilities significantly. The skill file gets shared. Others install it. The agent collectively gets smarter.

I think this is the right abstraction for business AI. Not because it's clever, but because it aligns incentives. The community that uses a tool has the strongest incentive to extend it. A skills marketplace where businesses can publish capabilities they've built is how AI agents get genuinely useful for specific industries and use cases, not just generic tasks.

Why Local-First Deployment Wins Long-Term#

I've been thinking about the long-term equilibrium for AI agents in business, and I keep coming back to the same conclusion: local-first deployment wins.

Not because cloud AI won't continue improving. It will. But because the economics and the trust model both favor local.

Economics: As models get smaller and more efficient, the cost of local inference approaches zero. The marginal cost of cloud inference doesn't. At scale, a local agent is dramatically cheaper than a cloud agent.

Trust: A local agent that has access to your most sensitive business data is aligned with you. A cloud agent that has access to that data is aligned with the platform that runs it. These aren't always the same interest.

Reliability: A local agent works when the internet is down, when the API is rate-limited, when the vendor has an outage. This sounds minor until it isn't.

Privacy: A local agent processes your data on your hardware. A cloud agent processes it on someone else's hardware. For most business data, the former is clearly preferable.

The future of AI agents in business is local-first, deeply contextual, proactively helpful, and built on skill architectures that the community can extend. We're building that at DenchClaw. I think we'll look back at the era of cloud-only AI assistants the same way we look at the era of cloud-only everything: as a necessary transition phase, not the permanent state.

Frequently Asked Questions#

Are AI agents ready to be a primary business tool today?#

For specific, well-defined tasks (structured data entry, information retrieval, draft generation, classification), yes. For complex judgment calls or novel situations, they're best as tools that support human decision-making rather than replacing it.

How do I keep my business data private when using AI?#

Use a local-first system where possible. When you do call cloud AI APIs, review what data is included in the context. Avoid sending sensitive details like specific deal terms or pricing to cloud models when a local model can handle the task.

What's the difference between AI features in SaaS CRMs and an AI-native system?#

SaaS CRMs bolt AI onto an existing form-and-table interface. An AI-native system like DenchClaw is designed from the ground up for AI operation — the data model, the interface, the skill architecture are all built around AI as the primary operator.

How do agents handle things they don't know about my business?#

The key is giving the agent structured context — clear descriptions of your processes, your terminology, your specific use case. DenchClaw's AGENTS.md, USER.md, and MEMORY.md files are how you provide this context. The agent is only as good as the context it has.

When will AI agents be reliable enough to operate autonomously?#

For narrow, well-defined tasks with clear success criteria, they're already reliable enough for supervised autonomy. For broad, open-ended tasks, I'd expect 3-5 more years of development before unsupervised autonomy is appropriate.

Ready to try DenchClaw? Install in one command: npx denchclaw. Full setup guide →

Kumar Abhirup

Written by

Kumar Abhirup

Building the future of AI CRM software.

Continue reading

DENCH

© 2026 DenchHQ · San Francisco, CA