Where AI development sits in 2026
An honest read on where AI development is in 2026: what's actually in production, what's still vaporware, and what it means for the kind of businesses we build software for.
We work with small businesses who keep getting pitched on AI. Most of the pitches are overstated. Some are not. This is a short accounting of what’s actually shipping in production right now, written for founders and operators who want to know what to take seriously and what to ignore.
Agentic AI is real, but narrower than the demos suggest
“Agents” is the word of the year. The demos look impressive: a language model picks tools, clicks through a browser, fills in a form, drafts an email, books the meeting.
In production, the useful version of this is much less ambitious. It looks like:
- A support agent that routes tickets, drafts a first reply, and waits for a human to send it.
- A data-entry agent that takes a PDF invoice, pulls out vendor, date, line items, and totals, and writes them to a spreadsheet.
- A research agent that reads your internal docs and answers questions with citations.
All of these are real and shipping. The fully-autonomous, multi-step agent that runs your business while you’re at the beach is not, and the companies trying to sell you one are optimistic.
GitHub’s own numbers back this up: commits and PR volume are up meaningfully year over year, largely because developers are using AI tools as a faster typewriter. They’re not firing engineers. They’re writing more code per engineer.
Small models got good enough for most tasks
Two years ago, you needed GPT-4 class models for anything non-trivial. In 2026, an 8B-parameter open-weight model (Llama 3.1, Qwen, Mistral) runs on a single consumer GPU and handles:
- Customer support replies.
- Document summarization.
- Classifying, tagging, and routing.
- Structured extraction from messy text.
For a specific task in a specific domain, a smaller fine-tuned model often beats a larger general one on both cost and accuracy. The frontier labs still matter for hard reasoning problems, but the median use case has moved down.
Practically, that means you can run a lot more AI locally than you could last year. Cheaper per request, no data leaving the building, and no dependency on whoever’s pricing page you’ve been staring at.
Multimodal is where the practical wins are
Text-only models were a bottleneck. A lot of the real work businesses want automated — invoices, receipts, ID documents, chart screenshots, warehouse photos — has always been in images or PDFs with mixed layouts.
In 2026, vision models are good enough to read those inputs reliably. The applications are unglamorous: extracting line items from a receipt, checking a warehouse photo for damage, reading a handwritten form. But they’re the ones that actually save operators time.
Physical AI is still slow
There’s a lot of excitement about robots and physical AI. In warehouse and manufacturing environments, where the task is constrained and repetitive, there’s real deployment happening. Anywhere else, it remains slower and more brittle than the videos suggest. If a vendor tells you their robot will handle a general task in an unstructured environment, ask for a reference customer.
What this means for a small business
If you own a tiffin service or a painting company, you don’t need a strategy for “AI.” You probably need one or two specific things automated:
- A cheaper first-line support reply that a human reviews before sending.
- Receipt or invoice processing that doesn’t require someone retyping totals into a spreadsheet.
- Search over your own content (FAQs, policies, SOPs) that doesn’t require paying an employee to remember everything.
Those are all tractable with current tools, run cheaply, and don’t require a platform rewrite. The rest — the autonomous agents, the digital workforce, the enterprise transformation — is mostly a longer timeline than the pitch suggests.
If you want help figuring out which one of those actually applies to your business, send us a note.