10 min read

The AI Worker That Fits in a Folder

Send someone a folder. They open it on a laptop. A competent specialist wakes up and starts working.

That sounds like a packaging detail. It is not. It is a different idea of software.

Most people still picture AI as a service you reach through somebody else's window. You subscribe, log in, paste data, hope the permissions are sane, and accept that the useful part lives on a server you do not control. The more capable the system gets, the more that arrangement starts to feel backward. If an AI can actually do work, then the interesting question is no longer only how smart the model is. It is how that capability gets packaged, moved, audited, and handed to someone else.

Anthropic's "skills" point toward this shift in a surprisingly concrete way.

Capability becomes packaging

A skill, at least in the form people are now experimenting with, is almost boring on the surface. It is a folder. Inside, there is usually a Markdown file with instructions. That file can reference other text files, examples, PDFs, schemas, code snippets, and full libraries. You are not just telling the model what to do. You are surrounding it with a working environment and a stack of context it can keep reaching for.

That matters because prompt engineering has always had a hidden weakness: it asks one text blob to do too much. It has to define the role, encode the rules, smuggle in the domain knowledge, and sometimes even describe tools that already exist elsewhere. A skill breaks that apart. Instructions live in one place. Reference material lives in another. Reusable code lives where code belongs. The model gets access to a system that already has shape.

This is closer to hiring than querying. A good employee is not just a brain with a job title. They arrive into a company with manuals, templates, prior decisions, approved tools, old reports, and weird process documents nobody fully understands but everybody depends on. Skills start to give AI the same kind of scaffolding. They turn "be helpful" into "here is the playbook, here are the forms, here are the functions, here is the body of knowledge, now execute."

That shift sounds incremental until you notice what moves. The center of gravity drifts away from the raw model and toward the package wrapped around it. Two teams can use the same base model and end up with radically different performance because one bundle has clean tools, reliable references, tight instructions, and working code. The other has a heroic prompt and vibes.

We have seen this movie before. A compiler is useful. A compiler plus libraries, package management, docs, build scripts, and conventions becomes a real ecosystem. AI is walking through the same door.

A pod that carries its own brain

Take that skill idea to its logical conclusion and you get something much more interesting than a folder of instructions. You get a self-contained pod.

Inside that pod sits the local model, the runtime that can execute it, the system prompt, the tool definitions, the MCP server or equivalent connectors, the function-calling layer, the skills directory, and any supporting code that has already been tested and cleaned up. Instead of rewriting logic inside the prompt every time, you ship the logic as working software and let the model call it.

Add a container around all of this and the unit becomes portable. Send it to someone. Install it. No API key. No dependency on a cloud service. No "connect your data to our platform" dance. The thing arrives with its own brain, its own procedures, and its own toolbox.

That is why "AI in a box" understates the idea. A box suggests static content. This is closer to a deployable worker. It can read documents, use embedded utilities, follow local instructions, and perform bounded tasks immediately after installation. The model is not reaching out for identity, capabilities, or permission from a distant vendor every five seconds. It already carries what it needs.

There is a strong practical appeal here, especially in places where cloud AI is awkward or impossible. Law firms do not want client files wandering through third-party APIs. Manufacturers have machines on networks that barely trust their own printers. Hospitals live inside compliance constraints that turn every integration into a miniature opera. A portable AI pod changes the conversation. You can inspect it, run it on isolated hardware, and keep the whole workflow inside your environment.

This is also a distribution story. Software bundles traditionally deliver features. These pods would deliver expertise. That is a different product category.

Function calling is the hinge

For this to work well, local models need to be good at a very specific skill: calling functions reliably.

A lot of AI disappointment comes from asking a general model to improvise like a polymath and execute like a deterministic tool router. Those are related abilities, but they are not identical. You can have a model that writes lovely explanations and still makes a mess of tool use. It picks the wrong function, fills parameters sloppily, or hesitates when the schema gets complex. In production, that kind of failure is not charming. It is expensive.

This is why specialized function-calling models are worth watching. Google's FunctionGemma, released in December 2025, is a clear example of the direction. The interesting part is not just that it handles tool invocation well out of the box. It is that a model like this can be fine-tuned for very specific function signatures and usage patterns. If your pod depends on a set of internal tools for document extraction, ERP queries, CAD validation, or lab workflow automation, you can train the model to become unusually competent at exactly those calls.

That changes the architecture. You no longer need a giant general-purpose model inside every portable worker. In many cases, you want a compact, efficient local model that is excellent at dispatch: read the request, interpret the context, select the right function, fill the arguments, inspect the result, and continue. The heavier reasoning can be partially encoded into the skill instructions, the reference materials, and the code itself.

This is one of the least glamorous truths in AI, and one of the most useful: a lot of valuable work is structured. It is not mystical cognition. It is routing, validation, lookup, extraction, formatting, checking, and decision support with clear constraints. A model that is excellent at function calling can feel far more capable than a larger but fuzzier model, because it spends less time improvising and more time driving real machinery.

That is also what makes the pod idea scalable. You could imagine a large catalog of portable specialists built on the same underlying function-calling model. The differentiation would live in the skills, the tools, the fine-tuning, and the reference corpus. One package becomes a contract analysis assistant. Another becomes a compliance checker for procurement workflows. Another helps a biotech team triage papers against internal experimental protocols. Same foundation, different embodied expertise.

The real product is a working specialist

If this sounds like SaaS with extra steps, it helps to look at what SaaS actually sells. In most cases, it sells access to centralized software plus a business model wrapped around recurring dependency. You do not just buy a capability. You rent proximity to the vendor's infrastructure.

A portable AI worker flips that. The vendor can sell you a self-contained specialist, or license it for a fixed purpose, or let a consultant package their own expertise into something transferable. Instead of paying indefinitely for a broad platform you half use, you might acquire a narrow but potent worker for a precise job.

Picture a company that handles public tenders across five countries. Today it probably uses a pile of templates, a knowledge base nobody loves, and a few overworked experts who remember how each process actually works. A portable AI bundle could contain the model, the tender rules, prior examples, local legal references, extraction functions, risk scoring logic, and document assembly tools. Install it on secure machines and it becomes a repeatable capability the organization can keep.

Or think about the long tail of professional services. There are thousands of workflows where the value is not a gigantic application. It is a disciplined method. Review this class of documents. Apply this decision tree. Pull these values from these systems. Flag anomalies according to this internal policy. Produce a draft in the house format. That is exactly where a packaged AI worker shines. It turns tacit know-how into a movable artifact.

The labor angle is uncomfortable and important. A lot of freelance work is really workflow packaging hidden inside a person. Clients pay for judgment, yes, but they also pay for repeatable procedure, memory, and access to tools. Once those ingredients can be encoded into a portable bundle, some services become easier to productize. That will not erase human experts. It will change what clients expect to receive from them. Selling pure hours gets weaker when expertise can be shipped.

There is another consequence, and it is more positive than the usual automation script. Ownership becomes legible. If a team depends on an AI capability, they can keep a local copy, version it, archive it, compare it, and run it without asking a platform for permission. That is a healthier default than pretending crucial workflows should live as opaque prompts inside somebody else's web app.

The messy parts do not disappear

This vision needs better hardware and leaner models than many people have today. Local inference is improving fast, but there is still a gap between what sounds elegant in a diagram and what feels pleasant on a mid-range laptop. A portable worker that takes forty seconds to think and melts your battery is not winning many hearts.

Security is also complicated in a new way. An offline bundle sounds safe because it does not call the cloud, but a self-contained package can still be malicious, incompetent, or stale. If it includes code, tools, and model behavior, you need a way to inspect what it can do. You need signatures, provenance, permission boundaries, and update policies. Otherwise the "freelancer in a folder" starts to resemble the oldest genre of computing mistake: running random stuff you found online because it promised to save time.

Maintenance matters too. A consultant can adapt when a regulation changes or an internal process shifts. A portable AI worker has to be updated deliberately. If its legal references are six months old or its tool schemas drifted from reality, confidence becomes dangerous. There will be a whole discipline around maintaining these bundles, testing them, and proving they still deserve trust.

And there is a social limit worth naming. Some work looks structured until you get close. Then you discover that the valuable part is not procedure but politics, persuasion, taste, or accountability. A portable AI worker can assist in those zones, sometimes powerfully, but it does not become the responsible party. That distinction will keep mattering even as the bundles get better.

Expertise starts to look like a file format

The most interesting thing about these pods is not that they are local. It is that they treat expertise as something that can be assembled, audited, and transferred with unusual precision.

For years, software has trained us to think in terms of applications and services. AI is nudging us toward a different unit. A capability can live inside a package that includes the model, the instructions, the references, the tools, and the working code. You can pass it from one machine to another. You can keep it private. You can inspect its contents. You can tailor it for one job and make it very good at that job.

When that packaging becomes normal, a surprising amount of software starts looking overbuilt. Many people do not need a sprawling platform. They need a competent specialist they can install, trust, and keep close to the work.

End of entry.

Published April 2026