
Each quarter, Every tells its writers to put down their pencils, stop scribbling, and think. We call these our Think Weeks, and they’re phenomenal opportunities to take a break from the daily hustle, step back, and look at the evolving tech and business landscape. Ahead of OpenAI’s DevDay taking place on October 1, we thought this was a perfect time to republish some of our sharpest writing on ChatGPT. First up, Evan Armstrong’s look-ahead about AI agents, which feel closer to reality than ever.—Kate Lee
Was this newsletter forwarded to you? Sign up to get it in your inbox.
Most races have a prize pool. The New York City marathon winner gets $100,000. The F1 winner in 2023 took home a $140 million pot.
The winner of the race that I’m going to describe will earn billions. Maybe tens of billions. They’ll bend the arc of the universe. They’ll materially increase the GDP of entire countries.
This is the race toward the AI agent. Agents are the next step in the AI race and the focus of every major tech giant, AI startup, and research lab.
I’ve spent months talking with founders, investors, and scientists, trying to understand this technology. Today, I’m going to share my findings. I’ll cover:
- What an AI agent is
- The major players
- The technical bets
- The future
Let’s get into it.
What is an agentic workflow?
An AI agent is a type of model architecture that enables a new kind of workflow.
The AI we started with formulates an answer and returns it to the user. Ask it something simple, like “Does an umbrella block the rain?” and OpenAI’s GPT-4 returns the answer, “Of course it does, you dumbass.” The large language model is able to answer the question without relying on external data because it uses internal data and executes on the prompt without making a plan. It's a straightforward line connecting input and output. And every time you want a new output, you have to provide a fresh prompt.
Agentic workflows are loops—they can run many times in a row without needing a human involved for each step in the task. Under this regime, a language model will make a plan based on your prompt, utilize tools like a web browser to execute on that plan, ask itself if that answer is right, and close the loop by getting back to you with that answer. If you ask, “What is the weather in Boston for the next seven days, and will I need to pack an umbrella?” the agentic workflow would form a plan, use a web browsing tool to check the weather, and use its existing corpus of knowledge to know that, if it is raining, you need an umbrella. Then, it would check if its answers are right and, finally, say, “It’ll be raining (like it always does in Boston, you dumbass) so, yes, pack an umbrella.”
What makes an agentic workflow so powerful is that because there are multiple steps to accomplish the task, you can optimize each step to be more performative. Perhaps it is faster and/or cheaper to have one model do the planning, while smaller, more specialized models do each sub-task contained within the plan—or maybe you can build specialized tools to incorporate into the workflow. You get the idea.
But agentic workflows are an architecture, not a product. It gets even more complicated when you incorporate agents into products that customers will buy.
Solving users problems > flashy demos
The only thing that matters in startups is solving customers’ problems. Agentic workflows are only useful as a product if they solve problems better than existing models. The tricky thing is that no one knows how to make AI agents a consistently better product right now. The pieces are all there, but no one has figured out how to put them all together.
This moment is strongly reminiscent of the early 1980s of personal computing, when Apple, Hewlett-Packard, and IBM were duking it out. They all had similar ideas about user interface (the use of a mouse, the need to display applications, etc.), but the details of implementation were closely guarded secrets. These companies competed on the quality of their technical components and how each of those components fit together to solve customer problems.
Companies that make AI agents are competing both on individual component quality and how these components are combined. In broad strokes, think of these arenas of competitive intensity scattered across five components:
Every illustration.- Data inputs: The agent needs access to unique data sets or be better able to parse public data sets (such as scraping the web). Where does the agent draw its data from? Can it access internal data repositories—note-taking systems for individuals or corporate knowledge bases for enterprise use cases—to make answers better?
- Models: For the past year, when you heard “AI,” it typically meant this component—the LLMs like GPT-4. Within the model companies like OpenAI, there are a variety of approaches that I’ll cover in a minute.
- Tools: Think of these like giving the handyman (an LLM) a new set of screwdrivers. This is an area I’m excited about. In 2023, I used a tool from OpenAI called Code Interpreter that has been able to replace many finance workflows. Code Interpreter provides the LLM with a coding environment that allows the LLM to modify spreadsheets.
- Interface: Knowing how to integrate these capabilities into a user’s workflow is just as important as—if not more than—what the agent can actually do. Is the agent nestled within a typical LLM chatbot? Is it operating behind the scenes as part of an application's code? Does the AI need to be in its own separate UI and app? Or should it be integrated into an existing workflow app like Salesforce or Excel?
- AI glue: This is my own term (you can tell because it sounds dumber than the rest), but in my interviews with founders building AI agent companies, the most common thing I heard was that “AI agents are an engineering problem, not an AI problem.” There is a sense that while each of the previous components is important, what matters is figuring out how to stick them all together. Glue is traditional, deterministic software that programs a set of logical steps.
There are infinite combinations of these components. Unlike with previous generations of software companies, investors and founders are underwriting science risk as well as product risk. In the 2000s era of SaaS, we knew the cloud worked, and we knew how to make software on it. The only question was if you could make a product that used the cloud in a way that benefited customers. With agents—both tooling and models—we haven’t completely figured out the science of making it work, let alone the product.
That is a mouthful, so let me repeat myself as simply as possible: This shit does not currently work, but investors are betting it can. Many assume we are just one or two scientific advancements away in models or tooling to make agents available at scale.
Ideas and Apps to
Thrive in the AI Age
The essential toolkit for those shaping the future
"This might be the best value you
can get from an AI subscription."
- Jay S.
Join 100,000+ leaders, builders, and innovators

Email address
Already have an account? Sign in
What is included in a subscription?
Daily insights from AI pioneers + early access to powerful AI tools
Comments
Don't have an account? Sign up!