20 People. $100M Revenue. The 5 Operations Behind Every Tiny Team Beating a Giant One.

Full Transcript — Download SRT & Markdown

00:00
Speaker A
Every workforce skill in history so far has had a finish line, a point where it was done.
00:06
Speaker A
AI doesn't, picture a bubble, the air inside is everything AI agents can do reliably today.
00:12
Speaker A
The air outside is everything that still requires a person.
00:15
Speaker A
The surface of that bubble, that thin curved membrane, that is where the interesting work is happening right now.
00:22
Speaker A
It's where you decide what to delegate and what to keep, how to verify agent output, where to intervene, how to structure that handoff, working on that surface well is the most valuable professional capability in the economy today.
00:36
Speaker A
But here's the thing, that bubble is inflating.
00:40
Speaker A
Every model release, every capability jump, every quarterly leap in reasoning or context or tool use, the bubble keeps getting bigger, tasks that sat on the surface migrate inside the bubble where agents are and the boundary continues to shift out.
00:53
Speaker A
And a person who learned to work on the surface of the November bubble is now standing inside it doing work that an agent handles better than she does, running verification checks against failure modes that don't exist for AI models anymore.
01:06
Speaker A
But notice what else happens.
01:10
Speaker A
But that's the simple part of the story.
01:13
Speaker A
Everyone's talking about that.
01:15
Speaker A
Let's go beyond it.
01:16
Speaker A
Notice what happens when a bubble expands.
01:20
Speaker A
The surface area increases.
01:24
Speaker A
The frontier doesn't shrink as AI gets more capable.
01:28
Speaker A
It actually grows.
01:32
Speaker A
There's more boundary to operate at, more places for human judgment, not less.
01:36
Speaker A
More seams between human and agent work, not less.
01:40
Speaker A
More judgment calls about what crosses that AI agent membrane and what doesn't.
01:45
Speaker A
More verification challenges at this new edge, more decisions about where human attention creates value that it didn't need to create before.
01:52
Speaker A
Every prior workforce skill, whether you're talking about literacy or numeracy or computer literacy or coding, was a destination.
02:00
Speaker A
You reached it, you got it, you're done.
02:03
Speaker A
The target doesn't move.
02:05
Speaker A
But the skill of working at the surface of this bubble in AI has no fixed destination because the surface is always expanding outward.
02:12
Speaker A
You can't learn it once.
02:14
Speaker A
You can learn to stay on it, to move with it as it expands.
02:20
Speaker A
To maintain your footing as the curvature shifts beneath you.
02:23
Speaker A
But that's a fundamentally different kind of skill than any workforce development system has ever been asked to produce.
02:30
Speaker A
We are trying to teach this expanding surface skill set mostly with fixed destination methods.
02:38
Speaker A
Every curriculum, every certification, every training program has assumed the target stands still.
02:44
Speaker A
This one doesn't, and the mismatch between the skill the economy needs and the infrastructure we've built is the most expensive gap I've seen in the global workforce.
02:52
Speaker A
And I'm giving that skill a name.
02:55
Speaker A
It's called frontier operations.
02:57
Speaker A
The surface of that bubble is the frontier.
03:00
Speaker A
Working on it, sensing where it sits, structuring handoffs between AI and human across it, maintaining a model of how agents fail at the current edge.
03:10
Speaker A
Forecasting where the surface expands next.
03:15
Speaker A
Deciding where your attention creates the most value.
03:20
Speaker A
That is the frontier operation.
03:22
Speaker A
That is frontier AI operations at its best.
03:25
Speaker A
This is not AI literacy.
03:27
Speaker A
That's just knowing what a language model is and how to write a prompt.
03:32
Speaker A
It's the equivalent of teaching someone the alphabet and then calling them a reader.
03:37
Speaker A
It's also not prompt engineering.
03:41
Speaker A
That's one technique inside one component of the practice.
03:46
Speaker A
It's like calling surgery scalpel handling.
03:49
Speaker A
And it's also not that vague gesture at human judgment that tends to fill a lot of keynotes about the future of work.
03:55
Speaker A
Because most people correctly identify that judgment matters, but they incorrectly assume that naming it is the same as teaching it in this fast-moving world.
04:03
Speaker A
Frontier operations is the specific, practicable, accessible version of the skill everybody is pointing to.
04:10
Speaker A
And very few people are building.
04:13
Speaker A
It does have components.
04:16
Speaker A
It does develop through practice.
04:18
Speaker A
It degrades if you don't maintain it.
04:20
Speaker A
And it's the first workforce skill in history that expires on a roughly quarterly cycle.
04:27
Speaker A
Which means everything we know about how to teach workforce skills doesn't work too well for this skill.
04:33
Speaker A
So what's inside the box?
04:36
Speaker A
I've named it as frontier operations.
04:40
Speaker A
But I'm not just going to point at it and stand there vaguely and tell you it's great.
04:44
Speaker A
There are five kinds of frontier operations.
04:50
Speaker A
If you picture the world as an expanding bubble of AI capability, which I think is very accurate.
04:56
Speaker A
There are five kinds of skills that stay persistent across the expanding surface area of that bubble.
05:02
Speaker A
The first one is what I call boundary sensing.
05:05
Speaker A
It's the ability to maintain accurate, up-to-date operational intuition.
05:11
Speaker A
About where the human agent boundaries sits for a given domain.
05:15
Speaker A
This is not static knowledge.
05:18
Speaker A
It updates with every single model release, every capability jump, every subtle shift in how agents handle long context or tool use.
05:25
Speaker A
Opus 4.5 couldn't reliably retrieve information from deep in a long document.
05:32
Speaker A
Three months later, Opus 4.6 scores 93% on retrieval at 256,000 tokens.
05:38
Speaker A
A person who calibrated his boundary sense against the November model and hasn't updated that is now either overtrusting or underusing the February model.
05:49
Speaker A
And both kinds of errors are very expensive.
05:52
Speaker A
So the skill is maintaining the calibration, not having it once.
05:57
Speaker A
So what does this look like in practice?
06:00
Speaker A
For a product manager, good boundary sensing looks like letting an agent draft a credible competitive analysis.
06:09
Speaker A
But realizing that that same agent will miss the political dynamics between two executives that it's never observed or been given context on.
06:18
Speaker A
And so the product manager doesn't hand the whole analysis to the agent.
06:24
Speaker A
They hand the agent the market sizing and the feature comparison piece.
06:29
Speaker A
And those are tasks that are now safely inside the AI bubble.
06:33
Speaker A
And then the product manager reserves the stakeholder dynamic section for themselves.
06:38
Speaker A
And now last quarter, that bubble was smaller and maybe the feature comparison was updated at least halfway manually.
06:44
Speaker A
Not anymore.
06:46
Speaker A
It's updated.
06:47
Speaker A
Let's say you're a marketing director, that might look like an agent producing a solid first draft campaign copy and AB test headline variants.
06:54
Speaker A
But maybe brand voice copy drifts subtly off tone after the third or fourth version.
07:00
Speaker A
Well, you use an agent for ideation, you use an agent for first drafts, you edit the voice yourself.
07:06
Speaker A
And you don't ask for it to iterate past version two.
07:09
Speaker A
What bad would look like here is the marketing director trusting everything and getting burned by hallucinations.
07:16
Speaker A
Or trusting nothing and doing everything manually.
07:20
Speaker A
Most commonly, the marketing director probably calibrated six months ago and hasn't noticed the boundary move.
07:28
Speaker A
The marketing director doesn't have boundary sensing as a skill set.
07:31
Speaker A
Here's the second skill set.
07:33
Speaker A
Seam design.
07:34
Speaker A
It's the ability to structure work so that transitions between human and agent phases are clean.
07:40
Speaker A
Verifiable and recoverable.
07:42
Speaker A
This is very much an architectural skill.
07:45
Speaker A
It's closer to how a good engineering manager thinks about system boundaries than to how an individual contributor thinks about their tasks.
07:52
Speaker A
The person doing seam design is asking, if I break this project into seven different phases.
08:00
Speaker A
Which three are fully agent executable?
08:05
Speaker A
Which two need human in the loop?
08:08
Speaker A
And which two are still irreducibly human?
08:11
Speaker A
What artifacts pass between each of these phases?
08:14
Speaker A
What do I need to see at each transition to know that things are on track?
08:18
Speaker A
The reason this is a very distinct skill and not just project management in a trench coat is that the answer changes as capabilities shift.
08:26
Speaker A
So the seam that was in the right place last quarter is in the wrong place this quarter.
08:32
Speaker A
And so the skill isn't in the design one off.
08:38
Speaker A
It's in the ability to redesign as agent capabilities.
08:42
Speaker A
So this might look like a software engineering lead at a company structuring the handoff so that ticket triage and work routing go to the agent.
08:50
Speaker A
Architectural decisions stay with humans.
08:55
Speaker A
And the boundary between those two is then defined by specific artifacts.
09:01
Speaker A
The content of the ticket, the structure of the code base.
09:05
Speaker A
The org chart, and then you have specific verification checks at that scene that ensure the handoff is clean.
09:10
Speaker A
Without those, you either go end to end with agent runs and you don't have the verification infrastructure in place yet to do that reliably.
09:19
Speaker A
Or you have humans manually reviewing things the agent now handles better than they do.
09:25
Speaker A
Another one might be a consulting engagement manager.
09:30
Speaker A
Let's say they break the strategy project into research, which is agent led with a human defined scope, and synthesis, which is human led with agent generated first pass frameworks, and a client presentation, which is human led with agent generated slide drafts.
09:44
Speaker A
The seam between research and synthesis is a structured deliverable.
09:50
Speaker A
It's a fact base with source citations that human can spot check in just a few minutes.
09:56
Speaker A
Months ago, that seam would have included manual fact verification on every data point, but the agent citation accuracy has improved dramatically, and so it made sense for the consulting engagement manager to move the scene.
10:06
Speaker A
The third skill is failure model maintenance.
10:08
Speaker A
It's the ability to maintain an accurate, current mental model of how agents fail.
10:15
Speaker A
Not that they fail, but the specific texture and shape of failure at the current capability level.
10:22
Speaker A
That's really important.
10:24
Speaker A
Early language models failed very obviously, right?
10:30
Speaker A
They had garbled text, they might have wrong facts.
10:33
Speaker A
They might have incoherent reasoning.
10:35
Speaker A
Current frontier models fail in very subtle ways, they have correct sounding analysis built on a misunderstood premise.
10:42
Speaker A
Or plausible code that handles the happy path and breaks on edge cases.
10:47
Speaker A
Or research summaries that are 98% accurate while the remaining 2% are confidently fabricated in a way that's very difficult to distinguish from the accurate parts unless you know the domain well.
10:56
Speaker A
So the skill here is not be skeptical of AI output.
11:01
Speaker A
That's necessary, but not particularly useful.
11:05
Speaker A
It's like saying the skill of surgery is to be careful.
11:08
Speaker A
The skill is actually maintaining a differentiated failure model, knowing that for task type A, the agent's failure mode is X, and here's how to check for it well in your domain.
11:19
Speaker A
While for task type B, the failure mode is Y, and and there's a different check for it.
11:24
Speaker A
This might look like a corporate counsel that knows an agent reviewing contracts is going to catch boilerplate issues, but miss indemnification clauses.
11:33
Speaker A
Or maybe miss non-standard termination language and maybe miss the interaction between a specific liability cap and section seven and a carve out varied in the exhibit.
11:40
Speaker A
So the failure model right now might say, trust the boilerplate scan, manually review cross references between liability provisions and exhibits.
11:49
Speaker A
And that's that's a very different check than read the whole thing again.
11:53
Speaker A
And it takes much less time and it allows you to be much more precise in your work and much more efficient.
11:57
Speaker A
A data scientist might know that an agent generating Python for data analysis will handle Pandas transformations and standard statistical tests very reliably.
12:07
Speaker A
But produce somewhat plausible sounding nonsense when the data has messy edge cases, like mixed data formats or implicit nulls or columns that changed meanings mid data set.
12:16
Speaker A
And so the failure model says, verify the data cleaning steps and the assumptions about column semantics.
12:23
Speaker A
And then trust a lot of the downstream analysis if the cleaning is correct.
12:28
Speaker A
And what does bad look like here?
12:30
Speaker A
Bad would look like applying the same generic skepticism to everything, which is very inefficient.
12:38
Speaker A
Right, it assumes memorized failure patterns from six months ago hold when that's actually incorrect.
12:45
Speaker A
Here's the fourth skill.
12:47
Speaker A
Capability forecasting.
12:49
Speaker A
It's the ability to make very reasonable short-term predictions about where that bubble boundary for AI will move next.
12:56
Speaker A
And to invest learning and workflow development accordingly.
12:59
Speaker A
This is not about predicting the future of AI.
13:02
Speaker A
It's actually something that nobody could do super reliably over long horizons.
13:06
Speaker A
It's about reading the trajectory well enough to make very sensible six to 12 month bets about what is likely to become agent territory.
13:15
Speaker A
Think of it like reading the swells on the ocean.
13:21
Speaker A
A surfer doesn't predict exactly what the next wave will look like, but a good surfer will read the sea.
13:29
Speaker A
Understand how the floor shapes wave at this particular break.
13:33
Speaker A
And position themselves where the next rideable wave is most likely to form.
13:37
Speaker A
And that skill is very much probabilistic positioning.
13:40
Speaker A
Not linear prediction.
13:41
Speaker A
So a person with good capability forecasting in early 2025 could look at the trajectory of coding agents, 30 minutes of sustained autonomy.
13:50
Speaker A
And then seeing how it scales from there and start investing more in code review and specification skills rather than just raw coding.
13:58
Speaker A
Meanwhile, a UX researcher watching agents get better at survey design and qualitative coding can start investing in interpretive synthesis, the skill of turning coded data into product insights that will shift a road map.
14:09
Speaker A
So the coding is migrating inside the bubble.
14:14
Speaker A
The so what of the coding is where the new surface happens to be.
14:18
Speaker A
So what does bad look like here?
14:20
Speaker A
It can look like chasing every new tool, which is exhausting and doesn't really generate compound returns.
14:26
Speaker A
Or it can look like ignoring developments until you're forced to catch up.
14:30
Speaker A
Most commonly, it tends to look like investing heavily in learning a particular platform, and then that platform's advantage might evaporate when the next model ships.
14:40
Speaker A
And either deletes that part of the workflow or provides leverage that that platform can't deliver on its own.
14:45
Speaker A
We've seen this over and over and over again with AI.
14:48
Speaker A
As large language models eat up more and more and more of the capability space.
14:53
Speaker A
And skill number five is leverage calibration.
14:55
Speaker A
It's the ability to make very high quality decisions about where to spend human attention, which is now the scarcest resource in an agent rich environment.
15:03
Speaker A
As agent capabilities continue to increase, the bottleneck shifts from getting things done to knowing what things are worth a human's attention.
15:10
Speaker A
Even McKinsey has a framework that describes two to five humans supervising 50 to 100 agents running an end-to-end process.
15:17
Speaker A
And that's not just McKinsey, that's a lot of different agentic patterns coalescing across the industry.
15:22
Speaker A
That roughly 10 to one ratio makes the math of attention very, very clear.
15:29
Speaker A
If you have 100 streams of agent output and eight hours a day, you cannot review everything at the same depth.
15:35
Speaker A
The skill is going to be triaging your own attention in real time.
15:39
Speaker A
So this might look like an engineering manager overseeing agent assisted development across five teams, and that person develops a hierarchical attention allocation.
15:50
Speaker A
Most agent generated code flows through automated test suites and gets linted on its own.
15:55
Speaker A
A smaller subset, like billing or data pipelines, might get flagged for human code review.
16:00
Speaker A
Only architectural decisions and cross-system changes might get flagged out for deep human engagement.
16:06
Speaker A
And so a good manager would recalibrate those thresholds monthly because the agents keep getting better at the routine tier.
16:14
Speaker A
And you keep finding new categories that belong in that middle in between like human touch tier.
16:19
Speaker A
Ahead of customer success overseeing agent handled support tickets.
16:22
Speaker A
Might review escalations and a random sample of resolved tickets to ensure quality.
16:27
Speaker A
But she wouldn't review the resolution of routine password resets.
16:30
Speaker A
She does review every ticket where the agent access account modification tools.
16:35
Speaker A
And that threshold ends up being calibrated to risk and adjusted as the agent's ability to use tools on her particular ticket system get better and better.
16:42
Speaker A
Now, what might bad look like here?
16:44
Speaker A
It might look like reviewing everything at the same depth, which creates a bottleneck.
16:49
Speaker A
Which creates human burnout.
16:51
Speaker A
Or it might look like reviewing nothing.
16:55
Speaker A
Which you would only do if you were intentionally piloting a dark factory floor scenario.
17:00
Speaker A
And that is something that very, very few teams are ready for.
17:05
Speaker A
Even if technically speaking it's possible today.
17:07
Speaker A
These are not a checklist.
17:09
Speaker A
These are five operations that are simultaneous, integrated and continuous.
17:14
Speaker A
It's a it's a way of practice, like driving involves steering and speed management and route awareness and hazard perception, all at the same time.
17:27
Speaker A
At any given moment, a person operating at the frontier is sensing the current boundary, they're designing seams around it, they're verifying against an updated failure model.
17:39
Speaker A
And they're making bets about where that boundary is going to move.
17:43
Speaker A
And allocating attention across that system.
17:45
Speaker A
The integration is what makes this a practice and not a curriculum.
17:50
Speaker A
You can teach each component in isolation, but a person who's good at all five individually.
17:56
Speaker A
And doesn't run them simultaneously, still isn't operating at the frontier.
18:02
Speaker A
Because they're thinking about putting those skills into practice.
18:07
Speaker A
Instead of doing them together seamlessly.
18:10
Speaker A
As a way of business.
18:12
Speaker A
As the way of working.
18:15
Speaker A
As the future of AI human collaboration.
18:18
Speaker A
This skill gap is structural.
18:21
Speaker A
Every other AI adjacent skill might eventually get absorbed into the technology itself.
18:26
Speaker A
Prompting techniques are starting to get baked into system defaults and prompting is moving up a level into things like intent engineering.
18:33
Speaker A
Integration patterns are getting productized.
18:36
Speaker A
But frontier AI operations, you can't really automate them because by definition, there's a surface of AI capability.
18:45
Speaker A
And so when a task migrates inside the AI bubble, the surface just expands outward.
18:51
Speaker A
And the person who operates at the surface moves with it.
18:55
Speaker A
The skill is sort of structurally resistant to its own obsolescence.
19:00
Speaker A
I will also call out that the structural gap compounds.
19:05
Speaker A
A person who develops this skill set six months sooner than their peers, they don't just have a six-month head start.
19:13
Speaker A
They have six months of updated calibration that the peer doesn't have.
19:18
Speaker A
And because capabilities are accelerating, the distance between calibrated and uncalibrated keeps getting wider with every model release.
19:24
Speaker A
And so the person whose boundary sense was current in February and the person whose boundary sense was current last August are operating worlds apart.
19:32
Speaker A
This is really the mechanism behind those gigantic leverage numbers that keep appearing in production deployments.
19:40
Speaker A
When you see, you know, cursor hitting stunning numbers with a small number of people from a revenue perspective, lovable doing the same thing, when you see the Anthropic team shipping every time I turn around.
19:52
Speaker A
The gap between shipping and revenue at AI native companies and shipping and revenue at traditional SAS companies isn't really explained by better tools.
20:02
Speaker A
It's explained by people who have developed the operational practice to be on the bubble.
20:10
Speaker A
And convert those tools into reliable output as AI continues to evolve.
20:14
Speaker A
I believe that this skill set is the single largest determinant of not only which businesses tend to succeed over the next decade.
20:23
Speaker A
But which economy starts to win over the next decade.
20:27
Speaker A
It's not going to just be about which economies built models.
20:32
Speaker A
Because models are portable over the internet.
20:34
Speaker A
It's not just going to be about which economies have compute.
20:37
Speaker A
Because compute can be rented.
20:39
Speaker A
It's going to be about which economies can field workers who are excellent at operating at the AI human frontier.
20:47
Speaker A
The models are starting to commoditize, the compute is being built out everywhere.
20:53
Speaker A
What is going to continue to be scarce is the human capacity to convert those inputs into economic output.
21:00
Speaker A
And that conversation is going to be something that we need to have a lot more of.
21:07
Speaker A
As compute starts to become more abundant.
21:10
Speaker A
So you might wonder, if you're trying to foster this skill set.
21:13
Speaker A
What does that actually look like?
21:15
Speaker A
Well, one, as a leader, you should be building practice environments, not coursework.
21:21
Speaker A
Just like flight simulation helps you learn to fly.
21:26
Speaker A
Simulating AI environments in sandboxes is important for AI era work.
21:30
Speaker A
This is much, much more important.
21:33
Speaker A
Practice environments are where agents have different capability levels, where failure modes are realistic.
21:40
Speaker A
Where rules can change so practitioners have to recalibrate.
21:43
Speaker A
This is much more practical.
21:46
Speaker A
Than just looking at a bunch of slides and saying you did an AI workshop.
21:50
Speaker A
You have to get in and touch the AI often in order to get skilled.
21:54
Speaker A
Measure calibration, not knowledge, is the second thing I would call out for leaders.
21:58
Speaker A
The right assessment is not, you know, can you write a good prompt?
22:03
Speaker A
It is given a task and an agent at capability level X, can you accurately predict where the agent is going to succeed?
22:11
Speaker A
Where it's going to fail?
22:14
Speaker A
And how to structure your work accordingly?
22:16
Speaker A
That is a much harder thing to do than writing a good prompt.
22:20
Speaker A
But it's a much, much more durable skill because it measures the ability of humans to work with AI as AI continues to scale.
22:27
Speaker A
Third, maximize feedback density, not training hours.
22:30
Speaker A
The speed of skill development is really a function of how many cycles a person gets through with AI per unit of time.
22:38
Speaker A
And it's not really a linear function of hours of training.
22:42
Speaker A
So a person that completes a 40-hour AI course offsite and then returns to the workforce, but never really touches an AI tool beyond some light chat GPT.
22:50
Speaker A
Has zero calibration cycles.
22:52
Speaker A
A person who skips that course and then delegates 10 real tasks a day to an agent and evaluates the output.
22:59
Speaker A
Is going to have 100 of them in 10 days.
23:01
Speaker A
Make sure that you understand how AI exposure is actually shaping feedback density for your team.
23:07
Speaker A
Fourth, create explicit roles for frontier operations.
23:10
Speaker A
The skill doesn't develop if it's just an undifferentiated part of somebody else's job.
23:15
Speaker A
Organizations need people whose specific function is to operate at the boundary, to maintain the failure models.
23:22
Speaker A
To update verification protocols, to redesign workflows when capabilities shift.
23:26
Speaker A
You can call them AI automation leads.
23:30
Speaker A
You can call them delegation architects.
23:32
Speaker A
You can call them frontier engineers.
23:34
Speaker A
The title matters less than recognizing that this process of evolving the automation frontier is very high leverage.
23:40
Speaker A
And it's a distinct specialty and it's something that you need someone to focus on.
23:45
Speaker A
So that you can aggressively socialize changes through the business.
23:49
Speaker A
A lot of the org chart built for the pre-agent era assumes that output scales with head count.
23:55
Speaker A
More people equals more output.
23:57
Speaker A
But with frontier operations, you're inverting this.
24:03
Speaker A
Output scales with leverage and leverage scales with how well a small number of humans operate at that boundary.
24:09
Speaker A
So the organizational unit that matters is like a tiny pod.
24:12
Speaker A
And I've seen two structures emerging here that are significant.
24:18
Speaker A
Right, the first is a team of one, a single person with a very strong frontier operation skill set.
24:25
Speaker A
Who runs multiple agent workflows across a domain.
24:29
Speaker A
That person does the boundary sensing, defines the seams, maintains the failure models.
24:35
Speaker A
And calibrates attention allocation for the pod.
24:37
Speaker A
And others are executing, don't worry, with lots of AI within those structures.
24:44
Speaker A
And developing their own frontier intuition as AI models evolve through practice.
24:50
Speaker A
This is a little bit like how a surgical team operates.
24:53
Speaker A
One lead that sees the whole field, others separated by complementary skill boundaries.
25:00
Speaker A
Executing in roles that mesh well together.
25:03
Speaker A
So in product development, this might look like one frontier operator who owns the human agent workflow across the product surface.
25:10
Speaker A
A couple of engineers executing heavy agent assisted development.
25:15
Speaker A
Maybe a designer running agent assisted prototyping and user research and also committing code.
25:21
Speaker A
And maybe a data scientist managing the analytics pipeline.
25:24
Speaker A
They ship at the pace of a 20-person team because the operator keeps the seams really current.
25:30
Speaker A
And the failure modes calibrated.
25:33
Speaker A
And oh, by the way, the operator is shipping too.
25:35
Speaker A
And the rest of the pod has enough frontier skill to execute without supervision.
25:39
Speaker A
So you might wonder, how does this ladder up?
25:41
Speaker A
I think the simplest way to think about this is to think about where you want to double down on your bets.
25:47
Speaker A
If teams of one are able to effectively discover opportunities for deeper engagement, and teams of five are able to build meaningful products.
25:57
Speaker A
A team that ladders up from there is really focused on one of two areas.
26:02
Speaker A
Either you have a portfolio of bets you're managing as a leader and you're farming that out across four or five teams of five.
26:13
Speaker A
Or you're in a position where you're looking for one big win and you pick something.
26:20
Speaker A
From a team of five or a team of one that's doing exploratory work in your area, and then you rally your whole team to double down on that big bet and produce something really polished.
26:29
Speaker A
And and how you allocate that depends a lot on your business strategy and product strategy.
26:34
Speaker A
Which has to devolve much farther down from executive leadership than it used to.
26:38
Speaker A
You need people who are managing four or five teams of five to be just as strategically informed as your CEO at this point.
26:44
Speaker A
So what do you look for when you're hiring for this skill set?
26:46
Speaker A
It's not necessarily a case for traditional hiring signals, right?
26:52
Speaker A
Credentials, years of experience, tool proficiency.
26:56
Speaker A
Those aren't necessarily helpful.
26:58
Speaker A
Instead, you want to look for, does this person track where agents succeed and fail in her domain?
27:03
Speaker A
Can she articulate specifically what an agent handles today and where it doesn't handle well?
27:08
Speaker A
Can she describe a new capability or immediately start redesigning a workflow?
27:14
Speaker A
Or does something just get filed under interesting and never get action?
27:17
Speaker A
Does this person have a failure model?
27:20
Speaker A
Not generic skepticism, but a differentiated understanding of how agents fail on which tasks.
27:26
Speaker A
Is there a reliable trend of forecasting this person has maintained where you can see them?
27:32
Speaker A
They've got really good instincts about where the future is going.
27:35
Speaker A
The person who can answer these questions well at high quality.
27:38
Speaker A
That's your frontier operator.
27:39
Speaker A
The person who answers them with, well, I'm good at prompting.
27:43
Speaker A
That's not your frontier operator.
27:45
Speaker A
Let's say you want to get better at this skill.
27:48
Speaker A
What does that mean for you?
27:50
Speaker A
If you're an individual contributor, you need to start tracking where your boundary sense is incorrect.
27:56
Speaker A
You need to track where agents surprise you.
27:59
Speaker A
The surprise is a signal.
28:01
Speaker A
You want to be collecting your surprises on purpose, logging them and starting to build your professional instincts.
28:10
Speaker A
So that you start to understand better where agents operate, where agents work, where agents don't work.
28:16
Speaker A
And if your agent hasn't surprised you recently.
28:20
Speaker A
Then you are definitely not operating at the boundary.
28:23
Speaker A
If you manage people, you want to look at how your team allocates attention across agent assisted work.
28:29
Speaker A
Are they reviewing everything at the same depth?
28:34
Speaker A
Is there a bottleneck that's masquerading as due diligence?
28:36
Speaker A
Are they reviewing nothing?
28:37
Speaker A
The right answer is going to be differentiated based on your domain.
28:44
Speaker A
But it's up to you to figure out where human attention ought to be allocated here.
28:47
Speaker A
And if your team can't articulate to you their philosophy of human attention.
28:52
Speaker A
You've got a problem.
28:53
Speaker A
If you run an organization, the question isn't, are we using AI?
28:58
Speaker A
It's, do we have people whose job it is to know where the evolving AI agent human boundary is and how we think about redesigning our workflows as it shifts?
29:08
Speaker A
If the answer is you can't name someone.
29:10
Speaker A
You are leaving one of the most consequential organizational capability decisions.
29:16
Speaker A
Of the decade to chance.
29:18
Speaker A
And I wouldn't do that.
29:19
Speaker A
We have seen context windows.
29:21
Speaker A
We have seen retrieval.
29:22
Speaker A
It is hard to overstate how much models have gained in capability between November of 2025 and February of 2026.
29:30
Speaker A
So within the last 60 to 90 days.
29:32
Speaker A
Anyone who is deep in AI has felt the difference.
29:36
Speaker A
And you feel it whether you're touching Opus 4.6 versus 4.5.
29:40
Speaker A
You feel it touching Codex 5.3.
29:42
Speaker A
You feel it touching Gemini 3.1 Pro.
29:44
Speaker A
And if you are listening to this and you can't feel it, if you're like, I don't know what the difference is.
29:50
Speaker A
That means you're not at the edge of the bubble.
29:54
Speaker A
And that was just one quarter, the next quarter, there's going to be another shift.
30:00
Speaker A
Everything we are seeing out of the labs is indicating that we are not slowing down.
30:04
Speaker A
And so if you are not in a place where these February agents are surprising you.
30:10
Speaker A
The best thing you can do to welcome yourself to the frontier is to find a way to give your agents a job that surprises you.
30:17
Speaker A
Whether they fail, whether they partly succeed.
30:22
Speaker A
Give them something that allows them to surprise you.
30:26
Speaker A
Because if you don't, you run the risk of missing this expanding capability bubble.
30:30
Speaker A
And this is the new workforce skill set that will define all of our career success for the next decade.

Get More with the Söz AI App

Transcribe recordings, audio files, and YouTube videos — with AI summaries, speaker detection, and unlimited transcriptions.

Or transcribe another YouTube video here →