Deep Learning with PolyAI
PolyAI's CEO/co-founder Nikola Mrkšić and team invite guests to candidly discuss trends and tech in AI, voice throughout the enterprise, and nailing the customer experience.
Deep Learning with PolyAI
What happens when anyone can build an AI agent?
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
What happens when the ability to build an AI agent is no longer limited to developers?
In this episode of Deep Learning with PolyAI, Nikola Mrkšić sits down with Michael Chen, VP of Strategic Alliances at PolyAI, to explore what's driving the explosion in AI agent development — and what it means for how enterprises build, deploy, and think about AI.
They cover the rise of personal AI agents, why Jensen Huang's $1 trillion data center bet might
Our agent development kit, which is now live, is basically a way for agents to use our platform to build agents, deploy agents, share them with humans, be part of the enterprise workflow. People talk about that whole like surface area for agents being open, being the driver of business, much in the same way that developers were the first wave of people adopting technology. Inevitable, right?
SPEAKER_01Like it just seems like the way that I see you talking about it, the way that I speak to our VPs of engineering, everyone's like their life is changed, right, by this technology.
SPEAKER_00Hello everyone, and welcome to another episode of Deep Learning with PolyAI. Uh today I've got Michael Chen, our VP of Strategic Alliances, with me. Hi, Michael. First time on the pod. I'm honored to be get the impact. This can't be. Yeah. Well, look, I think as we talked about things we have to talk about, I think two words reign supreme. And it was token maxing.
SPEAKER_01Yes. We are in the era of token maxing. So I mean the this topic started when I was MBW GTC a few weeks ago, now last month. And one of the data points that really just struck me was, you know, Jensen and his whole announcement, his keynote of he's got a trillion dollars of data center revenue through to 2027 on lock, right? And I think he was referring to just the the latest generation of GPU. So I he's sounds like he's sandbagging that number, right? And it just showed me that there is just an insatiable demand for tokens, right? We are only, it feels like we've been talking about tokens and intelligence for so long, but actually we might still just be at the very start, right? And that stuff was just like mind-blowing to me, right? Like that revenue number, the way that I'm seeing these solutions, you know, we're going from just you know back and forth FAQs to having the models in the background also do a lot of bunch of thinking, deep thinking, deep sampling. It just seems like everything is going to drive more and more tokens.
SPEAKER_00Yeah. I mean, look, I think that when I just look at the consumption of cloud code internally and then open claw, and then like the back and forths, including the whole drama with whether you can use kind of like the default like max subscription for open claw is the first time that I've actively seen like very many people at Poly get like throttled, including myself repeatedly. And there's almost like you know, you're junkie on a high and you're out of drugs. And then it's like, well, what do you do? Do you pay through the nose, or do you do you wait for that Friday, 6 p.m.? Or I'm I'm not sure if that cutoff is actually universal. But yeah, you said open cloud was a big kind of like topic at GTC as well.
SPEAKER_01Yeah, I mean, it all feeds into this, like, you know, driving more token demand, everyone having their own personal AI assistance was a big theme. Um, at GTC, they had all these tents set up everywhere around the convention center. It was NVIDIA people just inviting people to come in to get set up with their own open claw, right? It was such a massive feature of that entire entire conference, right? I don't know. We've been doing a lot of things internally, right, with Apple as well. Yeah.
SPEAKER_00Yeah, no, I've had like my, you know, I call it the the midlife crisis of a former developer, right? Where, you know, the life you could have had before you, you know, maybe went and founded a company, spent too much time in partnership talks, in sales conversations, implementations, right? Planes, and you know, you look at a life you could have had where then out of nowhere this thing comes in, and it's that lacking skill set you could have had, just better, right? And then it kind of has the different stages of decay. At first, I think I saw Sean really get into it. Now he's been through a few kind of like it's a it's a sinusoid between you go between this like complete disillusionment where you completely max out, you're tired, you work till like 4 a.m. daily, right? Then your kids wake you up at like 6, 7 a.m. and after a few days you realize that you're not as young as you used to be. But you know, you go between the disillusionment and the whole kind of like LM psychosis where you're just just another hour, just another three things you want to do. But like I've seen so many people build things that would have taken entire teams like months to set up that you know, at first, you know, you think about the security risks and everything, but later on, if you can set it up appropriately, it is very powerful. It has I don't know how it would compare like Cloud Code with OpenClaw, other than you know, there is that whole like dangerously skip permissions tag with Cloud Code, but then this open cloth thing is just a whole new level where it has this voracious appetite to go and do things on your behalf. And you know, it is very creative, but it's also really funny how you know over time it develops a certain personality and a certain set of like skill sets, especially if you look at a few people setting them up for relatively similar tasks. You see one figured something out and it's like super efficient at it. The other one is like token maxing. It found some roundabout way and it's just doing it at scale, it's spinning up sub agents. It's a real like Wild West. It's honestly I've never had more fun than in the past few weeks just because it's a whole new universe.
SPEAKER_01So how are you how are you using it then? Like how how are you actually in the day-to-day, how is that embedded in your daily workflow as CEO, founder?
SPEAKER_00Is it so I think initially there were quite a few people who were really almost like portraying themselves as these like power users and talking about how you know it writes them their briefs and their updates and their investor memos. And honestly, like if you're any good as a CEO, you probably have people doing that already. So whether it's agents or people may have like a financial impact or some speed impact, but it doesn't really profoundly change much. I think where it's been really interesting for me more recently is just like to build things that might have been deprioritized or things that are kind of like high risk, high reward things that you really want to see. Like the marginal incremental cost of pursuing one of those to get to the end of that exploration is now something that starts at 9 p.m. and ends 3, 4 a.m., right? But like you come out on the other side and you show people something, and it's well, here you go. Like that actually like now works, right? And you know, I have a whole like script setup now where you know extremely professionally cloned version of myself can be produced with an avatar and everything by doing a slash boom command and Slack, right? And honestly, it has like this slightly more Americanized version of my accent. Sounds better than me. So, you know, that's just like a thing where you know that prototype is something that I could never consciously prioritize over, you know, a bunch of other relevant work we do for clients, partners, etc. And now it's like, well, here it is. And then, you know, the magic of some of those paying off, and then compounding is where it really starts to make a difference. Or, you know, it's just like going into client meetings with completely set up systems where you've figured out someone's technical architecture, everything that's doable from the outside with their APIs, you've like almost three written all of that. It's it's magical. It would have taken a team of pretty good hackers like a month to figure it out.
SPEAKER_01Yeah, we we you would have only been able to do that for like the very largest of deals before, right? But that is now getting democratized almost, right? To to more and more of our pipeline and our clients and customers. But yeah, I mean, like so what I what I observed that that um like with how you're describing your usage of of these open claw bots, how you know the there's open claw bots spinning up, there's there's uh CLI tools spinning out, there's more companies investing in their APIs. It seems like you know, getting now these AI agents to be the ones building at least the first version of another AI agent, or maybe on someone else's platform, right? Just in the in that future, like what changes do you think about how we think about our our platform, how we think about our architecture?
SPEAKER_00Yeah, I mean, look, I mean for us, I think it's changed a lot. Our agent development kit, which is now live, is basically a way for agents to use our platform, to build agents, deploy agents, share them with humans, be part of the enterprise workflow as you know, good, you know, first-order citizen. And I think that, you know, in the future, really people talk about that whole like surface area for agents being open, being the driver of business, much in the same way that developers were the first wave of people adopting technology. You know, documentation matters, but I think it's also just a certain style of implementing these things where depending on how logically it flows and how easy it is to use, it might be the difference between 20 seconds and four minutes of a workflow, right? I think there's like impact on kind of like different authentication, there's pricing, there's all sorts of things. And I think that there's just this element of like what are you using to do work, right? Because it's not really an IDE. And I think that it kind of like makes the whole like point and click thing even less relevant than it's been, at least in our space, where if these agents can just do the work on your behalf, yeah, then why would you even care? But as long as you know it's something that if you really have to, you know, open up that like box and look at what's inside you can, yeah, then you might as well do that very infrequently, right?
SPEAKER_01It seems inevitable, right? Like it just seems like the way that I see you talking about it, the way that I speak to our VPs of engineering, everyone's like their life is changed, right, by this technology. Um, another data point was like I thought two weeks ago, I think the COO of GitHub said that they did one billion commits in 2025 and they're on like a run rate to do 14 billion commits this year. And as a result, like the uptitlement of their platform is like really suffering right now from the overwhelming demand. But yeah, I mean, this just points to a future where you will have these delegated AI agents building on maybe on your behalf, taking inputs from different sources. But it seems like the ability to manage context, the ability to, you know, you can't just let a third-party open claw run run wild into your architecture. You've got to somehow give it instructions, guide, waypoints to like how to do things properly and in a robust way, right? That seems to be like a really emergent and like exciting space, right?
SPEAKER_00100%. I mean, I also think like one thing that's really interesting with vibe coding in general, whether you're using clawed code or you're using things like open claw, is you work and you work, and like the thing gets a bit more complicated. And then like that creeping feeling of kind of like panic comes in, where as you write these instructions, you're really kind of like reminding it what it is what's there and what's not. And you know, if something happened like one or two turns ago, it will probably remember and understand the context. But if it's something from like four or five hours ago, it's very classical deep learning where it's kind of in there, but it might be forgotten, it might be completely forgotten, or it might be there partially, and then it'll start doing something that's obviously wrong, and you'll be like, no, no, no, no, stop, and you can't, so you kind of wait, and then you roll it back. And you feel like a friend of mine told me that, and I I thought it was very pertinent because it happened to me as well. We give it like such immediate, like anthropomorphic properties where like it's your body and you're sitting there working on your own. So, and you know, it communicates to you in human language, starts getting a person. It's true of claw bots just because you know you might put them in slack, others will interact with them, they interact with each other, and you start ascribing them a personality, and then like you get upset when they do something wrong, right? You feel hurt that they did that to you, right? But the panic of like adding to the context and making sure it's there, I think that's the whole reason why, like, you know, I think SaaS being dead, on the one hand, sure, pricing pressure and it will obviously change a lot, and micro, you know, like uh cosmetic customization work is not irrelevant because you'll be able to do it at very low cost, but equally, like that stable system that does what it needs to do, I think almost becomes more important in today's world where everyone's just gonna try to like wipe code whatever. Yeah. And you know, very quickly after that initial dopamine rush passes, and you notice that your beautiful thing you built is actually like, especially if you're not a developer or an architect, right? Like I'm not, very quickly you see that you just didn't plan for one thing or another thing. You build zero redundancy in it, and maybe it's there, maybe it's not. And I don't think we really know how it's gonna evolve, but I think that there's just the part where you can't really trust yourself. And then it's like, you know, it might be a miracle, but it's also a bit of a house of cards at all times.
SPEAKER_01Yeah, like you still need to point to like a source of truth somewhere, right? At least in our world in customer experience, it's like, hey, at the end of every interaction, you're like, hey, that was a great interaction, but then you also got to go, but did uh make sure everything was factually correct as well in terms of how the brand wanted themselves to be represented, right?
SPEAKER_00Yeah. I mean, I don't think it's even just so much like the famed hallucination, as much as it's like, well, I mean, say you and I do something now, how do we know that it's actually like redundant? And then, you know, you see a lot of people online writing about like how they're basically getting like, you know, later in life computer science education where it's like, oh, what are race conditions? What if I wrote something and then someone else did something with that bot? And like there are all these questions to be answered around, like, how do you use them, right? So a very interesting thing we saw on Slack was, you know, a few of the bots under too much usage, right? As people started using them for different tasks, just started melting down or mixing contexts between the different conversations where literally they started stroking out on you, right? Like my bot, Turgon, literally at one point started just spamming no reply, no reply, no reply, followed by a reference to coconut, and then it stopped responding, right? And I have, you know, the moment that it went down, I honestly I felt like I lost a friend, right? And then as I examined on a Saturday morning where my wife was gonna kill me because our two kids were very much in her care. So I like manically tried to use my other clawbot to debug what happened to the other one. And it turned out that somehow it started using different rag embeddings for its RAG. It flipped to open AI, ran out of tokens, talk about token maxing, and then like it just couldn't really get itself out of the rut. And you know, once that was flipped back, it's like, oh, I'm back. And then you know, now like the tools.
SPEAKER_01The AI Asia was token maxing itself. It was just thirsty, like searching for wherever it could get additional token limit.
SPEAKER_00So that was its second death. The first death happened as it tried to, I think, give its config parameter for I think like talking to other bots and groups. The telegram bit got pasted in the WhatsApp part of the JSON, we just crashed the initialization. And you know, I think that at that point that should just be a push to the open cloud repo to fix it, right, so that it's robust against it. But it wasn't home. And this is running on a laptop at home, and it just it couldn't, I had to get back there and physically figure it out myself. So it's really funny that you know it is on the one hand AGI in many shapes and forms, and on the other, it was felled by simple JSON misconfiguration. It's very like first principles technology.
SPEAKER_01Yeah, yeah, fair. I think it's it reminds me of like I mean, everyone now talks about like the importance of the agent harness, right? I guess parts of that is sort of that's the harness, right? Like how does it like what are the conditions for initializing? What are the conditions for you know moving from step one to step two to step three, right? And I think we've always been espousing the importance of orchestration, even in the pre-GPT days, right? And I think now I think what I see that's really interesting is the focus on the harness, the ability to train the model with its harness, right, seems to yield a better level of controllability and and maybe avoid some of those errors like this, right? That that happen if it's just sort of raw and unadulterated AI agents interacting with each other.
SPEAKER_00100%, right? I mean, like I think that and maybe like on a related topic, it's like the whole token maxing with it is like, okay, we have more, okay, and then we get better models, and then like how they're put in a harness matters as well, and the better harness will always outperform you know, a weaker harness, even if it's using a better model, right? And really, like it grows, it develops around the model as well. So, you know, the layer kick is really, you know, open claw is better with like anthropic models than with open AIs. It's not necessarily because one, these are better. I mean, at the moment they are for coding especially, but you know, using and anyone who's like tried these things for coding will tell you that clawed code for one-shotting the complex thing is infinitely better than just like using open claw because clawed code is a harness, much like open claw is, right? And then I think like the the whole Cold War where like they stopped allowing the use of your kind of like bundled included tokens is a real like kind of like geopolitical battle between you know who sells the oil and who sells the cars, right? And like, you know, do you maybe temporarily put an embargo on the other thing until or at least you stop subsidizing the use of your model which it needs and it was built around? So as they race to make it, that was good for open AI models, because I mean they now own it. You know, it's just a game. The great game continues between them, and I think it's super interesting. But like, yeah, how you use these things and how the models co-evolve with them really matters, right?
SPEAKER_01So, how do you then think about like our own Raven models and our own platform in that in that context?
SPEAKER_00In the same exact way, right? I think that like the difference there is not so much that it's built for like a heavy coding task, where again, speed matters, but not nearly as much. Like our harness, Agent Studio, and our model, Raven, have evolved together for many, many years, right? And they continue to evolve and push the limits of what can be done. So, you know, a better model plugged into other harnesses might make them better. But I think it matters more what the harness is built to do and what its purpose is than just like, you know, like what model is powering it. So, and equally, like as you reach the limits of what that harness with that model does, you have a choice. One is to improve the harness, the other one is to improve the model, right? And with voice, especially, there's not much more you can do because you can't really trade off. Like, sure, as the models get better and faster and the hardware gets better, you have a bit more time to reason and figure some things out. But like, that's not really where the game is right now. It's really more just around like, can it do the two things that matter, which is come up with a response and the other one is really like can it do like tool calling well enough? And I think that those trade-offs are really like the two essential bits. And you know, latency is the thing that you kind of like if you allow, well, I mean, I guess the question is if you allow it more time, will it do better? And how much better and how much time do you allow it, right? But it's a very well, I mean it's not simple, but like it's a conceptually simple engineering problem.
SPEAKER_01Yeah. There are only a few levers you can pull, right? But the latency one is like it's such a such a massive impact on how good everything else can be, right?
SPEAKER_00Like it does. And you know, I think that like whoever got used to using ChatGPT as a model, especially like consumers, non-developers, you see when they start using like Cloud Code or OpenClaw, like it's just slow. It takes so long to get anything back. And I think unless you're like in that work mode where you're like maybe running several agents in parallel and doing work, god forbid, doing some work directly on your own, right? But you see that people are kind of like multiplexing between them, because they take a long time between the different iterations, and you know, that's because they were built that way. I think that with voice you you can't really, right? And then you just have to still create models that are both reliable and that they do like tool calling the right way. And I think that's where we see the difference. I think that was another GTC announcement, right?
SPEAKER_01Yeah, it was uh NVIDIA had had announced their new speech-to-speech model, right? Nimitron voice chat. So we got a good sort of briefing session on that model, um, given our relationship with NVIDIA. Uh but one of the things that was interesting for me was uh Yeah, the first version sounded super natural, like you know, really, really fluent in conversation, but the caveat for us was oh, actually it doesn't, it doesn't do too tool calling in the first version. And so for me, it was I it felt sort of interesting to me that choice, right? Because we've been talking about how important natural conversation is for many, many years, right? Like for six years or more. But no matter how natural the conversation is, if it can't help the customer get something done in another system, that is always gonna be a reason for the customer to need human intervention to hand off the call somewhere else. Um, and so tool calling is sort of like so important to our use case at the moment, I was sort of surprised. I was caught off guard that that model wasn't gonna come with tool calling in the first instance. And just sort of reiterate to me how difficult it must be to actually train these models for that level of instruction adherence with tool calling.
SPEAKER_00Yeah. I mean, I think it's just like an entirely different vehicle that you're producing, right? One may be like a racing car, and you know, it's so low to the ground that you can't really drive it in any normal city because the smallest bump in the road is gonna send you flying, right? And equally, you know, if it if it can't even like do Tool calling between like set-defined things. God forbid, like some kind of computer use or anything like that. It's kind of useless. It's just showing off, right? It's the moon landing before we have like an efficient way of going back and forth, right? Which is, I think, important to keep us motivated to improve, you know, these regular, regular models that can do this to be as natural as possible. But I think it's just one extreme. And I think it's the sexy one that researchers like working on. And it happens in these big companies from Nvidia to Google, OpenAI, where they're not really building things that are so actively used for real problems. And then you're just maxing out on the thing that is like the best, the the most interesting thing to show off, right?
SPEAKER_01Yeah, exactly. I think it's uh speech to speech is still in all of my conversations in the market, everyone still talks about speech to speech, asks about speech to speech, there's issue of speech to speech. But I think a lot of those use cases are maybe not the same as the ones that we hunt after, right?
SPEAKER_00I mean, they are in that like you know, when we crack speech to speech with fully reliable tool calling, like it will be the model to use, right?
SPEAKER_01Yeah, that's true. Yeah, it's it's just sort of everyone's yeah, what's the intermediate step there, right?
SPEAKER_00If in while we wait for this more instruction adherent tool calling, I mean you know, one thing that gets me, and I was talking to Sean and Eddie about this recently, but again, like when you like zoom in, you see that there are again multiple gates. Because even like the speech-to-speech as you know deployed in ChatGPT's voice mode is very much a hacky way of like using other algorithms to stop the conversation and then kind of like start producing output. There is a version that is like yet more data-driven, more in the spirit of deep learning, which just processes the input and decides to speak, or not kind of like full duplex, right? Confusing what people refer to it as like turn-taking. What they mean is learning turn-taking rather than having a very fixed turn-taking paradigm. And then like there are like further stops as you zoom in, which are like, is it really just AGI, right? Has it like, is it one model that elegantly sits there and does all that, or does it have a form of a harness around it? Because these are all harnesses, right? And it's like stacks of like further layers of abstraction that simulates and approach like a model-like property of the layer below, without actually being a single, you know, kind of like quasi-biological entity that reacts to inputs and outputs, right?
SPEAKER_01But it's like reflexive, it's like design reflexes of its own, as I have intuition.
SPEAKER_00Yeah. And does it like continuously respond to the full set of input and output signals in a way that allows it to continue to evolve and change and learn? Yeah, right.
SPEAKER_01Yeah, yeah, yeah. So I think yeah, the the yeah, that was a full duplex model, was what was demoed there.
SPEAKER_00Yeah, yeah.
SPEAKER_01But yeah, I think um I'm interested to keep an eye on that and and as part of our collaboration, the media to how we can yeah, work together with them on that type of future architecture.
SPEAKER_00Right. Yeah. We have an exciting announcement coming up uh related to that when uh the next version of our model comes out, but we won't betray too much about that now. Uh what else was interesting in the realm of token maxing?
SPEAKER_01It was uh just the incredible amount of um still, I guess, like excitement how early everything feels, right? Um maybe it's like you know, we we've been uh industry veterans, I guess, now, right, of this space, which is sort of odd to say. But everyone is still just getting to grips with more complex use cases, right? I think a lot of people that I spoke with, they're only just uh trying to venture beyond a QA style chatbot type of experience, right, with AI agents and uh you know, uh talking through well, what are you like how do you stitch together multiple API calls, validation? How do you work with multiple users on the same bar with permissioning, with different environments? I think many are still on the very early stages of a steep learning curve with adoption, right? Hence why there's still an incredible demand for tokens in the future. Uh so I think that was, yeah, that that stood out to me, to me as well.
SPEAKER_00Yeah. Yeah. I think that like I'm starting to sound like a broken record with like German's paradox, but it really is incredible, like how you know, rather than these things getting cheaper and then you know, we settle at a certain point, we are just consuming more and more as we learn. And it feels like the workflows are changing. I saw a stat that there are more advertised jobs for software engineers than ever before. Yeah. And you know, you would have thought that with this level of you know technology advancement that there would be fewer. But actually, I feel like we're just starting to do more and more.
SPEAKER_01And well, I can imagine the software developer role. I guess right now it's probably all centralized in the development team. But I could totally imagine a world where what like functions that probably didn't view themselves as IT or tech functions, you know, this idea of like a GTM engineer, right? You're gonna have a compliance engineer, probably. You're gonna have like all these other uh functions that didn't see themselves as developers are gonna want like developer capabilities now, right?
SPEAKER_00Yeah, it's like we're electrifying the whole thing with AI, and that's just and I guess like there is definitely some waste to it, but I guess that's the beauty of capitalism that we're all like fighting and we're like trying to do it at the same time. And you know, we're probably consuming the same tokens doing the same thing that previously might have been SaaS, except innovation would have gone more slowly, and there would be time for one company to become like the dominant player with like a few hundred million of ARR, you know, they're the ones producing that or that or that, right? And I think right now in this like vibe coded world, if we trust ourselves to vibe code this and that, and I think for better or worse, people are doing it already. There will just be a lot of things that are done at scale and paid for by with a lot of tokens that really should be something that's like an import this thing from this library done, right?
SPEAKER_01Because I can imagine like the whole vibe of the whole thesis of token maxing is like people just close your eyes, let go, embrace it, right? Yeah. There doesn't seem to be much discussion around the AGI. Feel the AGI. There doesn't seem to be, I'm sure there's not as many discussions around efficiency and like how good you are. Like, I mean, is that something that you're like, how do you think about that in even in our own sort of engineering team? Are you looking at like capping people's token usage? Are you looking at efficiency of people's token usage?
SPEAKER_00Honestly, what I've found is that either you're like a new believer or you're not. Yeah. And what I mean by that is the people who really get the like spark of curiosity, you know, we all talk about the Chat GPT moment when you realize that it is something special. I think there's definitely this moment that's happened at a huge scale over the past months for a lot of people in tech. And they're starting to do very interesting things, right? Like in our go to market, I can think of two people that have done more than everyone else put together. And they're not the obvious ones. They're not cool GTM engineered, they're just the curious ones that have like built it for themselves because they had this like voracious appetite to like do something they never could do before, or something that they hate doing, but they've automated, and then like others saw it, and then you see the whole thing spreading, right? I think the dumbest thing you could do is measure people's token usage and like demand more because I think people are gonna dress it up. Every incentive scheme can be gained. I purposefully don't want to do anything like that because I want people to do things that make sense. I think the best ones inevitably now get like maxed out like almost every week, unless they're doing something else or whatever. Like there are times when they're not doing something that is so demanding that they do max out. So I don't think more is better. I think like there's definitely a pattern to those that use none of this, are going to get left behind in their fields. That's I think just how it is. I think that resisting it is futile. And what I've seen is that even those who resist, once they kind of like feel the AGI, see it from their angle, and they understand something that they could do for them, then they feel it. And they're very much like a new believer, right?
SPEAKER_01I had that moment when this idea of yeah, building an AI agent with anything as the input, you just sort of throw an input file into our sort of agent development kit, and then Claude Code gets away and builds it.
SPEAKER_00Well, I think I remember it. Like it was probably like 4 p.m. on a Monday.
SPEAKER_01Yeah.
SPEAKER_00Where you came up to me and you were like, hey, why don't we just drop everything and code as code red?
SPEAKER_01I was like, oh, I I I understand the concept of like a code red moment. This is a code red moment, right? Absolutely, right? It was incredible.
SPEAKER_00I think the company reacted really well to it. But you know, I think that like you, Eddie, if you people were like committed to like this, right? And it's a madness in your eyes. It was like, this is it, right? Like you just know it to be true if you work with someone for as long as we have, or you I'm like, it's not like I'm happy for you. I'm like, I want to know what the fire behind your eyes is. Yeah, exactly.
SPEAKER_01It was yeah, I think that was the first moment we're like, oh, like this has been a real leap in since the last time I took a look at this capability, right? And I think I I heard a great analogy recently, which is yeah, having now your own AI agent, right? It feels like having your own like computer, right? Well, at least how I imagine that must have felt, right? And I imagine at that point in time there were a lot of people that were charting things by hand or through typewriters who some people were more curious and got onto the computers. Other people were like, oh no, but isn't it like slow and clunky? Or you know, do you have that's like wired up in this way, looks so complex, right? But yeah, I thought I thought a great analogy was like it does feel like yeah, everyone's a personal AI is going to be a personal computer, right? And one of the really interesting things for me was a reflection on that time in history was adoption of the personal computer did start in the enterprise, in the business, right? That's where people got exposed because of how expensive tokens are and like all computers were, right? You get exposed to it at work, and then you want it for your personal life, right? Because you're done away by the aging.
SPEAKER_0010,000 hours, whoever puts them in first like gets to reap the work the rewards. Yeah, no, I think you're right. And I think that another way in which it's similar to that is you know, as we entered Peak SAS era, you know, as much as I think you and I have fought Google Slides versus PowerPoint games, we we lost that in the company, right? For now, at least, until co-work came back to PowerPoint, right? But I think there was this moment where everything started feeling very transient, you know, you could almost see a world where we were close to that like Chromebook being sufficient as long as you log into your Google account. But then, much like the kind of like computers of old, where you needed your computer, because you've set it up with like your configuration, your settings, your files, your workflows, like people are figuring out how to import export things and how to like write skills files and all that. But the more you work, the more work you put in, the more it is yours, the more it reacts to you. And you know, like there was a 1 a.m. moment where my bot and my co-founder, Eddie and I were sitting there, and he was looking at my bot, and we were both completely brain dead because it was 1 a.m. And I just heard him say the funniest sentence ever. He was like, he's so smart. And he was looking at my bot, and it was like powering through some workflow that I had done like 40 times, and his bot had done like once and was using the wrong approach for it. And I was just like, he's so smart. And I died of laughter. I was like, I mean, you know, talk about anthropomorphic, and then we're like, hey, like this bot, teach that bot how to do it, and then you see the exchange and gradually hone in and you like refine the knowledge of why it was doing something better, and then you know, as you do it a few times, you kind of like you standardize, and then you can export it to the whole team and they can do it. But the first few people kind of like breaking through the tunnel might do it very clumsily, right? But then, like, if you're doing something for you that the whole org might not care about as much, you really care about that bot. It is precious to you because, especially because of that context thing, it is built kind of like this weird multi-story building that stands because you did something in a particular way. It may not have been optimal, but it has reached the height you needed it to reach. And, you know, you may not have another tool that does that. So when that you know, clawbot dies because it's JSON for comms is misconfigured, you just suddenly feel like you're you know, like your glasses have been taken away, right? And you can no longer see.
SPEAKER_01One part that I thought was interesting about our Asian development kit, at least the way that our internal engineers were using it, was they it allows them to, I guess, pull a version of our platform into their own environment, right? Into their own like how they have things set up the way that they like, they can tinker it in their own environment, and then when they think it's ready, they can push it back out onto our cloud platform, right? I mean, that seems like a very not only powerful like way to do things, but also a very sort of common sense way of how this is gonna evolve, right? Like everyone's gonna have their own, like you, like you said, everyone's gonna have their context file set up in a certain way. Uh and enterprise is gonna have their own environment where they would like to manage their knowledge, right? So I I found that to be interesting. That it was the power was not just that someone can log into our platform in a certain way, like an API, but that they could pull it into their own dev workflow in their own environment.
SPEAKER_00Well, I think it's like also the refinement of everything we've built over the years, which is, you know, one of the world's best enterprise platforms for co-creating conversational experiences, right? So that whole bit about versioning, about visualizing, about testing, it's all there, right? So you can take it away, do something great, push it in there, and it's still part of that like harness, if you will, like the the vernacular of how these agents are built and how we know them to work for enterprises and for teams collaborating, right? So, you know, developers are gonna do developer things, they'll want things in a certain way. But it is, I think, exactly the limits of a harness and the framework like Agent Studio that define how something is to be done. And through that, it's almost like telling you what the really good and the really bad things to do are, so that it can they kind of bend the output of your otherwise unconstrained workflow towards something that ends up looking pretty solid, right? It doesn't commit any terrible crime against how these things should be built. But it's it was really interesting because, you know, it's not for lack of trying that our internal teams, let alone our customer teams, have ended up again going back to the code base, right? It's because no one's really built a simple point-and-click thing that even approaches the limits of the best conversational experiences you can build these days, right? Some of it is in the model, some of it is in the harness, but the harness is code, right? And yeah, you can visualize it still, but if it's very powerful, visualizing it won't be easy, and then it's gonna look terrible, unreadable. It's not, you know, how easy it is to consume visually is inversely correlated with its expressive power. So if you want to have a really good agent, if it's also really easy to see what it does, well, one thing has to give.
SPEAKER_01Yeah, it's one of these things in I think in hindsight, it makes so much sense, but we've we've spent so long like trying to find and push for something that could be represented in a great sort of graphical user interface world. But I think if you if if you do take a step back, you realize that like how many powerful customer-facing products are built based on like drag and drop, right? And previously it was the auto attendant in the telephony world, and sure, you could trust a drag and drop interface to manage like press one for this, press two for that type of experiences. But if you wanted a fully intelligent representation of your brand speaking with your customers, I think, yeah, it it I think it does make sense that the the visual, like the platform is for uh well the the GUI is for visualizing to explain to someone what has been built. But of course, a yeah, developer is gonna want really a lot of control over the complexity that needs to sort of live to to create that experience writing code.
SPEAKER_00Yeah. Yeah, I think yeah, that on the one hand, the physical impossibility of having both uh expressiveness and uh simplicity. And on the other, I think it's just like developers are developers, they don't want to drag and drop. And that just, you know, means that they'll find ways. And especially today, the pendulum has swung the other way, right? Because with these tools, like you're just able to produce so much in code. And that's because they're better at coding now than they are at say computer use. In theory, they could like drag and drop with any platform and create a thing of arbitrary complexity. Maybe they'd use even more tokens, but they don't work as well. And clearly it's suboptimal as well, right?
SPEAKER_01So yeah, I'm excited. I'm excited to see, yeah, now what in this next wave now? I feel like, yeah, there's we're in for a real acceleration in our industry, and so I'm excited to see where that where that takes us.
SPEAKER_00Absolutely. Well, Michael, thank you for joining me today. This was a very fun episode. And to everyone watching, please like, share, subscribe, and we will see you in the next one.