Knowledge Distillation

Why 35 Years of Web Infrastructure Isn't Ready for AI Agents

with Katrin Ribant
Podcast
Knowledge Distillation
Host
Katrin Ribant
Duration
65m

About This Episode

I joined Katrin Ribant on Knowledge Distillation for a deep, technical conversation about why the web isn't ready for AI agents and what we need to do about it. This one covered a lot of ground: from the real-world impact of vibe coding to the ChatGPT data exposure I helped uncover, to what analytics teams should be tracking in a world where bot traffic has crossed 50%.

The core argument: make boring sexy again. Semantic HTML, proper accessibility, fast page loads, clean code. These aren't exciting buzzwords, but they're exactly what AI agents need to understand your website. A span styled to look like a button fools humans just fine. It doesn't fool an agent. When AI shopping agents are trying to complete transactions on your site and your front-end code is a mess of JavaScript bundles and Tailwind utility classes masquerading as semantic elements, they will fail. And unlike humans, agents won't adapt to your broken site. They'll just go somewhere else.

We also talked about the shift from an attention economy to a utility economy. Bounce rate, time on page, page views: these metrics were always questionable for humans and they're meaningless for agents. What matters is task completion rate. Did the agent accomplish its goal on your website? How many wrong turns did it take? We don't have great tools for measuring this yet, but we're in the pre-Semrush era of LLM visibility. The tools are coming.

On the security side, I shared the full story of how we discovered ChatGPT prompts leaking into random Google Search Console accounts, what OpenAI's response was, and why cloud-based AI systems and sensitive data should never mix.

Key Topics Discussed

  • Why 35 years of web infrastructure isn't built for AI agents
  • Making boring tech sexy again: semantic HTML, accessibility, performance
  • The shift from attention economy to utility economy
  • How front-end A/B testing scripts add two seconds to every page load
  • Vibe coding vs. AI-assisted coding: a critical distinction
  • The ChatGPT data exposure and why cloud AI can't be trusted with sensitive data
  • How GA4 analysts are actually well-positioned for the agent analytics era
  • Red flags when evaluating AI-powered tools and wrappers
  • Digital tinnitus: protecting mental clarity in the age of AI hype
Transcript

Slobodan Manić: We're not optimizing the product pages for humans and for emotion and for triggers or urgency or whatever else anymore. We are optimizing for clarity. And we need to make sure that the LLMs and all those AI systems understand exactly why they would want to do business with analysis, are in a great position because session user, you know, abandoned rate, Bounce rate. Bounce rate is. There's no such thing as bounce rate for an agent. It's either completing the goal or not when it's on your website. So I think we're looking at different metrics, but the same kind of people and the same tool set still will apply. What happened with external CSS files and HTML for just for the meaning and CSS for the decoration when I was, when I was getting started. And again, this is old man yelling at the cloud a little bit. But HTML was the meaning, CSS was The decoration and JavaScript was the interactivity.

Katrin Ribant: That's how I learned it.

Slobodan Manić: There was nothing wrong with that. It looks like a button to you. It doesn't look like a button to a user who uses assistive technologies or to an AI agent. So use the correct HTML tag, use all of the correct HTML attributes, make it painfully obvious for the agent. We're talking about agent humans as well, but for the agent, make it painfully obvious what each element is and what it does. If you're loading that snippet in all of the pages on your website, you're adding 2 seconds of page load to every single page of your website. Furthermore, that JavaScript file is, it's a third party asset. It's loaded from a CDN that you don't control. It's. It's whatever the platform is, let's say Google optimizes because it's dead, so we don't throw anyone under the bus. But let's say you're loading that from a third party cdn. That's a cardinal sin in performance. You're loading a third party render blocking asset, which means if their CDN goes down, your website goes down.

Katrin Ribant: Welcome to Knowledge Distillation where we explore the rise of the AI analyst. I'm your host, Katherine Ripon, CEO and founder of Ask why. This is episode eight and today we're tackling a shift that's reshaping e commerce and the web at large. How websites must now satisfy two humans and AI agents. For this, my guest today is Slobodan Managed. Though everybody calls you Sani. Sani, you're a website optimization consultant with 15 plus years in web performance CRO also technical SEO. You work with teams at Instacart Klarna. You also wrote for Search Search Engine Journal and certainly did a whole bunch of things that my research hasn't yet unearthed. You're also a WordPress core contributor and obviously you host the NoHack podcast, which I think has 200 plus episodes at this point now. Subtitled Optimizing the Web for AI Agents. Is that correct?

Slobodan Manić: I think that is correct. First of all, what a wonderful intro and thank you for having me. It's an honor to be a guest out there.

Katrin Ribant: Well, let's thank Claude for the help. Not that I didn't know someone these things and I can read LinkedIn, but Claude was really helpful.

Slobodan Manić: Multiple conversations already before you did. So thank you, Claude. Let's say that.

Katrin Ribant: So you know all of these tells us exactly where your focus has landed and why you're my host on this episode. So would you tell us a little bit about why you and like how and why you decided to focus on this?

Slobodan Manić: So I, like many people, I felt into the trap of needing to know everything about AI as soon as it's out. And we've been living that hamster wheel right there the last two or three years. Like everything is the most important thing in the world. Everything. It's a. Right around the corner some big things are happening. It's going to be amazing. I'm tired of that, honestly. I. I spent the second half of last year just being tired of trying to keep up with everything. I don't know, maybe it's ADHD and the need to just know and explore everything. Whatever it is, most optimizers have something. I just got tired, honestly. And also I noticed that every new update and release is kind of the same as the old one. ChatGPT5 was that even a new model? Frontier model? Was that just a label on the old model? So I think it doesn't matter everything that's happening in the AI. Most things that are happening in the AI world, not everything don't matter. And I think the ratio is 99% hype and bullshit. 1% signal. If you don't know what the 1% is, it's best to sit it out and just wait to realize to see what. What's actually worth paying attention to. Also at the same time we have 35 years of web infrastructure, all the websites we've built over the 35 years, and now we have AI that needs to look at them and understand them and who's. Who's doing that work like that feels urgent. Necessary and extremely important. More important than keeping up with the latest models. So I'm going back to what I know how to do best. I optimize websites, make them technically healthy.

Katrin Ribant: So actually tell us what it is that you've been specializing on before this shift, like when it comes to web optimization, because I know you very sort of like specialized in one specific aspect and how that sort of leads into this.

Slobodan Manić: So when I started learning about web development back in mid 2000, knots, how do you call them? The zeros, the first decade.

Katrin Ribant: I think we say notes.

Slobodan Manić: I think that, that, that's, that's the term. Something like that. We're not native English speaker, they'll forgive us.

Katrin Ribant: We're just going to say notes. There you go.

Slobodan Manić: Let's, let's say that that's the term. So when, when I started learning about semantic web and how the browser can understand the website and then like why H1 is not the same thing as a span and why you should use one over the other, which it feels stupid that we have a need to talk about that still. In 2026, when I was getting started as a web developer, I was dead focused on semantics, accessibility and learning what code means. Because at that time the browser was still, you know, we had ie6, ie7, all those stupid browsers that needed everything to be perfect for them to be able to understand it. Then I moved into technical optimization, speed optimization, core web vitals. I even did some analytics briefly and technical SEO. The thing that still sticks and always will be there is that you need a healthy website. Whether that's for humans, whether that's for people who have accessibility needs or AI agents or LLM systems. Getting the basics done, I think is the most important thing before you go do your CRO or your, your SEO or whatever else. I think getting that website to be as healthy as possible, everyone benefits. So that's, that's why I'm kind of going back to that. Because now it feels like more people are abandoning websites and just building vibe coding and building whatever else and no one is really taken care of. Shaving two seconds off of every page load. Why are people not doing that?

Katrin Ribant: Yeah, and that does change things to the experience, for sure. And talking about which we collaborated recently on something really fun. It was that Halloween episode of the Floofies. For those of you in the audience who don't know, the Floofies are creatures that I created with AI and I do AI videos with them to explain the functioning of LLMs. Things like what is a context window what is attention mechanism? How do data flow in an LLM? How do tokens get generated? Things like that. And the flu fees are the central sort of like characters of that. And so the episode is called the flu fees and the Instant Checkout Crisis. It's really a parable about brands becoming ghosts where their data isn't optimized for AI discovery. So let's dig into the real version of that. What does that actually look like in the real world? And do you have any real example of this?

Slobodan Manić: So the real examples are kind of being released right now. We had the OpenAI agent, shopping agent, whatever it was, Perplexity has some version of that. So basically what this means is that the user doesn't have to go to your website to complete a transaction anymore. The user can or will be able to just search for a product, ask their LLM or AI assistant on their phone or on their computer, can you find this and this for me or that product for me? And then the agent will be able to identify five candidate websites and then maybe ask you which one do you want to use? And you say, I want to use this one. And then it goes and completes the transaction for you. The user never sees your colors, your logo, your hero image, your warm copy in your product page. So I think that that kind of. I just had an episode on my podcast released yesterday. We're moving and we need to upstream engineering. We're not optimizing the product pages for humans and for emotion and for triggers or urgency or whatever else anymore. We optimize it for clarity and we need to make sure that the LLMs and all those AI systems understand exactly why they would want to do business with you.

Katrin Ribant: And so we are optimizing for those AI agents, but humans still do visit websites and will continue to visit websites. And we still, I suppose, have a need for traditional SEO, be it simply to be present in those search results. So how do you look at those things, those, those two things together?

Slobodan Manić: That's an excellent question. The way you optimize for two different things is to find what they have in common and optimize for that. Any LLM, any AI will need the code to be semantic loading fast and be accessible and be clear and the information presented on the page not to be confusing and not to be self contradictory. What do humans need? They need the same thing. They need a fast website. They want a fast website. This is what Google has been telling us, by the way, for the last seven or eight years, that the experience of the user when they land Even though Google is like traditionally or seven, eight years ago was just SEO, it's just for rankings. No, they've been pushing the narrative of your website needs to be fast with corporate binance. You need to be seen as an expert with EAT and all those other things that they have been quietly serving for the last decade without telling us. This is going to be important in the AI age. Well, all those things are now extremely and even more important than when they were only when applied to humans. So whatever you find, whatever, basically whatever the machine and the human have in common and what their needs are and optimize for that and go heavily into optimizing for that.

Katrin Ribant: So you specialize in fixing what we could call the technical foundations. Right? The accessibility, the semantic structure, the performance. Like all of those really highly technical things that are very hard on larger websites, obviously. Right. And larger, complex websites and you know, complex E commerce, et cetera. How do those traditional web quality signals translate to AR readability? Specifically? Like, let's go. You know, our audience is technical. Let's go technical into like, what is it that you really actually need to do?

Slobodan Manić: You need to make your website as complicated as it needs to be and nothing more than that. So trim all the fat that you don't want to have and you don't need to have on your website. Because when, when an AI or an LLM is processing your websites, there's a token cost. This is not free. This is what currently is being presented as. It's free or almost free. So use as much of it as you want. But if an AI agent needs a hundred thousand tokens to complete an operation on your website, you'll start paying that at some point or your users will start paying. This is not going to be virtually free forever. And we all know that all these companies that have these frontier models, they're bleeding money left and right, so they will start charging more for these tokens at some point. So I think it's really important to. When it comes to speed, for example, the main problem with speed with most websites is that they're loading assets that they should not be loading. They're loading all those JavaScript files that they don't need need in a lot of pages, CSS files and CSS rules that they don't need. Just try to trim as much of that fat as possible and you'll do better than 95% of the websites.

Katrin Ribant: Could we translate that into like an example of the real asset? When you say JavaScript or CSS, like, what does that correspond to? In terms of an element.

Slobodan Manić: I'll give you an example. So if you have an A B testing script from, from a vendor on your website, front end A B testing. So in browser there's a JavaScript snippet that's being loaded that JavaScript.

Katrin Ribant: Sorry. Let's, let's take the example just to irritate all the A B testers out there. Kelly, you in particular, love you.

Slobodan Manić: Yes.

Katrin Ribant: So let's take the example of. We are. Do you want. We are testing the color of the button they all hate when you reduce it to that.

Slobodan Manić: So let's do that but let's reduce it to that. Let's say you're testing a button color in your about page.

Katrin Ribant: Exactly.

Slobodan Manić: You're testing a button color in your about page and beyond that about page there's a. Or maybe just in that about page there's a form where you sign up for something. Nothing beyond that it's one page test. There's no checkout, there's no card. Now most people will load that snippet in every single page of their website and an A B testing the front end snippet is to prevent the flicker of content shifting and moving left and right between the when it's loading a variation it will block showing the page for a second or two until it loads the changes changes it while it's invisible and then shows you the page that adds two seconds to every single page load nipped in all of the pages on your website. You're adding two seconds of page load to every single page of your website. Furthermore, that JavaScript file is. It's a third party assets. It's loaded from a CDN that you don't control. It's. It's whatever the platform is. Let's say Google Optimize because it's dead so we don't throw anyone under the bus. But let's say you're loading that from a third party CDN that's a. That's a cardinal stain in performance. You're loading a third party render blocking asset which means if their CDN goes down, your website goes down. It doesn't make you panic and rethink everything you're doing about how your optimizing your website. This is more important than anything CRO can do in my opinion. So just load the assets you need only when you need them and only at the time when you need them, not sooner.

Katrin Ribant: So what you're just describing is some is a best practice that is valid AI agent or not AI agent. Right.

Slobodan Manić: Those two seconds will be added for every single user. That's AI agent, that's a human. So you're making your website slower because you just added it to the head section of every page of the website because that's what the instructions said. That's not what you're supposed to do.

Katrin Ribant: And is there something specific like. So in the floofy stories we talked about brands with like sparse descriptions and missing details becoming ghosts basically present but unseen. When you look at a site for AI agent compatibility specifically, what are the common failures that you see specifically for AI agents of, you know of that, you know of that sort of level and if you could describe the technical aspect of it, but also sort of like what it translates to into a real use case so we can link it.

Slobodan Manić: So the. I don't want to make this strictly an accessibility story, but this is, this is about accessibility. If you have buttons and you're using something like a tail to style those buttons, you can use a span tag and just give it 20 classes and it's going to look like a button. It looks like a button to you. It doesn't look like a button to a user who uses assistive technologies or to an AI agent. So use the correct HTML tag, use all of the correct HTML attributes. Make it painfully obvious for the agent. We're talking about agent humans as well, but for the agent, make it painfully obvious what each element is and what it does. That's when they hallucinate. Not always. They will not always be wrong. But an AI agent, as you know, they guess.

Katrin Ribant: Yeah.

Slobodan Manić: Of course, if you don't, if you don't feed them a hundred percent accurate information, they will guess and they'll guess wrong sometimes. So make it painfully obvious on every single level. And that's the visual of your page, that is the code of your page, that is the content you have in the page. Painfully obvious what it's about. That, that is an.

Katrin Ribant: I see what you mean when you're saying that. It goes back to the fundamentals of what you were doing in the old. And I'm just going to like say that across the episode because I like that we're saying that when you had to optimize for those, you know, crappy browsers that we had at the time that were requiring perfection for everything. Basically, paradoxically, we are working with a guessing engine that tolerate ambiguity. But if we want that engine so the LLM to be able to guess correctly about our website, we have to be painfully accurate.

Slobodan Manić: It needs help. It needs as. As much help as we can give it. Yes. And even then, who knows? But it's more likely to get the right answer. I mean, if you go to any LLM and ask it something that's kind of ambiguous and you expect it not to know the answer, it's going to give you a different answer as many times as you ask it. When you have a messy page with messy and confusing HTML rendered in the browser, or messy which leads to a messy and confusing accessibility tree, which is what these agents are doing when they control the browser, what they're using. It's more likely, it's infinitely more likely to make a mistake if it doesn't know what the subscribe button is or like whatever the action is in the page.

Katrin Ribant: So if we think about this in this way, where it means that basically we have to really fix the accuracy of the technical structure of the webpage, we are ultimately also creating a new data layer for the analyst. Right, because all of these, these elements are what gets tagged and what then feeds into the base data that we use in digital analytics that we all complain about being so messy because, yeah, I mean, you know, you hack one thing on top of one thing on top of one thing, and when you have an older website, you end up with all of these subsequent layers of hacks where you end up with a data layer that is just like extremely messy. So good news is this may force a lot of brands to sort of clean up some of that aspect.

Slobodan Manić: I sure hope so. I, I'm not. That would be amazing. Because we know the answer. We know the answer to this problem. Just fix like, if you have a bad road, the answer is not to build the bridge on top of it, it's fix the road. Like we have the foundational layer and that is the HTML of our websites, which is the language of the browser. And these AI agents are using the browser to access and understand and read the website. We have it already. We know what it needs to be. All the specs, all of the HTML specs, they barely have changed at all in the last 10, 15 years. We know exactly what we need to do. But then again, for, I don't know, developer experience, we have REACT to make it easy for developers to write components and just do single page apps. So we have tailwind that gives you 20 classes to make the button red. And what happened with external CSS files and HTML for just for the meaning and CSS for the decoration When I was, when I was getting started. And again, this is old man yelling at the cloud a little bit, but HTML was the meaning, CSS was The decoration and JavaScript was the interactivity.

Katrin Ribant: That's how I learned it.

Slobodan Manić: There was nothing wrong with that. And. But. But now we have everything in one big JavaScript bundle and we call it a single page app and. Cool. I just don't get it.

Katrin Ribant: So note to everyone out there, if you're in an organization where you hear chatter about redoing the website for AI agents, I suppose this is in every organization that has E commerce at this point, right? Quite frankly, probably every organization, period. Because everybody needs to be discoverable by LLMs. Try to get into the conversation because this is your chance to maybe fix some, at least some of the data layer. It's a chance that's not going to come back because, I mean, whoever redoes these fundamentals, if they're not forced to it.

Slobodan Manić: This is so painfully simple. This is like going to a doctor and they tell you you need to walk 10,000 steps. Yes, but is there a pill? Forget about the pill. Just do the 10,000 steps and you'll be healthy. If they tell you that, just do the basic thing and things are going to be better for you. Doing the fundamentals and the basics. I don't know. Let's make boring sexy again. If we're going to have a slogan, let's try to do that.

Katrin Ribant: That's a great slogan. I like it. I just want to come back to your HTML, CSS, JavaScript separation and bundling, et cetera, because I definitively know your opinion about Vibe coding. You're not shy.

Slobodan Manić: I'm not shy. I'm not. Sh.

Katrin Ribant: Not shy.

Slobodan Manić: I'm not. I'm not. Not at all.

Katrin Ribant: So do you see brands trying to fix these issues with some Vibe coding of web pages? And if so, feel free to rant about it.

Slobodan Manić: So that's an interesting question because it's fixing a problem by not understanding what the problem is, essentially, because Vibe coding. And let me just define what. What I mean when I say Vibe coding and why. I'm so not as. I'm not fired up. I just. I don't think it works.

Katrin Ribant: Come on. You're fired up.

Slobodan Manić: Okay. Basically, AI assisted coding, when you understand what you're doing, is not Vibe coding. Like, that's not what I mean at all. If you have a developer who understands what they're trying to build, what good output is, what versus bad output, if you have that, that's not Vibe coding. Vibe coding is I have a dream, I have a. I want to build this. I don't know how, but I know how to type a sentence or two and I build it and I call it a product. That's what I call Vibe coding. And I don't think long term that can work. So basically, if you have a website that's essentially broken works, but it's kind of not. Doesn't work really well and you don't know why, and then you ask AI to fix something, you don't know what it is. It just doesn't make any sense on any level to fix it like that. One thing that I saw, I think it applies here. I saw a screenshot of a Vibe coding prompt that says, can you add class red 400 to this element? That was a prompt to just type red 400 where the classes are. Why waste anything? So rather than using AI to fix something in a way that we have no idea what it is, it's maybe just look at the code and understand what's wrong and try to understand or hire someone who does. Because long term it's going to be cheaper, I guarantee you that.

Katrin Ribant: Yeah, that's true. But I suppose we're in that weird limbo where a lot of new functionalities needed to handle this efficiently aren't really covered in a mature way. Right. And we are stuck between building our own tools or trying to understand, you know, the different aspects of what we confronted with. On that note, actually, Shameless Plug episode six with Jim Giannuoglio is talking about episode six of Knowledge Distillation where Jim Giannohlio is talking about just that subject, like the AI analyst as a builder of tools and a tool and a consumer of tools. And when I was asking the previous question, I was just thinking, I can totally see somebody going like, oh, I understand that this page isn't really discoverable by an agent. Let me just take the code, throw it into Cloud or, you know, or Gemini, and ask it to fix it for a while.

Slobodan Manić: I love how you're leaving ChatGPT out. That's the way in 2026. Let's, let's leave it out of the conversation, please. Yeah, sorry.

Katrin Ribant: I tend to. I tend to. And like, my question is, do you see a lot of that? And that was kind of what I was sort of, you know, thinking about it in terms of Vibe coding. Because it's basically vibe coding, the code of the, of the page, right? It's like sort of.

Slobodan Manić: There's an assumption in that process. There's an assumption that the output is going to be correct. Why are we assuming that? Do we have a reason? Why are we assuming that LLMs are correct, they're correct a percentage of the time. Just we don't know when they're wrong.

Katrin Ribant: If you don't know what that percentage.

Slobodan Manić: Is, we don't even know what the percentage. Exactly. Exactly. So if you don't know what the correct answer is so you can verify what the LLM told you, don't use an LLM. If you know what the correct answer is, you can do it yourself faster. So why.

Katrin Ribant: Yeah, well, you know, first of all, where there's a vacuum, there's going to be people trying to fill the gap. Right. And one of the things you've written about this space is that it's poisoned by AI, that most AI wrapper sort of SaaS products will be worthless in 12, 24 months. Thinking about, you know, people in the audience either responsible for websites building data layers, analysts that are evaluating some of these new offerings, and I have to confess, I know nothing about these offerings. I just assume they exist because again, when there's a vacuum, there's something filling the vacuum.

Slobodan Manić: I love that.

Katrin Ribant: So for people evaluating like AI powered e commerce of analytics tools, etc. What are like, can you point some red flags that people should look at not to fall in that trap.

Slobodan Manić: When analyzing a tool before using it? Right.

Katrin Ribant: Yeah. Like an AI wrapper or like something that would, that would, that would promise to help you with this particular problem.

Slobodan Manić: Well, if you can see who built the tool, just see if they've done it without AI ever.

Katrin Ribant: Mm.

Slobodan Manić: The answer is no. You sit, you spare yourself a long thinking session because if that person doesn't know how to do with it without AI, if it's something new that they figured if I just write a prompt and have a wrapper, it'll do it for me. Don't just, just walk away slowly, carefully. But if that person is someone who knows the process and now is using AI to automate part of that process, read on and pay attention to that. I think that that's the first red flag. If it's someone who never has any track record of doing what they're using AI to do for them now because they don't know if it's correct or if it's giving them, it comes back.

Katrin Ribant: It comes back to the same thing. Right. If you don't know the answer exactly. Then it can't help you.

Slobodan Manić: Exactly. I mean, there's just no way. So I think that will be number one. Number two, I don't know if you can try it. Just try and see what it does and Then if you try it again and tell something, tells you something else that's completely different, just walk away. So I guess, I guess those are.

Katrin Ribant: Because again, if you don't know what the answer is and you can't judge how correct it is or not, at least you can correct. You can judge is it consistent.

Slobodan Manić: At least you have that exact. Great point. Great point. So I think those two would be the first two signs for me, the first two things I would look at also if you, if you, the user assessing the tool you need to be using, if you don't know what good looks like, it's try to figure that out first before you use AI or any AI tool or any AI wrapper. Also those thin wrappers that just have a prompt and a ui, no one needs that. It's time for that to go away. Yeah.

Katrin Ribant: Speaking of which, like, amongst many things that you know, I think we both agree on is the fact that learning about AI technologies, not just prompting and context engineering, but like, I mean really understanding how LLMs work and the difference between the models, the applicative ecosystem on top of the model and what the labs are building around that is really how you, you spot when someone is selling your, your hype. And it's really the thing that if I were to tell anybody to invest in is actually understand how models work and what is the model and what is the applicative layer on top of the model. Because even within the labs they're building applications obviously and agents on top of them and they're vastly different in qualities amongst them. I know you have opinions about the labs and what they offer and in terms of like, you know, what those, what those applications are, where would you say people should spend some time?

Slobodan Manić: Digging, I think again, make boring sexy. Again, learn the basics of how these things work. If you don't know what an LLM is, what it does like what a system prompt is versus a user prompt, all those things. If you don't know any of that, just do that first. It will take like a weekend of reading to get the very, very, very basics. That's what I'm talking about. You don't need a PhD in.

Katrin Ribant: You can watch the fluffy videos.

Slobodan Manić: You can watch. Absolutely. That, that, no, that's a, that's a great example. You need to know that. I've had too many conversations with people that take everything chat GPD told them for granted and can you ask it again?

Katrin Ribant: That's fascinating. I mean that's literally really fascinating because I mean, I mean, how do people not Understand what token prediction is like. Literally it is probabilistic prediction because no one told them.

Slobodan Manić: Because Sam Altman is not going to tell them that when he's hyping ChatGPT. That's why. And most people don't need to dig deeper. And look, I'm not blaming them, they just don't need to do that. I'm talking about an average ChatGPT user. So I think humanity was so primed for something like this to come in and just do all the damage that it's doing and will keep doing that. A lot of smart people will be analyzing what happened in, in 2000s for a very long time. That's the way it is.

Katrin Ribant: I'm sure that's true. It's definitively a historical moment. Well, and you know, let's talk about security, like, you know, AI security for a moment. It's a thorny subject. You did an episode of no hacks on ChatGPT, exposing conversations from users for analysts, you know, whose teams are using AI tools and uploading data, even if it's, it seems to be non critical data. What would you say? Like, how would you characterize the real exposure risk?

Slobodan Manić: Huge, Vast in one word. I think you cannot trust these cloud based systems that just with any sensitive data, if you want to have secure AI working for you, it has to be something you fully controlled, so something local, something that's open source, installed on your machines, on your computers. I don't think security and any of these major frontier models go hand in hand and never will because the least bad thing that can happen is them using your data for a next training run. And do you want that? Do you really want something like that?

Katrin Ribant: I think that even if they tell you they don't and you believe them, let's imagine they tell you they don't and you believe them. Working theory, right? They tell you, they tell you. Then though they don't, you believe in them. The way those systems are architected means that it's not just that you cannot trust them, they cannot trust themselves. They do not actually have control over where the information is when it goes across the layers of the transformer. Right. Like it just that that information is not traceable at that point. And so would they want to sort of spend all of these resources securing it? They wouldn't necessarily have the ability and the way to sort of bypass that. If you are a malicious player, there's a very vast attack surface there.

Slobodan Manić: It's extremely easy and especially when you have middleware between the user and the frontier model, there's an app that's a wrapper. Basically. Leaks can happen and exposure can happen in more places than you would like to accept. And I think, you know what I would like to ask all those big tech AI people. And Sam Altman talked about he would not be able to raise a baby without chatgpt. Like he had that it kind of went viral, that quote. I want to ask those that have preteen and teen kids, do you allow them to use your LLM? Are they on LLMs or is that like, you know, wealthy people don't have kids on social media and like responsible people don't have kids on social media. Are LLMs the same? If it's not safe enough for a basic user, it's not safe enough for a business. And I mean another thing that me and Jason Packer and web analyst, another person when we discovered those leaks in November that ChatGPT prompts were leaking into random Google search console account because ChatGPT was using Google to web search and they're using their operator agent, whatever it was that to control the browser, not via API, they literally open a browser, type in the search and give you the results. So we, we saw prompts like I think my boyfriend is cheating on me, what should I do? And stuff like that. How can anyone feel comfortable around this? And now they have a ChatGPT has a ChatGPT health campaign. They want to manage your health. I'm seeing YouTube ads all the time for this. We should trust them. When they prove to us we should trust them and not before that. Like we cannot trust big tech after Cambridge Analytica, after Facebook, after social media. We should never trust big tech unless it spends a decade convincing us we should trust them. So please don't. There's no safety, there's no security in using cloud based LLMs. There are so many levels this could go wrong that it feels like, I don't know, Stranger Things Season 5, which Topical went sideways in many ways.

Katrin Ribant: Yes. Yes it did. Well, let's bring it back to the analyst listening. If AI agents are becoming a real traffic source, how do you actually even measure that today? I mean like technically, right? Like what should analysts be tracking?

Slobodan Manić: I think it's currently impossible to know. There are no tools that will tell you this is an agent. There are IP ranges that you can look at. There are patterns in behavior. So make boring sexy again for the third time. Server logs. Just look at the way they behave and how quickly they go from one page to another. If it goes from One page to another in five seconds and you know, gets all the information and then another page, another page that's unlikely to be a human. So I don't think we have great tools just yet. I'm just like LLM visibility. We are in the pre Semrush and Ahrefs era. There's going to be some tools that tell us that we don't have good analytics tools for agents just yet. I'm sure they will happen especially when they, when they do ads finally in those LLMs, they'll be forced to give us a way to track what's happening. So I think the worst thing for LLMs ads also the best things for LLM.

Katrin Ribant: And so, you know, let's imagine we have some data. You've talked about moving from the attention, an attention economy to like a utility economy. Right. How would you say analysts should adapt their KPIs when success shifts from capturing eyeballs to delivering utilities to AI agents?

Slobodan Manić: Basically, yeah, this is fun because attention get just fighting for attention, you know, when 10, 15 years ago. Our time on page is low. DEFCON 1. Who cares about time on page? Those are stupid. Those were always stupid metrics. You don't want people reading an intro to your blog post before they get to the call to action. Nobody cares about that. So moving to utility economy, capturing that, I, I think it's tracking failure rate for those agents because if an agent is on your website, has a goal, it already knows it wants to complete the goal. Tracking failure rate is really more important than anything else. So how quickly does it do it? How many steps, how many wrong turns does it take along the way tracking those things? And hopefully we get good analytics tools for that, then I feel like they should be obligated to provide the tools, the providers of these agents and models. Once we get that, we will be able to fully track. But I think seven extra seconds on page is less meaningful than 85% completion rate on a task.

Katrin Ribant: Yes, I can see that. And obviously we don't really have the tracking tools yet. Right. But like, let's put ourselves into a hypothetical probably relatively, you know, close future where we have some of those and let's imagine that some of those tools are built the way you think they should be built and they're tracking useful and they're useful.

Slobodan Manić: Right, right.

Katrin Ribant: So for analysts who've built their career on Google Analytics and traditional education analysts.

Slobodan Manić: We feel for you, I'll just say that analysts that have built their careers in Google Analytics, we feel especially GA4.

Katrin Ribant: We do, we do you know, and you know, traditional attribution, etc. If you imagine those tools and the future of analytics with those tools and you know, the traditional aspects, what skills from the old era do you think transfer into the old era? What needs to be learned unlearned? Like how do you see that necessity to transition from one to the other, knowing that the other doesn't really exist yet?

Slobodan Manić: That's a super important question. I don't have, I don't think I have the best possible answer. I think most of it does translate because you're still looking at website usage metrics. I think people who have done server log analysis, some kind of technical log position because session user, you know, abandoned rate, bounce rate. Bounce rate is, there's no such thing as bounce rate for an agent. It's either completing the goal or not when it's on your website. So I think we're looking at different metrics, but the same kind of people in the same toolset still will apply. I'm not 100% confident in that, but I think that's how it's going to be.

Katrin Ribant: You see a return to an era of looking at very granular data and rebuilding KPIs from very granular data, emulating what we think is important in the behavior and in the, in, in, you know, in the sort of like hitting the goals while we're sort of like guessing that this might be important and this other thing might be, might be important. And, and, and so, you know, because that's what we were doing in the beginning. Like in the beginning we thought that time on site was important in the beginning of the web number of paid.

Slobodan Manić: Page views were like the hottest metric in the world. We had those trackers, you know, on every website. That was the most important thing in the early days. And that was.

Katrin Ribant: Well, there also was nothing else.

Slobodan Manić: Exactly, to be fair. Absolutely. But, but, but Internet and we evolved from that into something that was kind of meaningful, not, not perfect. It was never perfect, but it was more meaningful than, than. How do you call them? Stat counters, like hit counters, something like that. I think there are two things here with, with, with doing analytics for, for non human users, let's just call them that because LLMs, AI agents, whatever, whatever there is. There's the standard website that was built for humans and how they interact with that. But there's also going to be, there are new protocols there, there are going to be new ways that they get things done on a website. There's going to be an agent handshake with the Website and then you know, this is an agent, this is what it's trying to do. And then you can do analytics based on that. We're not there yet. Yet. We have 35 years of legacy infrastructure we need to make better.

Katrin Ribant: Before, but also until now, I mean mostly we wanted to filter out bot traffic, not analyze.

Slobodan Manić: That's, that, that's another can of worms. And yes, this is something that I honestly, I don't know what the solution for this will. Because there's so many bots. I mean bot traffic is now, is it 60% or on the Internet? It went above 50% sometime last year and a lot of people were blocking all kinds of bots except for Google and Bing and whatnot. Now we will have actual agents that are not bots that are scraping. You don't need Captcha. If you put Captcha on a website and it needs to check I'm not a bot before it can access your website, what are we even doing there? But there has to, there are some protocols emerging. If the agent can shake hands with the website and introduce itself and say I do this, this is my goal, I want to do this. The website will be able to identify that and you'll be able to track and analyze. We're not even close. We're not even close to this being a reality which you know, in, in how fast everything is moving. It could be three months, it could be a year or three years.

Katrin Ribant: I was just going to say we're not even close. That means it could be here in three months, right?

Slobodan Manić: Yes, yes, yes. So. But it's not, it's, it's nowhere in sight right now.

Katrin Ribant: The good news though is that all of those GA4 analysts, you know, that had to learn to use BigQuery and go back to event level data and have built some of that skill set that's actually probably their best preparation for what?

Slobodan Manić: 100%1 if we still had universal analytics and that's the only thing people know how to do now. Screwed. It's not happening. But GA4 was there because it's event driven. This is significantly closer to what we will be doing.

Katrin Ribant: And if you had to guess, because obviously that's all we can do is guess. Right. Would you think that GA4 is going to be able to evolve its current structure and capture structure to accommodate for analyzing this type of traffic within its current capabilities or minimal evolution of its capabilities?

Slobodan Manić: J5 I don't know. But if there's a tool that's going to be broadly accepted and used, what's the most likely one. It's ga whatever version for sure.

Katrin Ribant: But there's also a situation where you can say, well, first of all, Snowplow might be faster or something along those lines might adapt faster and then you need to do an integration or you can just possibly have niche tools that only do that and then you have to sort of overlay in the early days.

Slobodan Manić: In the early days that that makes a ton of sense and it's probably going to be like that. But if this agentic web and agents sort of being our avatars online and doing things for us online fully happens, I think it will. I think there's no way to put that genie back in the bottle. And when that happens, Google's going to want to have a piece of that pie for sure. And which company has the most expertise to build something like that? This and to make it broadly accepted Also, they still control most of web traffic online. A lot of it goes through Google still. Even if it's AI overviews and whatever else, it's still Google ChatGPT. As hypey as it is, it doesn't matter. Compared to Google, it really doesn't matter. And whatever, everything that Google has in its ecosystem, not just Google Gemini. So I think if there's a tool, it absolutely will be broadly, broadly accepted tool, it absolutely will be something from Google. And why would they have a new tool for this? To just put it in Google Analytics.

Katrin Ribant: Well, you know, I feel like after having spoken about AI for about an hour to wrap up, we should just talk about AI fatigue. You called it like digital tinnitus, I think, which I love that term. I'm stealing it, if I may.

Slobodan Manić: Please, please do.

Katrin Ribant: Yes, I will attribute it as much as I can, but like, like, I really like that. And as a result of like two years of AI hype and pivot or die mandates and things like that. So for like, for others who feel that, who are burned out by the constant pressure to adopt AI and you know, productivity, etc. What's your advice to like protecting mental clarity while staying current. Because you really have to balance the two, right? Like you kind of have to stay current, but you kind of also have to not get caught into the hamster wheel.

Slobodan Manić: And it's really difficult to find a balance. It's really, really, really. Especially if you're on LinkedIn or following something like, it's just, it's nearly impossible. I struggled with this a lot, as I told you earlier. I struggled with. I tried to keep up with everything, the good, the bad. You need to stop caring about it because most of it legitimately doesn't matter and never will matter. If you want to bet on finding the 1% of things that will matter and obsess over it and say I will, I said it first, go ahead, by all means do that. But you will also waste 99% of your time following trails that lead to nowhere. So I think maybe just catching up, find a newsletter that gives you monthly updates or something like that. Or once a month, on the 1st of every month, just ask AI overviews or Gemini or whatever tool, what happened last month? Like, is there something worth paying attention to? Because when they release whatever new frontier model, I mean when ChatGPT releases a new model, Gemini does it the next week and then Anthropic Entropic does it the week after, which means they're just pretending that they're doing something. To me at least I don't see any change. I think just ignoring most of it, ignoring all of it and occasionally paying attention and just focusing on doing your job as well as you can and being as good as you possibly can be at what you do is more important than using AI to squeeze 5 extra percent of productivity or do whatever bullshit.

Katrin Ribant: Well, as you said, the fundamentals still matter, right? If you don't know the answer, you really shouldn't trust what the AI tells you. Especially if you, if it tells you different things with, with different versions of the question. Right?

Slobodan Manić: And it does. With the same version with the same question, it tells you different things. Yeah, most of the time. And I know it's difficult. I know people are afraid. Job security, what is it? People forget about it completely. People feel stressed, anxious. I'm just saying what if most of it doesn't matter? What if most of the stories they're trying to hype and give us are just to make the bubble bigger and do whatever they're trying to do and just maybe it's not.

Katrin Ribant: And that is the characteristic of a bubble, right? You have to keep it going. So that's going to continue for a while. Well, we've talked about a lot and we've covered sort of vast array of subjects from technical foundations to the new data layer to security vibe coding. So for specifically for analysts in E commerce teams, which are probably really at the forefront of this, Right. What's the one thing that you'd like them to think about illustrating? You know, making boring, sexy.

Slobodan Manić: Great question. Another great question. So the specifically for analysts in E commerce teams, I mean, if you can find the three most important things you should be doing and just spend time doing only that for a week and ignoring everything else. What would that life be like? And this is something anyone can test. Just focus on the most important thing and nothing else. Because there's always running in too many directions. Every Single person in 20, 25, 26 is running in all the possible directions. What if you just chose one thing to do for a week and just making it as. As perfect as you can. And this applies to analysts in E commerce, to anyone. But I think if you just have what's the most important KPI for for us, what is everything we can do to make this better in a week or for one week? Maybe try that.

Katrin Ribant: That's great advice. So for people who want to learn more, listen to nohacks work with you on website optimization or accessibility to AI agent. Just follow your thinking. Where should they go?

Slobodan Manić: So no hacks podcast obviously. So nohackspod.com there's also a substack, it's kind of a sibling to the podcast called it's nohacks.subsect.com or my personal website. My first name, last name. I'm not going to spell it.com if you just search for me.

Katrin Ribant: We'll put all the links in the show notes.

Slobodan Manić: Thank Perfect. Excellent. But yeah, the podcast is the best way to follow my work. Get in touch because if you care about optimizing for AI agents, optimizing the web infrastructure and websites for AI agents, guess what I'm talking about? Nothing else this year. This is the only thing I care about.

Katrin Ribant: Great. So we can learn all we need to learn about it there. We'll put all those links in the show notes, including our flu fees, episode on GPT and commerce. For anyone who wants a lighter take on why brands become invisible to AI. And so that's it for episode eight of Knowledge Distillation. Sani. Thank you very much.

Slobodan Manić: Thank you so much.

Katrin Ribant: And if today's conversation made you think about how AI is changing data analytics, visit us at Asky AI and try Prism, our platform helping analysts navigate complexity with context. Thanks for listening. And remember, bots don't think AI analysts do. Thanks to Tom Fuller for the editing magic on this episode. If you want to work with Tom, head to askqui AI and check out the show notes for his contact info.

Want me on your podcast or at your event?

I speak about optimising websites for AI agents, CRO, and web performance.