How I AI

“I haven’t written a single line of front-end code in 3 months”: How Notion’s design team uses Claude Code to prototype

52 min
Feb 23, 2026about 2 months ago
Listen to Episode
Summary

Brian Lovin, designer at Notion AI, shares how Notion's design team uses Claude Code and a custom Next.js-based Prototype Playground to rapidly prototype AI-powered features. The episode demonstrates how AI-assisted coding tools enable designers to move from Figma mockups to functional prototypes in production, with emphasis on teaching Claude to solve problems autonomously rather than asking for manual intervention.

Insights
  • AI-assisted coding fundamentally changes design workflows by enabling designers to prototype in production-like environments rather than static design tools, revealing real interaction problems earlier
  • Teaching AI tools to solve problems autonomously (via skills, scripts, and MCP tools) reduces friction and extends productive sessions without manual handoffs
  • Designing AI products requires code-first prototyping to understand actual model capabilities and failure modes—Figma mockups cannot capture conversational or agentic behavior
  • Shared team infrastructure (Prototype Playground) with global and local Claude rules creates collaborative AI-assisted workflows while maintaining individual customization
  • Context quality directly impacts AI output quality; tired/lazy prompts produce poor results, making prompt discipline and rest critical to effective AI collaboration
Trends
Design teams adopting code-first prototyping workflows as AI product complexity increasesShift from designer-engineer handoffs to designers directly building interactive prototypes with AI assistanceMCP (Model Context Protocol) tools becoming essential for connecting AI agents to design systems, Figma, and development environmentsSlash commands and skills emerging as abstraction layers to make AI tools more accessible to non-technical team membersAutonomous verification loops (browser testing, CI monitoring) reducing manual QA steps in prototype developmentDesign systems and shared component libraries becoming prerequisites for efficient AI-assisted prototyping at scaleNotion AI and similar agents becoming design tools themselves for prototyping and ideation workflowsGrowing recognition that model capabilities require continuous designer education and experimentationInternal tool development (like Prototype Playground) becoming competitive advantage for design-forward companiesIntegration of voice-to-prompt tools (Monologue) reducing friction in AI-assisted coding workflows
Companies
Notion
Brian Lovin works as a designer at Notion AI; the company's design team uses Claude Code and Prototype Playground for...
Anthropic
Claude and Claude Code are the primary AI tools used throughout the episode for code generation and design prototyping
Vercel
Prototype Playground is deployed on Vercel; mentioned as the hosting platform for the team's shared prototyping infra...
Figma
Used as primary design tool; Figma MCP integration enables automated code generation from design files
GitHub
Used for version control and CI/CD workflows; slash deploy command automates GitHub PR creation and branch management
OpenAI
Mentioned as user of WorkOS for enterprise authentication; competitor in AI coding assistant space
Perplexity
Mentioned as user of WorkOS for enterprise authentication; competitor in AI search/research space
Cursor
AI-powered code editor used alongside Claude Code; offers hover-based fix suggestions and agent mode
People
Brian Lovin
Designer at Notion AI; demonstrates Prototype Playground setup, Claude Code workflows, and design-first approach to A...
Claire Vell
Host of How I AI podcast; product leader and AI enthusiast conducting the interview and providing design perspective
Quotes
"The way I think about designing B2B SaaS is you want your designs to encounter reality as early as possible."
Brian Lovin
"I haven't written a single line of front-end code in 3 months"
Brian LovinEpisode title
"Every time somebody is like a little anti-AI assisted coding, I'm like, do you know that I used to have to walk uphill both ways for my CSS? It was not fun to do this."
Brian Lovin
"Anytime the AI asks you to do something, you should, before responding, try your best to see if you could teach the AI to answer that question for itself."
Claire Vell
"I think as more and more people are designing apps that both are for AI or incorporate AI in some way they're gonna need some other like native code first way of working to actually understand what the models can do"
Brian Lovin
Full Transcript
The way I think about designing B2B SaaS is you want your designs to encounter reality as early as possible. I've always been into prototyping, and then all of a sudden these AI coding tools come along, and now I can prototype faster, I can prototype in production. So explain to us what this prototype playground is. It's just the Next.js app. All of our prototypes are in one place. Seeing what other people are working on is really fun and interesting. And oftentimes you spot cool ideas and you're like, ooh, I want to try that. The code is all in one place. It's just in one repo. And so I can just yoink cool ideas from other people's prototypes and incorporate them into mine. Every time somebody is like a little anti-AI assisted coding, I'm like, do you know that I used to have to walk uphill both ways for my CSS? It was not fun to do this. I mean, even just sitting here watching this, I still just find this magical. Welcome back to How I AI. I'm Claire Vell, product leader and AI obsessive here on a mission to help you build better with these new tools. Today, we have a designer-centric episode with Brian Lovin, designer at Notion AI, who's going to show us how he set up a prototype playground for the entire Notion design team to vibe code using CloudCode any prototype they need. This is a great one for someone looking to either shift their design organization into a code-first prototyping mode or learn some advanced techniques with cloud code. Let's get to it. This episode is brought to you by WorkOS. AI has already changed how we work. Tools are helping teams write better code, analyze customer data, and even handle support tickets automatically. But there's a catch. These tools only work well when they have deep access to company systems. Your copilot needs to see your entire code base. Your chatbot needs to search across internal docs. And for enterprise buyers, that raises serious security concerns. That's why these apps face intense IT scrutiny from day one. To pass, they need secure authentication, access controls, audit logs, the whole suite of enterprise features. Building all that from scratch? It's a massive lift. That's where WorkOS comes in. WorkOS gives you drop-in APIs for enterprise features so your app can become enterprise-ready and scale up market faster. Think of it like Stripe for enterprise features. OpenAI, Perplexity, and Cursor are already using WorkOS to move faster and meet enterprise demands. Join them and hundreds of other industry leaders at WorkOS.com. Start building today. Brian, welcome to How I AI. What I am so excited about in terms of our conversation today is you're going to show us about how one of the best designed products out there, Notion, is being designed by people like you using these new AI tools like Cloud Code. So why did you make this shift to how you were doing design, what it meant to prototype design and build things, especially for a product and in a company who values design so highly? The way I think about designing B2B SaaS is you want your designs to encounter reality as early as possible. And, you know, if you imagine this gradient of like, I'm scribbling on a napkin on one side to I'm shipping to production and showing customers on the other side. Our goal as designers is to move up that gradient towards prod as quickly as possible. So I'd say for most of my career, I've sort of biased towards being interested in programming mostly at the prototyping level i just find that when you're designing something in figma and then you actually try it in the browser in the browser you notice a ton of problems you know all of a sudden you're clicking things you notice loading states you notice oh that didn't quite work on this screen size so you encounter some version of reality sooner and you end up getting to a better design more quickly so you know i've always been into prototyping and then all of a sudden these ai coding tools come along and now i can prototype faster i can prototype in production i can uh or what i most often do now at notion is just prototype in a little internal tool we've built called prototype playground and again the idea is just like how do we get something that's somewhat realistic in a kind of real environment in our case the browser uh as quickly as possible. And I think that just helps you move faster and end up with better designs. So explain to us what this prototype playground is and how you set it up and how you might use it. Okay. So prototype playground is nothing magical. It is just a Next.js project. So actually here to source apps and there's an app directory. And you'll notice here in our app directory where normally in a next js uh app you would see pages um well we've just namespaced every designer on the team or pm or engineer whoever signs up and wants to use it can just namespace some directory so here's brian and then every directory inside of that is some prototype and so it's just the next js app but each page is sort of a standalone there's no global layout there's no global i don't know like structure that you have to adhere to and so what that looks like on the front end is this this is what we call prototype playground and it's just a list of prototypes ordered by who is working on stuff recently so here's a few from december and then a bunch from november and it's really cool because having everybody's prototypes in one place is useful in two dimensions one just from a visibility point of view seeing what other people are working on is really fun and interesting and oftentimes you spot cool ideas and you're like oh i want to try that and then on the other dimensions like if you spot a cool idea and you want to try it the code is all in one place it's just in one repo and so i can just yoink cool ideas from other people's prototypes and incorporate them into mine usually usually by just telling claude to do that uh yeah i think before prototype playground there was a lot of designers at notion who prototyped in code the difference was we were all creating our own repository our own next js instance and so it was hard to know where everyone's stuff was everyone was rebuilding it in different ways or if people were trying to recreate something that looked notion-y we were all doing it from scratch so anyways prototype playground next js app all of our prototypes are in one place and then we have a few shared components and shared styles so if you want to make something that looks notion e you can do that pretty quickly so for example we have some templates here i can show you like notion ui is just a notion e sidebar and actually this isn't even very notion e i think at some point i slipped this new button in here which obviously doesn't exist in the the the product i don't think these things do anything but it's close enough if you're like oh i needed to prototype something with a notion sidebar i can just come in here and duplicate this template and then we of course pull in a bunch of our colors typography icons so that again just getting to close enough notion styles uh without a whole lot of effort yeah and i want to call out for people a couple many episodes ago i showed how you could build a very similar next js app for yourself that had a combination of docs you were working on markdown docs you're working on and prototypes in a very similar format where it was like here's my folder of just stuff i'm working on very minimal shared components, very minimal shared styles. I like this too, because it's nice to have that team level organization. So you can pop in and see who your teammates are working with. I have a question from an operational perspective. Did you set this up? Was this a passion project for you? Did engineering set it up for you? How did this actually get created? Yeah, I set it up with another engineer. I mean, it's just the NextJS app but then operationally just a few approvals it's deployed on Vercel so we had to like go through a little bit of process to get that project spun up get a few people added as members otherwise yeah it's not that much again it's just a pretty basic Next.js app which you can literally use Claude to like help me make a Next.js app and it's just going to get you the default I like you know I like keyboard hands everybody does the same keyboard hands motion where it's just this. I have one more question, which is, of the people now working in this repository, how many before were working in code versus this is their first repository that they've cloned to their desktop or deployed? Was the design team pretty technically adept already, so this was very natural, or were there some people that needed to be onboarded? I think so. I mean, to be honest, Prototype Playground is still really for me. I think I use it the most you can see here there's a bunch of other people that that are creating things but if you were to go through i probably use it the most i think there's maybe like five to ten people at notion that use it quite a lot and then a bunch of people who either have tried it and it didn't stick and we can get into two reasons why that is or they're just not interested or they prototype separately right like we still have people prototyping in figma we have people that prototype in their own code base still they just prefer their own stack maybe they don't like next js maybe they don't like react so they do something else and i think all that is totally fine in fact one of the the features i added recently was this ability to link to an external prototype so if you prefer using v0 or lovable or a figma make file whatever it might be you could just link to it here and in fact this is what it'll show up as in prototype playground just have this little external icon and so you can click it and it'll open in a new tab so in theory this could be the prototype playground or repo for any prototyping tool my hope is that over time we make this thing useful enough that more people will want to prototype in it because it's just faster than those other tools and we got to figure out how to lower the the onboarding complexity for people who aren't technical before. So to answer your question, I don't know, I'd say some people who weren't technical made their very first code prototypes or like AI assisted prototypes in the playground. But probably the majority of us that are still using it daily had some technical background. Got it. Perfect. Well, let's prototype something. I want to see how it actually works. let's do it um okay so there's a few ways to make new things in xjs right like we could be in cursor and we could come in here and create a new folder and create a bunch of page.tsx and metadata files and that sucks i don't want to do it um so there's two ways around that the first is when you're running in localhost you can actually just click this button that says new and you give your prototype a name and a description. I'll call this one howiai. And then this is for fun. And I create that. And all that's doing under the hood, if we bounce back over to cursor, is it just created those files for me on my computer. This is my favorite part is like, there's no backend for prototype playground. It's just all files on disk. And then we can just push all this to GitHub. So here we have like a little metadata file. These get sort of collated to render the list on the homepage. We have an actual prototype file here with some code and then this is kind of nice like it automatically gives you a button to open it in cursor. So now I can just come in here and start prototyping. Now typically my workflow is I just bust open Claude in the terminal. I know this isn't how you're supposed to use cursor but it's just it's just how I do it it's probably not even how you're supposed to use clod code but I just do it we're just equal opportunity offending these two tools I know I know sorry everybody um but this is how I like to work um and in fact I have a little shortcut here where I can just press uh caps lock g and then I can get these two things side by side in my computer so I usually um clodding over here reviewing the changes here and then monitoring sort of the output over here so let see here I want to make a prototype and I don know let just come up with some contrived example Like maybe you can help me think of a good use case Can we make a prototype for oh, like a little, a video and audio. This may be complicated. A video and audio like display module for my podcasts. Video and audio. So it's like video and then maybe like an audio player. Let's see. You know, it's Opus 4.5. I think you can do it. Okay, let's try it. So normally, let's walk through like my actual workflow. There's sort of two steps. One is you can type a lot. That's not that fun. I do use this tool called Monologue where you can just talk to your computer. There's many products like this. I think Monologue is just nice and cute. So we can just talk and it's just much faster than typing our prompt. The second thing you'll notice with Cloud Code is I switched over to plan mode. I think it's really, really important to plan before doing anything. For whatever reason, you just get better outcomes. now the key thing about using plan mode is to actually read the plan and i think this is where having a development background just gives you an edge because you can read the plan and be like oh that part actually doesn't look quite right whereas if you maybe don't have as much programming experience it would be harder to tell that but in either case i still find that having the plan mode and creating some structure before actually writing code is better so let's just do both of these things at the same time so we're in plan mode and i'm gonna invoke monologue here and it's recording and so let's say i want to build a new prototype in this how i ai directory and we are a podcast and i want to build a detail page for a podcast episode that has both a video player and an audio player underneath the page should have the title of the episode description and how about if you hit play there's little confetti that shoots up out of the player and so we end that and now I will delete this and we plan so I have to give you props on two things one I am also a plan mode slash like write your spec, write your PRD person, obviously. I think the second thing is I am still just such a read the code, read the outputs girl when it comes to AI. It's actually one of my challenges when I use something like Claude Code or watch people use Claude Code is if you don't do it inside a cursor or something that gives you this sort of, I love your three pane window, your like code window, your Claude window, your output window, because I see people with like 17 tabs of quad code going, just accepting a bunch of changes. And I have to read. I think this is also just like engineering development background where you can just spot things that make no sense in the moment as opposed to having to go back and debug something. So I am very much aligned with you on that. Yeah, it's helpful. And you know, this is probably obvious to a lot of people who are familiar with using cloud code but maybe if you aren't like another piece that's really important here is getting the right context up front right like we just typed in some prompts um but under the hood i can show you we actually have uh some other files helping us out here so we have a clod.md file at the root of our project with just some rough instructions around like the tooling that we use like we use bun we use tailwind it has like a rough outline of the project structure another thing that we do is anytime someone runs the project locally we create a clod.local.md file and that local file is not committed to the git repo so it's personal per computer and it adds a little bit of extra context like hey this is my username in prototype playground It tells Claude where my directory is. And it gives some instructions like, hey, you know, don't go around and mess with other people's files. Like prefer to work in my directory. And a little bit more about the workspace structure and how like individual projects can be built. So a couple of those things are working under the hood here. And while you're accepting some of these Claude code changes and questions, I do want to call this out for folks. because I think people are pretty aware of the Cloud MD global settings, but I think people forget that there are actually locally scoped versions of these that you can implement. And so it's really useful to get one version deployed to everybody that gives you your master rules for using Cloud. And then you can set up your own custom one with your own particular preferences. And I think that's a really nice way to create a good collaborative environment where people are using a similar AI tool or agent to work in the repo. Totally. Yeah. Okay. I don't know. We'll see how this goes, but it's going to install some sort of confetti. It's going to have a player, audio player. Look at that wireframe. This is really awesome. Like it does a wireframe in the plan, which is crazy. And here, I don't know. We can just kind of skim this for the sake of this example. This looks fine. So let's auto accept edits. Now, I have a tip for people because I think when you spend enough time on Twitter or watching other people use these coding tools, people are always like, how do you get it to run for longer? Or, you know, they find themselves constantly getting stuck or the agent does the wrong thing or it's asking for their input. And my philosophy on this has been any time the AI asks you to do something, you should, before responding, try your best to see if you could teach the AI to answer that question for itself. There's a good example. Oh, wow. That was very fast. Ooh la la. Well, here, let's hold on that and see if the confetti works. Well, actually here, the example is I've already taught Claude to like always lint itself after it's done, right? Like what's really annoying is when it builds a bunch of stuff and then you go and look in your browser and there's some error, right? So for example, I've taught Claude, hey, check your work. One, you can run commands like, what was this, like eslint, right? And like look for actual TypeScript errors. The second is you can give it access to MCP tools. So Playwright is one, the Chrome DevTools MCP is another one. And you can say, well, actually, you know, before installing this, Claude would say to you, hey, I've implemented your feature. Go take a look at it and let me know what you think. And remember, our rule is anytime Claude tells you to do something, ask if you can teach it to do that thing for itself. So I don't want to have to look at the browser every time to see if I did it correctly. so instead I teach Claude actually you should be the one to go and open the browser so it knows how to launch Chrome it knows how to navigate here it knows how to click the play button look for confetti make sure the audio is working all that kind of stuff and so now we we were able to run this task for much longer without my input and actually get to something that is working well I'm actually very impressed with this prototype it's much more lovely than I thought it was going to to end up much more robust and the confetti looks great the confetti looks great yeah well here i'll show you another example this is i think where the power of mcp gets crazy so let's let's actually clear this we're just going to start start a new new conversation here i'm going to just totally undo everything let's just start from scratch so uh a couple other things that built in you know i think uh remember like i'm trying to make the onboarding flow as simple as possible for people on my team yep um so what claude has is called slash commands and you can just build these yourselves and they're basically glorified prompts but they can also run scripts and so we have some slash commands in the project that help people get going really quickly so i have one called create prototype and then you can give it an optional name so we'll call this one how I AI. And that's going to do the same thing as clicking the new button on the browser, which is what we did earlier. The difference, of course, is I don't have to click things. I kind of want to design this so that I basically live over in the terminal. And can you show us really quickly in your repo just how these commands get defined? Perfect. Thank you. Yeah, sure. So again it's basically a glorified prompt it has a name a description and then some instructions so in our case we say kind of how to come up with a name based on what the the user provides tells it where to look to determine the current user's username how to create the new thing it actually provides some sample code to use for both creating the page and the the metadata file I think I need to also prove this. So it goes, let's just do blank for now, as well as creating the metadata. So AI is better with good context, but it's also really, really good if you just provided examples of how to do things. So the reason it's important to provide these code snippets is to show it what success looks like, right? If this was like just instructions to create blank files, it wouldn't know what to create. So in our case, we're just showing it an example of success. And we could probably simplify this. It's actually quite a long command, but here we go. So it created this and a blank piece of text. That's great. So that's just one way to start. You just type slash create a prototype and then that'll create. but maybe we have some design in Figma and we want to build this. This might not work, but let's try it. So we can connect to the Figma MCP and I can just copy a link to this frame and say like, let's build this, this notion UI. So before you could just paste a link to a Figma URL and try and manually invoke the Figma MCP. and it would sometimes ask clarifying questions and sometimes it would build it and then sort of stop halfway through i don't like any of that so we actually built a command called slash figma and it roughly does a couple of things the first is it actually checks that you have the mcp server installed and running you know for people on the team who have never done mcp stuff before they might not know how to do this and so it detects if you have it installed and if it doesn't if it finds that it's not installed, it'll just teach you how to do it. So it actually returns instructions to the user on how to set all this stuff up. And then it moves on to phase A, designing or extracting the design from Figma, then it'll implement it. And then the most important thing is we enter this third phase called the verification loop, where it's going to open the browser and compare the implementation it created to the original Figma file. And I think my instructions are basically keep looping until you've gone through like two loops where there were no more changes oh yeah here repeat until the implementation matches or after three iterations with no changes and then stop iterating so let's just see what happens this i would say it gets like 80 correct 80 of the time but that's just that's just how ai is right now i was gonna say about 60 percent. So I think I was right. Well, actually, you know, I think it is 60 percent, but this command and this loop and these instructions and like the pairing of the two MCPs actually gets us to 80 percent. I want to call this out for folks because one of the things that I find most frustrating using MCPs, even as a fairly sophisticated user, is one, you just have to use these like magic keywords to invoke the MCP and the right tool and the right thing. And, you know, sometimes I have one of the challenges I have is I have a lot of MCPs that use the same tool names because so much across SAS is named the same like everybody has the concept of projects everybody has the concepts of pages or documents and so I like this idea of like force invoking a specific MCP via a slash command and not even just force invoking that specific MCP but force invoking a specific set of tools in that MCP Super super useful And then I will give you props for the instructions at the top that teach somebody if you have no idea what you doing here how do you even get this thing installed? That's such a nice piece to add in as user experience for a consumer of this slash command that might not be you. And so that's something that people should really, really think about. Yeah. Yeah. I would say also, it's funny because I've actually watched a bunch of these videos and looking even back at the ones from six months ago, it's crazy how far the tooling has come. And so I imagine that people who, for whatever reason, might be watching this video in six months will look at what we're doing here and be like, oh, how naive. You know, we've come so far. MCP is no longer a thing or something like that. Right. And I kind of feel that way now where mcp is it's like not the best thing but it's the best we have so far right like it's very context inefficient sometimes it runs forever sometimes it yeah it just like blows up your context window but it's the best we have right now so even just watching this right like here's our design that got built this was literally just pasting the link to the figma file no other custom instructions and now over here on the right it should be uh i think i ran into an issue earlier yeah but something got busted with this let's try the chrome dev tools mcp again i think i quit it midway through because it was detecting some conflict with the window but anyways this is pretty good by default and then from here i would iterate you know some things things you might notice would be like there's no hover states some of these images are broken but those are just easy follow-up tasks. Well, and you're doing this from a kind of design perspective, but think about how many engineers sit there and like pixel pull over Figma prototypes into the front end. And, you know, if you have a great design system, maybe that's easier to do, but it's not what the 27 seconds that we just watched to scaffold stuff out. And so I just think, you know, the friction reduction in these, you know, asset to asset handoffs, which for my entire career, 20 plus years in tech have been the most expensive part of implementing something where a designer gives you a design and then you have to get into the front end or the front end has to be hooked up to the back end. All those little pieces can be smoothed out and done much faster. And then you can spend the time on the optimizations the performance the how it feels how it how it works and i think that's just really it's really fun from a builder perspective totally it's so fun and yeah i mean even just sitting here watching this i still just find this magical right like now that it's using the the chrome dev tools mcp they like looped and fixed the broken images and like created this checklist of stuff like okay everything appears to be right it's got this bottom bar these things are obviously wrong but we could go and fix those with the follow-up prompt but again the goal is like can we get 80 percent in literally one prompt i just pasted a link and it just iterated itself towards something that's roughly complete i know and every time somebody is like a little anti-ai assisted coding i'm like do you know that i used to have to walk uphill both ways for my css like yeah it was not fun to do this like i i find this just mesmerizing. This is so cool. Mesmerizing. This is great. This episode is brought to you by Orcus, the company behind Open Source Conductor, the platform powering complex workflows and process orchestration for modern enterprise apps and agentic workflows. Legacy business process automation tools are breaking down. Siloed low-code platforms, outdated process management systems, and disconnected API management tools weren't built for today's event-driven, AI-powered, cloud-native world. Orcus changes that. With Orcus Conductor, you get a modern orchestration layer that scales with high reliability, supports both visual and code-first development, and brings human, AI, and systems together in real time. It's not just about tasks. It's about orchestrating everything. APIs, microservices, data pipelines, human in the loop actions, and even autonomous agents. So build, test, and debug complex workflows with ease. Add human approvals, automate backend processes, and orchestrate agentic workflows at enterprise scale, all while maintaining enterprise-grade security, compliance, and observability. Whether you're modernizing legacy systems or scaling next-gen AI-driven apps, Orcus helps you go from idea to production fast. Orcus, orchestrate the future of work. Learn more and start building at orcus.io. That's O-R-K-E-S dot I-O. Are there any other commands that you think are super useful? Yeah, yeah, I can show you a couple. So I want to scroll back up a little ways. Actually, there was this step very early on where you can see it was running over and over again this skill called bunrun clod skills find icon what's that well if you look over here in our design we actually have a bunch of very notion specific icons right like we have this ai face we've got home inbox we have all the icons in our project the problem is ai is really bad at estimating what the name of an icon should be, or rather it uses like the most obvious name possible, which doesn't always match what's in code. So for example, like this face icon, there's no way AI would know what we call this. Or a very common one is it will, if you have like a search magnifying glass, right? It will just assume that it's called search icon, when in fact, in our code, it's called magnifying glass icon. And so this icon hallucination was getting really, really annoying. So I wrote a little skill called find icon. And the skill basically says, like, anytime you're going to implement an icon, first go and actually look through the whole project, but also look for synonyms or closely related words to the icon. So if you're going to look up something called search icon, also try search for magnifying glass icon. and it actually wrote a typescript script to do this to just iterate through all of the the files in our icons directory which is like 5 000 right it's a lot so it'd actually be very inefficient for it to try and load all that up into context it needs to write itself a script to do more effective searching so in that that loop here um yeah you can see it like found it looked up magnifying and found the magnifying glass icon it looked up inbox and it looked up gear and trash in order to get all these things correctly now this only this skill had to exist after all of us on the team just got really really frustrated with it hallucinating over and over and over again it's sad because it obviously missed these bottom three it didn't get them correct but the fact that it got these on on the first pass is a huge step up so the way i think about it is you know we have these commands that you run manually and skills are these capabilities that the AI should detect automatically and sort of use at the appropriate time and it'll know to do that based on the description and title you've given it so in this case find icon and then how to search for icons and of course the best part is just letting it do things programmatically on your computer by calling actual coded scripts so this was really helpful saves us a lot of time and just fixing imports and nope search icon does not exist those kinds of annoying knowing steps well what i like about this uh is one this is exactly what you would do to like a junior designer or engineer onboarding you would like explain you'd be like sometimes we call it search but not really it's magnifying glass you just got to go find like the closest synonym and the ability to be able to describe that to an agent or a skill or a tool and then let it do it programmatically for you this really useful. We do have a How I AI episode on Claude's skills, but one piece we don't go into in detail, which I think is really important, is Claude's skills can be bundled with scripts. And so the ability to combine both natural language prompting, which is in the skill.markdown, with a set of programmatic tools in terms of scripts is a very powerful combination, and Claude's very good at making these. So you can have Claude's. Like all this, like I did not type a single line of code in this, right? Like this is 100% like, hey, I just need this problem to be solved, create a skill for it. And in creating that skill, also create a script so that you can work more effectively. Like this is 100% prompted. Show us your last command, because I think this is a really useful one. Okay, this is fairly new. I think I merged this last week. Going back to sort of the problem with Prototype Playground, it's still a Next.js app. It's still React and TypeScript and Git and branches. And it's just a lot of concepts to throw at someone who maybe is used to only prototyping in Figma or they're intimidated by a terminal or code. And so I'm trying to figure out, like, how do we make this thing more approachable? How do we make it easier to onboard, but also not dumbed down, right? Like, I want people to learn how to use computers. I want people to even subconsciously absorb the ideas of git and branching and pull requests and merging so I don't know the best way to do that but my first attempt is to create this skill called or this command called deploy and deploy does basically two things the first is it like goes through prerequisites and makes sure that it has the github cli tool installed on your computer and that you're authenticated. And if you're not, it like walks you through those steps, how to do it. And then the second step is it will just walk you through step by step how to get this prototype you've just created deployed so that you can share the link with someone on your team. Let's see what happens. I'm going to try it now. I'm going to hit deploy and we'll see what happens. There's a couple of really cool loops in here that I think save people a lot of time. So we can see it going through the prerequisite steps here. It's making sure I'm logged into GitHub. Now, the first thing here, look, it's looking to see if I'm on a Git branch. It notices I'm not. I'm on main. And it shouldn't be doing that, right? Like, we never want to push to main. So I think what it should do is help me create a new branch. And we'll see if it actually does it correctly. it's also trying to find some typescript errors and it's going to run some tests i basically told it to do all this stuff because it's really annoying if you push code to github wait for all the checks there to pass if they fail then you got to come back to your computer fix stuff okay great so it created a branch now it's staging our changes branch name perfect creating the commit it i'll give i love this this is a great idea i will also give my just like hack to learning git for anybody who hasn't used it i just love the git git hub desktop app it just like it gives you buttons for all this you can see your diffs you can like create branches with buttons so i think this is awesome and if you are intimidated by the command line there's like literally a beautifully designed desktop app that you can that's true it's pretty nice well now check this out so it's created the pr and in the instructions i've told claude hey whenever you create the pr open it in the user's default browser so now we have our pr opened here and uh this check to deploy it to versell will fail but that's okay because i give it one more step here and all this red looks scary, but it's not. I tell Claude to just monitor the CI every 30 seconds or every 60 seconds until all of the checks pass. And I tell him the specific checks that I care about. And if any of the checks fail just fix yourself and then push the changes So you know if people push something to GitHub and there a TypeScript error they see some error over here in the GitHub UI They take a screenshot They send it to me on Slack and be like why is my thing not working? I want to just avoid that entirely. And, you know, going all the way back to my first principles, like if the AI is asking you to do something, like check the PR or tell me the CI status, you should really be thinking about how do I teach Claude to just do that for itself so over here this slash deploy command literally is just end to end i just sit back and watch it loop over and over and over again checking its its commit status its ci status making sure everything works and then when all of the check marks over here are green the script will stop i think this is pretty awesome i i feel uh i hope it lowers the barrier and like the intimidation factor of having to learn all these tools. But at the same time, you know, if you are curious, you can just sort of read along and understand what's happening. It's like instructed to communicate in clear English what it's doing. My favorite part of this, and it's not going to be what people think. I think the slash command is amazing. I think running through all the pre-projects, great. I love that you just open it up in a browser window. It's one of those things that, you know, even if you created the branch created the pull request said it was ready to go people are like okay well now now what do i like now what do i do and just forcing open the browser window and saying like this is where it lives on github my question is do you have to get your code reviewed in prototype playground or do you just for prototype playground no i mean people can always ask for it but no we pretty much just yolo merge i think the thing that i mostly check for is like did my PR accidentally mess up someone else's prototype? Yeah. But again, like that happened a couple of times and that was annoying. So then we created this Claude local file that's like important, do not do this, you know? And that seems to have fixed the problem. So yeah, a lot of YOLO enable auto merge. And of course it's not perfect. I don't know. It seems to be hallucinating some stuff here. Like it thinks these passed, even though they haven't. I don't know. it it's close so i'm just gonna zoom out everything that we went over you created a shared repo for your entire team where you could have name level directories no database we're just using metadata json and and and shared code to put different prototypes inside this repo you set it up with both global clod rules as well as local clod rules plus clod commands and clod skills to sort of guide people along common paths. My favorite one is going to be Figma to code. It's so beautiful. It's so good. And then the number one rule that I've heard from you today is when asked to do something by Claude, teach Claude to do it. It's, it's, yeah. So you have this amazing prototype playground. You've set all this stuff up. How has, let's just do a couple lightning round questions and get you on your way. And my first one is, how has this shift from doing things, you know, maybe exclusively in Figma or in these lower fidelity prototype models to really leaning on things like cloud code, code-based prototyping? how has that changed the design team? Has it changed a small part of the design team? Do you feel like overall things in the organization are shifting in a way? How do you feel like it's changing the way people work together? I still use Figma. I probably still spend 60 to 70 percent of my time in Figma. You know, there's just certain things that you're making that don't need to be in the browser. They don't need to be coded up. You can just look at it and be like yeah that's roughly right we should just ship that i find that as you're designing for things that use ai that is not true though so for example if you are building a chat bot or in my case i work on notion ai i don't think you can design a good chat experience in figma you can design what the chat input looks like. You could design a little chat bubble and a send button and like a dropdown for model picker. I think all that's fine in Figma. But what you can't design in Figma is what it actually will feel like to use that thing. I probably should have said this at the very beginning, but the reason Prototype Playground existed is because when I started working on Notion AI, I was literally designing conversations in Figma. You know, it was like, the user's going to say this and then the AI is going to say this and then it's going to work perfectly and create a page or a database and like you mock these golden paths in Figma and then the engineers go and they build it and then it just doesn't work that way right it you send a message the AI gets stuck or it asks a follow-up question or it does the wrong thing and you need to correct it and prototype playground was for me a way to connect to real ai models and just start feeling out like okay how are the models going to work if i submit this kind of prompt what happens if i connect it to the notion mcp doesn't even know how to create a page what happens if it runs into an error oh right we need to design an error state for this what happens if the model is thinking for two minutes and the user's staring at an empty chat screen like what should we do in that intermediary time to help them feel confident that it's working, that it's doing the right thing? Is there any way to show incremental progress? I just found those things very, very hard to design in Figma. So to go all the way back to answering your question, I think as more and more people are designing apps that both are for AI or incorporate AI in some way they're gonna need some other like native code first way of working to actually understand what the models can do it feels honestly kind of bad it feels like a lot of wasted time where every month the whole freaking industry has to learn like oh what are the new capabilities of this model 4.3.2 max pro and then a month later it's all irrelevant because the new has come out and then you learn that it feels like a waste of time unfortunately i think it's necessary because the model capabilities are still advancing quite steadily with each release and it's really important as designers to understand what models are capable of doing so that we can create product experiences and designs that sort of live right at the edge of what the model is going to be able to do well what's really frustrating is if you design something that's like you know oh a user is just going to ask for a cool website and it's going to be this perfect output website on the other side models can't do it right or or they require a bunch of fine tuning and and sort of like intermediary prompting to get that right designers just have to know what's going on under the hood there to to design something that's plausible and possible So I suppose the more products incorporate AI, the more designers will have to shift to thinking sort of prototype first, but probably prototype first with actual code under the hood, where you can incorporate modern models and see where they break and see where they're good and see where they're bad and actually form an opinion about which models are good for which things, that kind of stuff. So speaking of which models are good for which things, and you're using my current fave, my babe, Opus 4.5. Why Claude Code? Why Cursor in this non-cursory configuration? Tell me how you arrived at this is your tool stack. I need to play with more of the Cursor stuff. I actually think Cursor agent mode is pretty awesome. I've clearly tried it a little bit. I just haven't gotten that that far the thing that I still really appreciate about cursor I actually technically use both like if I have I don't know like some file and there's there's like some error here I still really appreciate being able to just hover over the error and there's a button that says fix and chat yeah that's still faster than like copying and pasting it down into to cloud code so I actually use both a little bit I just think cloud code does the best work I don't know how else to describe it there's this weird feeling as you use all the different models for different things it's like to different people they just feel right for me opus 4.5 is just insanely good at doing what i want and i like the way it approaches problems i like the way it plans i like the way it executes i like the way it communicates back to me and the follow-up questions it asks and then you know why not use opus 4.5 in cursor versus in the the terminal ui i don't know i think this is just purely personal preference like some people look at this and they're like this looks like shit like give me buttons and ui and components and and drop downs and things like that and for me i don't know this just feels nice and easy it just feels good as our friends over at every say each model has a mouth feel yeah yeah exactly and Claude Code and Opus have a good one okay and then my very last question because you are you seem like a expert prompter but when AI is not listening when it's not listening when it makes up you know CI checks that passed where it didn't it didn't actually pass what is your prompting technique basically I noticed there's a direct correlation with how good of things I can make and how tired I am. And if I ever get to the point where, man, Claude just sucks, it's doing the wrong stuff, and I go back and I reread the thing that I said, I realized I made no sense. And so the best solution for me to write better prompts is like, go to bed, try again tomorrow. Which, I don't know if that's a cough out answer. It's not actually writing better prompts but uh you know your your output's just directly correlated with how good of context you give the thing and if you're giving it sleepy tired lazy please fix this type uh commands it's going to do bad work i don't know if this is what you intended but you gave me very good both relationship and parenting advice there which i'm thinking about i was trying to ask my kid this morning to do something, I'm pretty tired. And I clearly, the inputs, we're not going to get the outputs that I want. Well, it's easy. I mean, just go take a nap. Can't you do that at any point that you need? I love that. You know, one of my favorite little agents, Devin, does have a sleep. You can send the agent to sleep, baby. We just need the agents to send us to sleep. Well, Brian, this has been awesome. Just a deep dive, I think in a very forward-looking view into how design teams, as you say, especially ones that are going to be building AI products, are going to start doing their work. So where can we find you and how can we be helpful to you in Notion? You can find me, I'm mostly on Twitter or X, Brian underscore Levin, or my website, BrianLevin.com. And then I work on Notion AI. And I think it's genuinely one of the few useful sort of knowledge work agents. So if you haven't tried it, try it and send me feedback. We're always trying to make it better, help it do more things better, faster. So try Notion AI. Yeah. And we're big fans of Notion AI too here at the podcast. So definitely give it a look and definitely get some feedback and we will send it directly to Claude and put it in prototype playground. Brian, thank you for joining How I AI. Thank you for having me. Thanks so much for watching. If you enjoyed the show, please like, and subscribe here on YouTube, or even better, leave us a comment with your thoughts. You can also find this podcast on Apple podcasts, Spotify, or your favorite podcast app. Please consider leaving us a rating and review, which will help others find the show. You can see all our episodes and learn more about the show at howiaipod.com. See you next time.