The exact AI playbook (MCPs, GPTs, Granola) that saved ElevenLabs $100k+ & helps them ship daily

howiaipodcast 5Byg-9K8JnM Watch on YouTube Published June 01, 2025
Scored
Duration
44:24
Views
25,836
Likes
533

Scores

Composite
0.54
Freshness
0.00
Quality
0.85
Relevance
1.00
7,889 words Language: en Auto-generated

when you're editing as much as possible try and edit the underlying chunk rather than the actual output. I like that you have here the UR and then gives a very specific identity and job to be done at the top of this and then you have very specific instructions where you say you must do ABC D and it's quite particular. This saved us $40,000 a year for the tool so immediately canceled it over $100,000 in agency costs. I think the highlight though is just not having to deal with more SAS vendors, more agencies, constantly trying to get arts sold. Welcome back to How I AI. I'm Claire Vo, product leader and AI obsessive here on a mission to help you build better with these new tools. Today we have a great conversation with Luke Har's, head of growth at 11 Labs. Luke shows us how to make everything a launch by making everything automated with AI. He shows us his secret flows for generating case studies and tweets on the fly, how he saved his company tens of thousands of dollars by, yes, being a marketer that coded in cursor, and explains what an MCP is and how he hooked it up to WhatsApp. Let's get to it. This episode is brought to you by Orcus, the company behind open-source conductor, the platform powering complex workflows and process orchestration for modern enterprise apps and agentic workflows. Legacy business process automation tools are breaking down. Siloed, low code platforms, outdated process management systems, and disconnected API management tools weren't built for today's event-driven AI powered cloudnative world. Orcus changes that. With Orcus Conductor, you get a modern orchestration layer that scales with high reliability, supports both visual and code first development, and brings human, AI, and systems together in real time. It's not just about tasks. It's about orchestrating everything. APIs, microservices, data pipelines, human in the loop actions, and even autonomous agents. So build, test, and debug complex workflows with ease. Add human approvals, automate back-end processes, and orchestrate agentic workflows at enterprise scale, all while maintaining enterprisegrade security, compliance, and observability. Whether you're modernizing legacy systems or scaling next-gen AIdriven apps, Orcus helps you go from idea to production fast. Orcus orchestrate the future of work. Learn more and start building at orcus.io. That's o k.io. Hey Luke, thanks for joining. Thanks for having me. In 2025, we've talked a lot about vibe coding, cursor this and vzero that, but we have not talked enough, I think, about vibe marketing. So, what do you think the future of an AI CMO is? In the next couple years, there's all these tools like Lovable and Cursor, and the rate of software production is going to go exponential, but it's not going to matter if no one's actually using your tool. And so what's important is actually getting the product into market, getting people to know about your new features. At 11 Labs, we have this launch process. So basically every new feature we do, every new model, we run it through this massive checklist, which is like, okay, we need to first work out what are the value props. Then we need to work out what's the core messaging. Then we need to work out who's it for. Then we need to turn that into the blog post, the X post. And it it takes a lot of time, these massive launch processes. And so the thing I'm really excited for the AI CMO is being able to go from every single new feature or new product and translating that into your entire launch process, making the assets, making the videos, making images, but then also going beyond a launch. So what are those then evergreen channels that you'll be testing? And so uh let's say 11 Labs, we launched the best speech to text model. Okay, we need to be running Google ads for that. So then it will spin up understand all the various keywords, spin up the Google ads, it will optimize the landing pages. So I think this entire thing is going to change massively. Um, and we're already using a few of these different workflows and I'm excited to talk you through them. Yeah. And I think one of the best ways that companies can market is actually just telling great customer stories. And I know you have a workflow for getting case studies out. You're lucky enough to have probably tons of customers that love you. So, can you walk us through how you use AI to make case study writing really easy? With case studies, you know, you sign a great customer and you now want to be able to tell the story about how they actually use your product. And so, what we're going to do live, Claire, is I'm going to have you write a case study for 11 Labs. I know you've used 11 Labs. And what we're going to be using is the tools Granola, which is a fantastic transcription tool. It's a note-taking tool. Uh, and we're going to use chat GPT actually with a custom GPT which will then uh translate that into an excellent blog post and as a bonus we'll write a tweet in a in our company voice in as well. Okay, I'm excited and I am a happy 11 Labs customer and you didn't pay me to say that. So I I did not. Um, so what I'm going to do is I'm going to open up Granola and I'm going to do a case study interview with CLA. I'm going to just ask you two quick questions and then we'll use the transcript for this. So cla fantastic to meet you and I would love to understand how you use 11 labs in your work. Yes. So at uh my company we have to produce a lot of customerf facing live events. It's the way that we connect with developers and customers in our community. And those live events can be either in-person events or they can be recorded streamed events. And we put a lot of preparation into the messaging and the way we present our products at those events. So, ahead of our user conference that's coming up in a couple weeks, we actually build a script for what our product keynote will look like. And it's me, it's our CEO, it's our SVP of product, it's engineers demoing, it's a Q&A with customers, and we like to run through and rehearse those keynotes. And the rehearsals are very expensive from a time perspective. I just named 10 people that have to be in the room. We have to walk through the script. We have to record it and then listen to it later. And we're really trying to nail a very specific set of timing. You know, we only have 30 minutes or so to get all this content in. And if you're participating in the dry run, it's actually really hard to listen for is this good as a third-party observer of this keynote. So, one of the things that I do with 11 Labs that I find super useful is actually build prototypes of keynotes. So, I load in the script into the I think it's called the studio flow and I give everybody in the company and our customers different accents. My boss has this lovely British accent. So, I give him a British accent. I give myself a different act. So I pick voices for everybody in the keynote and then I actually generate a proto like an audio prototype of the keynote and send it around for people to listen to for two things. One, timing to make sure do we have enough content? Do we have too little content from a timing perspective? And then two, does it narratively flow and sound natural and is easy to understand and listen to because it's a virtual event. And I found that that little flow, which I guess for the how I AI listeners is also a little a mini how I AI built in built into the flow has been really useful for us to make sure we get high quality events going. Fantastic. And so what I do is, you know, basically chat through in more detail and I try and get out some some concrete concrete facts the use cases. Cla's already said she's using the studio product. Um, and maybe you would even go as far as using voice cloning of your different customers. So, we may include that in the case study. And maybe you could give me a metric. So, how has this driven ROI for your company? How has this doubled the revenue of launch? Yeah. So, I I think it saves us significant amount of time internally. So, there's definitely hours and hours saved in terms of iterating on something like a keynote. And then if we make it high quality, then our customers hear more about our great products and then of course we get to sell more. Yep. Fantastic. So, and maybe even you you say a statistic like it saves you uh 5 hours of meeting prep time. Yes. It saves me 10 hours of meeting prep time. Let's make this one a good case study for you. Great. Uh so then I click stop transcript uh and generate the notes. So all this time, Granola is sort of recording what we're saying and um analyzing it on on the back end. So now what we're getting is this auto summary based on all the stuff that I just said. Yeah. And um Granola is super smart. I was actually speaking with the one of the founders. It's pretty cool. It actually would take CLA's email, use that to enrich it to work out your job title. And so it pulls in all this extra context to make these fantastic uh summaries and transcripts whether it's for case studies or or meeting notes. And what we then do is I've created a custom GPT which we use throughout the company and it's the 11 Labs copy editor. And I'll just edit the GPT. So you can uh see here what the prompt is. And for those listening, uh, a GPT is a chat GPT sort of like customized chatbot that has content and instructions in it. Yeah, you can think of it as a as a very easy way to share a prompt. Um, and so this is the real prompt we use. And what I've done is I've fed in our toner voice guide. And so I've said how you're an expert editor, your writing assistant specializing the 11 Labs communication style. Uh, you must enforce American English spelling, even though it breaks my heart. You're a serious research-led tone of voice similar to Palanteer SpaceX. So, it goes into quite a lot of detail. Talks about preferences for types of words. And then it includes some um example blog posts that we've done. It includes some example tweets that we've done. So, different case studies and tweets that we're very happy with. What I'm then able to do is use this GPT. I paste in the granola summary. So I'll say create a case study of how CLA uses 11 Labs for launch darkly. And then I go here's the summary of the call. And I find Granola normally gives the best summary as well as pulls in this extra context. And then for bonus points, I then actually copy the raw transcript from Granola as well in case it wants to grab any points. So here's the raw transcript. Can I tell you what an amateur granola user I am, which is not know I did not know how to get to that transcript. So little tip for Claire here. Yeah. And then so I'll also just say for the studio product and then I hit send. And I find pretty much the first time it gets something that's usable. Here we go. It's now writing that out. How launch darkly uses 11 Lab Studio to prototype product keynotes. And one of the like key prompts that I'd given to the GPT is to make the headers like skimmable summaries of the article. So we've got cutting prep time in half for live events, prototyping keynotes, and because I've done the raw transcript as well, it pulls that one out, too. Um, I'll often do like one or two iterations actually just in the asking. So maybe it's including certain hyperlinks to other SEO articles, maybe it's got certain product details wrong if you want to include pricing and then we put that live on our blog. The other thing then is if you think about I love to treat everything I can as launches. So if you think about your case study as a launch like first of all you have to write it but then the distribution really matters. And so what I've also done in this GPT is tell it what a great tweet looks like. Even the aesthetics of a great tweet like okay come up with a hook line and then do a few like either bullet points or a short paragraph then do an image. And so if I go write it as a write a tweet thread for this, it will then rewrite that into a tweet summary. And it writes these handy brackets around like what should the assets actually be. Uh this is the bit where I don't think we have a full AI AI CMO just yet which is I'm really excited about the IM image generation models the new GPT models because then I think you will actually be able to do like end to end launches and case studies by pulling this in. Yeah. So this this tweet generation chat right now not only writes the content but actually puts these placeholders of what media you would need to make an effective tweet. So, a screenshot or or something like that. And so, here you've gone from I don't know, we spent three minutes where I I blabbed a little bit about a a use case to a very polished case study, a tweet, and then I'm presuming you're going to do three or four other other things off this one one asset pretty quickly. Yeah. So, then you would do the LinkedIn post. uh you maybe would write it in the style of you could have a uh GPT in the style of your founders's tone of voice. So then you basically paste it into that and he's got the asset too. Um and then the way I'd always zoom out and think about this is how do you actually put this into a workflow? So I think the best growth systems you can do these one-off efforts but things get busy you know your time gets taken up and really how do you build it into a system and so concretely set up a Zapia such that each time you get to close one deal in your salesforce it sends them an email with your calendarly link and you just get booked fantastic different customers maybe it's a month into their into their contract where all you have to do is rock up have a nice chat with them you can even Get GPT to summarize or pre-prepare what the different bullet points and topics you should cover. Chuck it through your granola and then chat GPT flow and you'll you'll be turning out five case studies a month in no time. This is a great flow because I often find things like case studies or little marketing assets are easy to make but you have to remember to do them and if they take time you you know you get put in a meeting or you have to pass it to somebody else and you just sort of forget and you slow down the next steps and then you produce less assets. So I think it not only makes it easier to produce the assets but it makes sure that that engine keeps going cuz you as a human are not responsible for that that next step. And a common theme uh I think we touch on in in one of the next examples as well is like when you're editing as much as possible try and edit the underlying prompt rather than the actual output. And so if you're like, ah, it always does headings which don't, you know, maybe they're not particularly strong or I like more numbers or I like more concrete stats, make sure to incorporate that back into the underlying prompt. And on that point, can we go back to the the GPT just for a quick minute because I'd like to call out some things that I think you do pretty well here in in the prompting that I think folks can learn from. Okay. So I from from a prompt perspective, you know, very commonly everybody starts with the UR. And so I like that you have here the UR and then gives a very specific identity and job to be done at the top of this. Basically making sure that copy that comes out of 11 Labs matches matches the strategy or matches the the tone of voice and the brand. And then you have very specific instructions where you say you must do ABC D. And it's quite quite particular which I think is nice. Um some folks I know love very general prompting but I find that if you have a point of view of what your tone of voice should be, this sort of like very precise formatting prompting is is very important. And then you've broken down those instructions by types of content generated. So you have instructions for tweets, instructions for blog posts, and then the last thing at the end, um, which I also think people underutilize as good examples. And I have a question. Do you use any bad examples in here? Is it is it all good examples? I mean, they were comically bad when when I was coming up. We do actually use for the next workflow I'll show you. We do actually use bad examples for that for translation, but you know, a bad blog post is clearly a bad a bad blog post. Um and actually if I was one extra thing I found sometimes I think you know to draw back the learning from the granola team like give it as much context as possible. So if I was to extend this further, I think it would I would give it a lot of information around like what's the core messaging for each different product that we want to get across and really nail and then it knows when I'm doing the different interviews. Ah the studio product we really want to emphasize how you can do like multi- dialogue complex speech and then it would draw that out too. Yeah. Now, the other thing that I think people worry about is that like AI on top of AI on top of AI becomes very lossy. And I like the idea that you use the granola summary, but then you also use the raw transcript. So then you have both sort of the the highlevel summary as well as some raw context. And because these contact windows are so big, the chat can make sense of it. And without doing the raw transcript, you wouldn't get any of the lovely quotes as well from what the customer exactly. I didn't even think of that. Okay. Well, this is I'm going to steal this workflow. This is so so great and so fast. And uh I love your philosophy of everything is a launch. So, that's that's a really good way to to think about things. This episode is brought to you by Retool. There's a huge gap between impressive AI demos and AI apps that deliver real value inside your business. While most AI solutions can only generate text, Retool lets you build apps that take meaningful action by connecting directly to your business systems and data. With Retool, developers combine the power of code with the speed of visual building to create AI powered tools that solve real problems. No more writing endless integration code or building UIs from scratch. The results speak volumes. The University of Texas medical branch increased diagnostic capacity 10fold. Amazon's Genai team uses Retool to make complex AI accessible to enterprise customers. And RAMP saved $8 million while boosting efficiency by 20%. That's why over 10,000 companies from startups to Fortune 500s trust retool as their AI app layer. Retool because AI should do more than talk. it should work. So I know you had another use case where you were using a external tool or some sort of tool and you actually just built a solution that saved the company quite a bit of money. Yeah. So, this one was um for 11 Labs, we're in a whole bunch of different countries and it's very important to us that we localize all our content and so we want our homepage to be in Hindi, in Spanish, in German, in Polish, in Japanese. Um, and I set out about this process of how do you go about localizing the website? And I spoke to loads of the top experts and apparently what you're meant to do, you set up a very expensive localization tool. Uh so the one we chose, I won't name the name, but it was $40,000 a year. Uh and it quickly went up or they kept on trying to push it up. So it's you're now paying $40,000 a year for this tool. And then the tool you then need lots of humans inside to actually do all the translation work. So then you're getting agencies which you're paying about $100,000 for. Um, and we set up this flow. There was actually quite a lot of engineering work to connect it to our CMS, to connect it to our codebase. Um, and I was like, okay, fantastic. I've done all that. Uh, but the AI translation is terrible. And then we found out that agencies and the humans were terrible because you're constantly playing this cost game um of trying to minimize the cost so you can't get anyone who's any good. And meanwhile, we have AI which is like utterly taking off. And I had this situation where my key team kept on sending me screenshots back from chat GPT being like, oh no, this one's better instead. And I'm like, well, if we're just using chat GPT for the reference of what's better, why don't we just use chat GPT for the whole thing? Um, and so I've got this uh Figma board where I've kind of laid out in a bit more detail what we started with and what we went to. But basically we ripped out this entire tool all of the agencies, wrote a very small server where all it does is take the string has a prompt per language explaining what the tone of voice is for that language and the context sends it back whether that's into GitHub or payload. And this saved us $40,000 a year for the tool. So immediately canceled it. Over $100,000 in agency costs. And previously we were waiting days to get the translations back. Whereas this is now instant. And if anything is very sensitive like say our pricing page, we just have one of our team. So we're already a decently large team. One of our team just to do a quick sense check. And if anything's wrong again, we update the prompt rather than the source code as much as we can to make it make it better. So you just replace this tool. And what I think is so interesting is I have this debate internally as somebody who provides SAS software. You provide SAS software and I uh I think one of the existential threats in the SAS industry is the cost of building going to zero. And I talked to so many people and they say teams will never build this in ter like why would you? But then what I think you're showing is it it can actually be quite cost effective and improve quality to think about building these tools yourselves. And did you I mean who built this? Was it the engineering team? Like who actually this I did the first 90% in one day and then I got one of the engineers to help. So I literally built it all in cursor. I was actually ill. So, I was meant to be skiing. I was lying in bed and I was having to deal with the fact that we had just gone through three different agencies who didn't meet our quality bar for translation and I was like, I cannot be bothered to get a fourth agency and I'm just going to rewrite it all. Uh, so I did the bulk and then um yeah, one of the fantastic engineers on our team, he helped uh helped get into production. I think the highlight though is just not having to deal with more SAS vendors, more agencies, constantly trying to get upsold. And my broader the broader question of like is all SAS dead? I think no. But I think human in the loop SAS like if your job is about putting low-skilled workers in some sort of flow, which translation is, I think that's very risky because just the AI and like at the moment we still do need a little bit of like every week or two we have someone just give a quick scan. and is it all great? And they do make little tweaks, but you know, give it 2 years time. I would much rather bet on AI costs getting cheaper and the quality going up rather than paying for more agencies. So maybe I have three three takeaways here. One is you really should reconsider looking at build versus buy on some things, especially if you're not satisfied with the quality of the buy. It's worth worth the investment. So I think that's that's thing one. thing too is look out. Your marketers are going to hop into cursor and get it 90% done and then hand it to you engineers. So, you might as well do it yourself. They'll be making more work. Yeah, they'll be making more. And three, um, do you know how many software products I have built out of the frustration that I'm supposed to be on vacation, but I am actually sick. That is like the perfect the perfect time to get something new done. Um, so winter season is is a highly productive season for me because I'm always sick. You you need to get them in the multi-year deals, not ending in a holiday. I have to ask, do you feel like this is should we should we SAS eats its own tail? Would you ever productionize this and sell this to others? Is it very No, I think we're going to open source it. So, there's a couple of there's a couple of things that I'll just show you. Um, if enough people in your YouTube comments, okay, uh, say to open source it, we're open source it. But just to add a, you know, this basically summarizes what it was is you have all your code in GitHub, you have your like strings and then you push it into the SAS tool. And the biggest issue truthfully was they didn't allow you to edit the prompt of the AI. And so there was just no way the translation would be any good. understand your brand language, understand the glossery. And they quoted us about 6 months before they were going to ship the ability to edit prompts. And I think it's because the whole business model is based on no get these humans in it. Um, and so instead what we shipped was this whereby it's just a GitHub action which understands runs every time you changes change the keys in your translation dictionary sends it to an LLM with a prompt per lm and saves it back and that works way better. Um, and then the same with our CMS we just built into it a uh translate button which again just sends it in. And it was so nice just having one source of truth. All the text is either in your well two I guess either in your CMS for all the blogs or it's in the code base for your core pages. Also another shout out this is this is not a paid advertisement cursor but we wrote this cursor rule which does all the lift. One of the things with the engineers was like it's quite annoying to have to extract all your strings. And so we wrote this one cursor rule where you just say extra like translate the strings and it grabs them all into this en.json file nicely wraps it handle server side or or client side rendering. So that was pretty fun. So I'm making the first request. Everybody in the comments do it for me as well. I would like this open source because I would 100% use this use this flow. And you know, I think people get scared when you hear, you know, human the loop is out. But I do think there's this opportunity for folks to operate at a at a higher level of their craft. And so, you know, this is not the fun part of translating is taking string A into string B. Then you can start doing things like, does this match localized style? Is this appropriate? Is this how we want to talk to our customers in this region? And so I do think there's this ability for humans to then add a layer of quality um and use of their intellect and skills that is higher level than than this. Yeah. And the really cool so we spoke through the cursor rule the get app GitHub action is here which is basically instantly on each push it generates that but exactly as you said we just have this prompt file per um per language which talks through again our brand guidelines translation style keywords and the cool thing is you can define that and like you can really take the care and the nuance exactly how you want to represent your brand and then you'll be pretty confident that that's then uh scaled up across any content you're then putting out in future. And before truthfully, we weren't planning on translating, you know, every blog page, but now like you actually can do and it's a much better experience for your users. Have you tested putting these language specific prompts in the in the language itself? So, we explicitly decided not to because I wouldn't have then been otherwise able to vet. Yeah. That we were consistent. Um, but I did I'm not sure if you saw that tweet I did, but I did say that recently of basically someone asked a doc system a question in Arabic and it replied, well, because the docs are in English, I'll reply in English. It's like no, you you do definitely want it to reply in the language uh the language you want. I love this and I love that you built built most of this. And is the maintenance cost very high? There's none. There's none. The um well, it hasn't broken so far. Um and it's just because it's just a GitHub action which is updating the JSON strings. Yeah. So, um Great. Yeah. Well, this is super useful and a good selfc customer story. You saved yourself $40,000. So, yes, you know, give yourself a case study. Okay. And then we saved I'm not saying the best one for last, but I love this last example. Uh, one, because you get to explain to the audience what an MCP is for those that are still confused by the concept, and two, I think you built a really fun one. So, you want to talk about your WhatsApp MCP? what an MCP is, it's a model context protocol. And so it's a protocol uh written which enables anyone to expose tools to AI agents. And so the example and why I built the WhatsApp one was we all get tons of messages. We're all in tons of different WhatsApp groups and it's really hard to stay on top. But currently, if you ask a tool like Claude, it has no idea about any of your WhatsApp messages. It can't help you out. And so, what again, actually, that same weekend that I was ill, I did one I did the Saturday was uh ripped out the uh ripped out our translation software. And then the Sunday was okay, can I actually connect WhatsApp to my AI system using an NTP? And it's part part of my broader thesis of basically like I think a personal AI assistant really only needs your WhatsApp, your calendar and your email. And then it knows everything about you and it can even organize tasks, send emails, you know, organize dinners, dinners with your with your friends. So uh that's why I built it and there's also some like cool um use cases for work as well uh which we can jump into. So let me show you. Cool. So this is the WhatsApp MCP repo and how it works is it has two main parts. So it has um a bridge which it pretends you know WhatsApp web when you sign in you scan that barcode that's exactly how it works. So it actually pretends to be WhatsApp web and it uses fantastic library called what's meow to do this. Uh so you when you run it in your terminal, you scan your barcode. Um and the first thing it does is it downloads all your messages onto your local computer, saves it in a SQLite database. Um and what that means is you can then keep querying it as much as you want with an AI. Uh and you have no risk of or very low risk of being banned because it's only downloaded it once. So yeah, to be fair, this is unofficial uh unofficial stuff. Um and then the other bits the WhatsApp MCP server which basically gives the ability for your AI to query this uh SQL like database as well as like sending messages, sending voice notes. So if we jump over to Claude Mhm. And this is the desktop Claude instance. Yeah. Desktop Claude. They've also now shipped the ability for you to run it from claude.ai uh if you host it too. Um, and what I've got is I've got this MTP called WhatsApp here. And I can just type into the chat, uh, so what, uh, are some recent messages WhatsApp received? Um, and what that will then do is it will use the tools that the WhatsApp MTP exposes and then summarize and use back that information. And here you can see it's talking about 11 Labs new features. And so uh 11 Labs is launching conversational AI agent or a new speech to text offering. Um as well as a few tests like how are you and hello world. Um, and a few examples of like why you may actually want to use this is I'm in a whole bunch of different WhatsApp groups. And what you can do is you can use it to summarize, oh, what over the last week were people actually talking about? And often now some of the best uh way to keep up with trends or the best thoughts on new tools, they're all in these WhatsApp groups with hundreds of messages per day. And so, uh, if you're looking for a way to get an edge on Twitter or LinkedIn, you can say, uh, summarize the thoughts on 11 Labs from the messages. Um, that will then search the messages that you've received on WhatsApp, which is talking about it. Uh, and then you could take something like this, chuck it into your GPT, which is already trained on your tone of voice, uh, and generate a tweet thread for you. Everything's a launch. Everything's launch. Uh, and also to plug it, if you're interested in how we run launches, the add a blog post which goes through like all the different sets that we use as well. Yeah, we'll we'll link to that in the in the show notes. So, okay, just recapping this for folks that are still have have their mind blown. So, you built this MCP which you've open sourced, which again is unofficial but friendly. Um, implemented a nice way. uh you download this code, run it locally in your CLI, it does a one-time pull of your data. So if you want it updated, do you just run and refresh that again or does it pull? How does it When you start it, it will pull all the recent messages between the last time it ran and now and then whilst it's running, it will also receive any new messages. Got it. So actually I can al and I can use it to so send a message to and then I'll do a phone number and then potentially if my phone's not on silent. Yeah. So that's just came through saying hello. Um and then you can also use it connect. The cool thing about MCPs as well is you can connect to lots of MCPS at once. Y and so I have an 11 maps MCP installed so I can use it to you could like generate a voice round up. So you generate text to speech of this. Yeah. So you can stitch these all together. Okay. So see you have this MCP. It pulls down your stuff. It gets regular updates. It's all local. So none of this stuff is going to the cloud. Um, and then you've connected to that ser or that server through your your local desktop claude or or if you you hosted it, you could do it on the the web version. And then now just in this chat box, you have access to all these different tools. And you know, one of my hypotheses with AI is like tabs start to go away. If you're like me, you have 500 tabs open, everything along the bottom of your desktop dock, and you're switching context. And this sort of centralized chat interface that can access all these tools just makes you much more efficient and also allows you to get really creative about how you stitch these tools together. Yeah. And so and so for this one for example, we just generated this text and then you can say send this to the phone number. And behind the scenes, I mean, for folks that aren't seeing this, it really is basically using natural language to select from a list of tools which hits a list of publicly available APIs that have been appropriate appropriately authenticated. And so for anybody who, you know, kind of knows what an API is, but maybe isn't an engineer, but wants to be able to say to a system, do this thing for me and use sort of the exposed endpoints, this might be a more accessible a more accessible framework. Yeah. My the overall thesis with tools and agents is with we spoke about that granola flow earlier which was um when something happens. So when a deal is closed then send out you know use Zapia to send out a calendarly the person books in then you generate you do the call you take the transcript you put into your GBT and then and that's all very static and very rigid but let's say hypothetically I actually want to do a roundup of five leading startups which are all doing this well suddenly my workflow is completely broken. if you had perfectly scripted that all out in a tool like you know Zappia or NATM like that's actually now not usable and you'd spend a ton of time resetting it up and so the really cool thing about these chatbased MTP tools is it can be much more you know it's trying lots of different ways to like okay how do I actually send this audio message you know on the spot we were like okay generate this now send this and the AI as the models get smarter and smarter are able to get deal with these like higher level abstraction tasks. And so a genuine one you could do is uh why not have your AI actually phone up CLA and have the conversation for you about that. So uh if you want we could actually try that now. I'm not sure if it will work. Why not? Let's try. So we can do that. So create a conversational AI that can uh do case study interviews. And so this will then use 11 Labs to actually create an AI agent that you can speak to about case studies. So first of all, it's going to yeah list the agents. I love this because you're using AI to create more AI. You're really just replicating agents on agents on agents. And then what this would do is create a specialized agent on the spot for a specific use case um that then you could use to give me a little call. Yes. And get get a case study done. Great. Yes. Exactly. And so you can actually see that here. It went through and created a prompt. So it's got the first message. Hi, I'm your case study interview coach. that if you go to our Twitter, you can see Louie doing this workflow where he then um phones up and orders a pizza using an AI and you just say, "I would like to create a pizza ordering AI agent." But yeah, hopefully this gives the listeners a glimpse into like you can on the spot come up with these agents which I think will be more and more abstract which can then do these tasks for you using the tools as they go. So that's the promise of MCP and transparently it's still very early and I think most of these are tools uh are toys but I do think it's it's going that way. As a parent of kids who really like pepperoni pizza I'm very worried about the ability to spin up a pizza ordering agent in my house because we will end up with with a lot of pizza. Okay Luke, this has been incredible. Uh you made a case study with me using Granola AI and your magical GPT. We eliminated $40,000 of spend um by coding in cursor as a as a marketer. And then you built a way to use Claude or an AI to query your source of both personal information and industry news, which is WhatsApp, and do a bunch of really interesting stuff there. So, uh this has been very eye opening to me. I've learned so many things from what you shown me, um including that everything is a launch. I'm just going to keep that in my mind. Well, let's wrap with a couple lightning round questions. The first one is, you know, we've we've talked a lot about coding and textbased flows. Even in what you showed, a lot of it is coding and textbased flows, but I think what's so interesting about your point of view is you're starting to bring the idea of voice as input and output into um into the the industry and into how people build products. you know, very quickly, what do you think kind of like voice modalities unlock maybe for product managers to think about in terms of what they're building? I think there's two broad types of things they unlock. So, one of which is new experiences for customers which just wouldn't have otherwise been possible. And so, one great example is like if you're in education, suddenly you can make something which is way more engaging. So, chess.com shipped a app with uh Professor Wolf which enables you to get like turnbyturn guidance on your different chess. And so you can imagine a world where every whether you're learning a language, you're playing chess, you're you're uh learning maths that everyone can have this like interactive tutor and that's kind of yeah these new experiences. The other type which I think is really exciting if you're a product manager is you probably have a lot of back office functions and but so particularly if you're an internal PM if you look at all the places that you currently have um you know people doing manning phones. So often this would maybe be like say you're doing research collection you know you're a scaled marketplace and you have large numbers of people collecting data or you're doing customer support. Well, maybe actually currently you're not able to expand internationally because your team only speaks English. Well, as now you could spin up an entire team of customer support agents who are fluent in French and in Spanish and in German. So, I think that's really exciting, too. Yeah, I love the I love the international uh angle to this. It's something that I haven't heard very many people speak to. Okay, Luke, this has been so great. Where can we find you and what can we do for you? Thanks so much for having me on the podcast. It's been a lot of fun. Uh you can find me, my website is harry's.co. Um and a couple of blog posts people may enjoy. So I've got the uh how to launch your products where I literally talk through the checklist that we use uh through our launches to go from idea to value props to core assets to distribution. I also talk about how to hire your first growth marketer. Um and you can follow me on Twitter. That's Luke carries and then underscore. Great. Well, this has been so fun. Thank you so much. Cool. Thanks so much, Cla. Thanks so much for watching. If you enjoyed this show, please like and subscribe here on YouTube, or even better, leave us a comment with your thoughts. You can also find this podcast on Apple Podcasts, Spotify, or your favorite podcast app. Please consider leaving us a rating and review, which will help others find the show. You can see all our episodes and learn more about the show at howiipod.com. See you next time.

Summary

Luke Har's from Eleven Labs shares AI-powered workflows that automate marketing, save significant costs, and enable personal AI assistants using tools like Granola, custom GPTs, MCPs, and cursor. He demonstrates how to turn every product feature into a launch, generate case studies and tweets instantly, replace expensive SaaS tools with custom AI solutions, and build a personal AI assistant that accesses WhatsApp data via an open-source MCP.

Key Points

  • Luke Har's, Head of Growth at Eleven Labs, demonstrates AI workflows that automate marketing and save the company over $100,000 annually.
  • He uses Granola for transcription and note-taking, then feeds the summary into a custom GPT to generate polished case studies and tweets in the company's voice.
  • To save $40,000/year on a localization tool and $100,000 in agency fees, he built a custom AI translation system using a GitHub action and LLMs.
  • The translation system is open-source and uses language-specific prompts to maintain brand consistency, with a human-in-the-loop for quality checks.
  • He built an open-source WhatsApp MCP (Model Context Protocol) that connects to WhatsApp web, downloads messages locally, and exposes them to AI agents like Claude for querying and summarizing.
  • The WhatsApp MCP enables AI to access personal information, summarize group chats, and generate content like tweet threads based on real-time conversations.
  • The core thesis is that everything can be a launch, and AI should be used to automate not just content creation but also complex workflows and personal data access.
  • He emphasizes building custom solutions when off-the-shelf tools don't meet quality standards, especially when AI quality is rapidly improving.
  • The key to effective AI prompting is providing a clear identity, specific instructions, and examples (both good and bad) to guide the output.
  • The future of AI includes agentic workflows where AI can create other AI agents to perform complex tasks, like making a phone call for a case study.

Key Takeaways

  • Build your own AI solutions when commercial tools don't meet your quality standards; it can be cheaper and more effective.
  • Use a combination of high-level summaries and raw transcripts to give AI context for better content generation.
  • Create custom GPTs with specific prompts, tone of voice guidelines, and examples to ensure consistent output.
  • Implement automated workflows (e.g., with Zapier) to turn customer success into case studies, ensuring a steady stream of content.
  • Explore Model Context Protocols (MCPs) to connect AI agents to personal data sources like WhatsApp for powerful, real-time applications.

Primary Category

AI Business & Strategy

Secondary Categories

AI Engineering AI Tools & Frameworks Programming & Development

Topics

AI marketing case study automation custom GPTs Model Context Protocols AI translation WhatsApp integration AI agents prompt engineering AI workflow automation cost savings with AI

Entities

people
Luke Harries Claire Vo
organizations
ElevenLabs Orkes Retool Cursor Granola Zapier
products
technologies
domain_specific
products technologies

Sentiment

0.85 (Positive)

Content Type

interview

Difficulty

intermediate

Tone

educational instructional entertaining technical promotional