Back/How I AI
Blog
How I AI

How I AI: My 24 Hours with Clawdbot (aka Moltbot)—3 Workflows for a Powerful (and Terrifying) AI Agent

I went from zero to one with Clawdbot, the viral autonomous AI agent. This is the real story of setting it up securely, testing it as a personal assistant for my calendar, and using it to build a web app and conduct market research.

Claire Vo's profile picture

Claire Vo

January 28, 202611 min read
How I AI: My 24 Hours with Clawdbot (aka Moltbot)—3 Workflows for a Powerful (and Terrifying) AI Agent

The whole time I was using Clawdbot, I had two thoughts fighting in my head. The first was: This is so scary. This is a terrible idea. Nobody should be doing this. The second was: Boy, oh boy, I want this thing. This tension sits at the heart of the current wave of autonomous AI agents. They promise to be the always-on, do-anything assistants we've dreamed of, but they come with a laundry list of security concerns that can make even a seasoned product leader nervous.

In this episode, I wanted to cut through the hype and give you my unfiltered experience. I invited a Clawdbot instance—which I named Polly—onto the podcast, gave it access to my machine, and spent 24 hours testing it as a real personal assistant. My goal was to see what it can actually do, what goes wrong, and what the experience feels like for someone who is both excited by and deeply paranoid about this technology.

Clawdbot is pitched as an open-source AI agent that actually does things. It can run on a local machine or in the cloud, spin up sub-agents, and take real actions like sending emails, managing your calendar, and even writing code. People are loving it for personal productivity and hating it for the high likelihood you'll do something very dumb with it. I wanted to see for myself. I’ll walk you through how I installed it, my workflows for testing it as a personal assistant and an async teammate, and my final verdict on whether it lives up to the hype.

Workflow 1: The Zero-to-One Clawdbot Setup (and the Security Headaches)

Before you can get an AI agent to do your bidding, you have to set it up. The Clawdbot site promises a quick, one-line terminal install, but my experience was much more involved. It took about two hours of wrestling with dependencies and security decisions before Polly was up and running. This is definitely a tool for tinkerers and developers right now, not the average consumer.

A podcaster shares her space, featuring a microphone, art, and a whimsical 'Please Join my Cult' poster in the background.

Preparing the Battlefield: A Dedicated Machine and User

My first move was to find a machine I could sacrifice. I grabbed an old MacBook Air that wasn't in use and created a brand-new, dedicated user account just for the bot. While Clawdbot has access to the file system, my hope was that containing it within its own user account would offer a thin layer of protection. If I were to use this long-term, I would wipe the machine completely and make it a dedicated bot machine.

Dependency Hell: More Than a One-Liner

The quick start guide was optimistic. Even on a relatively fresh laptop, I hit a wall of missing dependencies. I had to:

  1. Install Homebrew.
  2. Install and update Node.js and NPM.
  3. Install Xcode command-line tools.

After all that, the npm install command finally worked. This setup process alone makes it clear this isn't a plug-and-play tool for everyone. It requires a comfort level with the command line and troubleshooting installation issues.

Connecting the Brain: Telegram and Model Selection

Once installed, the onboarding flow walks you through connecting a messaging app. I started with WhatsApp, but the docs recommended a burner phone, so I quickly pivoted to Telegram, which I barely use. The process involves messaging a bot called "BotFather" to create a new bot and get an API token. It feels a bit shady, but it’s the standard Telegram flow.

You then give Clawdbot that token and create a personalized share token to ensure only your instance of Telegram can communicate with it. This is critical—if someone else could message your bot, they could access your files and send emails on your behalf.

During this setup, I also had to choose an LLM. I went with Anthropic's Sonnet 4.5 for a few reasons:

  • Fear: Honestly, I was nervous about what a more powerful model like Opus might do autonomously.
  • Practicality: The tasks I planned—scheduling and email—didn't seem to require top-tier reasoning.
  • Cost: I used my own API key to monitor spending, and I wanted to see how a more cost-conscious choice would perform.

Building a Digital Cage: Scoped Permissions and Separate Accounts

This was the most important part of the setup. I refused to give the bot direct access to my personal accounts. Instead, I created a completely separate digital identity for it:

  • A new email address: I set up a new Google Workspace account (polly.the.bot@...) for it to use.
  • A limited password vault: I use 1Password, so I created a new vault named "Claude" that only the bot could access. I stored its own passwords and an Anthropic API key there—never my own credentials.

This approach limited the potential damage but also, as I'd find out, some of the bot's functionality.

Workflow 2: Hiring an AI Assistant for Calendar Mayhem

With Polly set up, my first goal was to test it as an Executive Assistant (EA). I've onboarded EAs before, and the process is gradual. You start with read-only access and simple tasks. I tried to do the same here.

Key Lesson: Question Every Permission Scope

I asked Polly how to grant it calendar access. It walked me through the Google Cloud Console to enable APIs and generate a client secrets JSON file. Then, it gave me a URL for OAuth authorization. This is where I hit my first major red flag.

A detailed look at the broad Google OAuth permissions requested by an application, including access to Google Drive, Contacts, Docs, Sheets, Calendar, and Gmail, alongside 'Clawdbot onboarding' security instructions. This highlights critical considerations for app integration and data access.

The permission screen asked for the ability to see, edit, create, and delete everything: my files, contacts, spreadsheets, calendar, and email. All I wanted was for it to look at my calendar. I pushed back directly in the chat:

Do you really need all these scopes?

It immediately responded with the classic AI apology: "You are absolutely right. I do not need these scopes." It then generated a new URL with only calendar-read permissions. This is a huge lesson: always be critical of the permissions an app or agent requests and push back when they seem excessive.

The Simple Task: Scheduling One Event

My first task for Polly was to add an upcoming visit to the Vercel studio to my calendar. It couldn't find the event online, so I forwarded the confirmation email to its dedicated inbox. This is exactly how I'd work with a human assistant.

It did a great job ingesting the email details and even recommended adding travel buffer time. However, when it came time to add the event, it asked for write access to my personal calendar. I said no. My workaround was to treat it like a colleague:

Hey, can you just create an event on your calendar and invite me to it?

It worked perfectly. It created the event on its own calendar and sent me an invite. It even correctly handled deleting a duplicate event later. For a single, simple event, it performed well, albeit with some negotiation.

The Stress Test: Managing a Family Calendar

Feeling confident, I gave it write access to our shared family calendar. This was a mistake. I asked it to do a few things: fill in my kids' new basketball schedule, add recurring piano practices, and identify conflicts.

At first, on Telegram, it seemed like it was doing a great job. It confirmed every action. Then I opened my calendar. It was total chaos.

Demonstrating an AI assistant's ability to identify and correct scheduling errors in a chat interface, showing the interaction on a mobile device.

Everything was on the wrong day. It was consistently off by exactly one day. To make matters worse, the command-line tool it was using couldn't create recurring events, so my calendar was now flooded with dozens of individual, incorrect appointments that I had to delete one-by-one. While I was deleting them, the bot was trying to add them back, thinking something was broken. It was truly stressful.

I ended up sending it frustrated voice notes from Target while pushing a shopping cart. The exchange about time zones was hilarious and revealing. It told me:

The issue is I've been trying to quote unquote, 'mentally calculate' which day of the week each date falls on, even though the API is telling me what the date of week is.

I replied, exasperated, with my baby crying in the background: "You are a computer, you are not doing anything, quote unquote mentally. You are making calculations." This highlights a core weakness: LLMs have no real sense of time or space. They can't reliably handle dates and time zones, which remains one of the hardest problems in software.

Workflow 3: The Asynchronous Teammate for Coding and Research

Beyond administrative tasks, I wanted to test Polly's ability to do more complex, asynchronous work, like coding and research.

The Coding Project: A Next.js Conversation Viewer

I decided to have it code something. Via a voice note on Telegram, I gave it a detailed prompt:

Okay, let's use voice from here on out. I want you to document our conversation in a Next.js web app. That shows the back and forth of our full conversation from the very beginning today till the end in a ui, I want you to redact anything that is a secret key, a person's name, or a specific place, and I want to toggle between two UI versions of this display. I want you to be able to show me a terminal style conversation back and forth... or I want you to show me a telegram style text back and forth. The content should be in JSON, the same. Again, redact names, emails, dates, et cetera. Replace them with placeholders or redacted blocks, and then generate the next JS app. We are eventually going to deploy this to Versel. Can you let me know when it's deployed to Versel so I can look at it?

It successfully built the Next.js app locally. However, the experience wasn't great. The latency between sending a message and getting a response is just too slow for the tight feedback loops required for coding. Tools like Devin or Cursor's background agents feel much better suited for this. Deployment was also a non-starter, as the bot didn't have GitHub or Vercel accounts, and setting them up felt like too much hassle.

I ended up just Airdropping the repo to my main laptop and finishing it myself. The one magical moment was when I was on the go and asked for a status update. I said, "Hey, like shoot me a screenshot of what it looks like," and it did, right in Telegram. Being able to get files and screenshots from a remote machine is an underappreciated superpower.

The Star Performer: Market Research on Reddit

This was my favorite use case and where Clawdbot truly shined. I sent it a voice note: "Go on Reddit. Find what people want from ChatPRD, find what they want from a product AI platform, and email me a report."

This workflow was perfect for an agent like this for a few reasons:

  1. Latency as a Feature: For a research task, I don't expect an instant response. The fact that it took time to come back felt natural, like delegating to a human employee.
  2. Multimodal Communication: The experience of giving a command via voice and receiving a detailed report in my email inbox felt seamless and powerful.
  3. High-Quality Output: The report it emailed me was excellent. It was a well-structured Markdown document with key insights, bullet points, and links to the relevant Reddit threads. It was actionable, accurate, and presented exactly how I'd want a product manager on my team to deliver research.
A podcast host discusses AI-powered product research, with a detailed Markdown report visible on screen summarizing Reddit insights on PRD tools, integration limitations, and the challenges of long PRDs.

This task showed the true potential. An agent that can take a high-level goal, use tools to research it, and deliver a synthesized, professional report is something I absolutely want.

Conclusion: I'm Scared, and I Want One

After 24 hours, my feelings about Clawdbot are as conflicted as when I started. It's powerful, the interface of chatting with your computer from anywhere is compelling, and it can do some things remarkably well. The Reddit research workflow felt like a glimpse into the future of productivity.

But the product isn't there yet. It's too technical for non-developers, the latency can be frustrating, and the security implications are genuinely terrifying. I found myself constantly fighting its bias to act as me rather than for me, impersonating me in emails and grabbing the widest possible permissions. I've since uninstalled it, deleted the keys, and removed the bot. It just feels too risky.

This raises the million-dollar question: who is going to build this for real? The big players like Google, Microsoft, and Apple have the data, the models, and the device integration, but do they have the risk tolerance to build something this autonomous? Startups will try, but they'll face huge hurdles getting the deep data access needed to be truly useful. Clawdbot has shown us what’s possible and ignited the imagination of every builder in the space. It has product-market fit as a category, even if this specific open-source tool isn't the final form. We're going to see a lot more agents like this, and I for one will be watching—and cautiously testing—to see who finally gets it right.

Brought to you by

  • Lovable—Build apps by simply chatting with AI

Try These Workflows

Step-by-step guides extracted from this episode.

Become a 10x PM.
For just $15 / month.

We've made ChatPRD affordable so everyone from engineers to founders to Chief Product Officers can benefit from an AI PM.

How I AI: My 24 Hours with Clawdbot (aka Moltbot)—3 Workflows for a Powerful (and Terrifying) AI Agent | ChatPRD Blog