Build a Speed-to-Lead Bot in a Weekend
- 1
Set up a simple form using Typeform or Tally that captures name, phone number, and inquiry type.
- 2
Connect the form to a webhook using Make.com or n8n. When a form is submitted, the webhook fires.
- 3
Use an AI code tool (Claude, Cursor, or Replit) to write a script that triggers a call via Twilio or a voice AI provider like Thoughtly, ElevenLabs, or Bland AI. Pass the prospect's name and inquiry type into the opening script.
- 4
Prompt your AI engineer to write a call script with a natural opener, two qualifying questions, and a calendar booking link delivered via SMS at the end.
- 5
Run 10 test calls on warm leads. Measure hangup rate, conversation length, and booking rate against your current follow-up method.
Stretch goal: Feed three of your best real sales calls into the knowledge base and instruct the agent to mirror your objection handling style. Compare conversion rates between the generic and trained versions.
Build a Government Contract Opportunity Scanner
- 1
Pick one government procurement portal relevant to your industry — SAM.gov for federal contracts, your state's purchasing portal, or USASpending.gov for award history.
- 2
Use Claude or Cursor to write a Python script that pulls new contract opportunities matching three keyword filters you define (product category, agency type, minimum value).
- 3
Pipe the results into a Google Sheet that auto-refreshes daily. Add a column for AI-generated fit score: prompt Claude to rate each opportunity 1-5 based on your capability description.
- 4
Set a Make.com automation to Slack you any opportunity rated 4 or 5, with a one-paragraph summary of why it fits and the submission deadline.
- 5
Submit a response to one opportunity this week — even a no-bid letter. The process of responding teaches you the procurement language faster than any research.
Stretch goal: Use Prol's core insight: build a relationship map of agency contacts who purchase in your category. One warm introduction to a contracting officer is worth 50 cold portal submissions.
Reprice One Client on Sprint Cadence Instead of Hours
- 1
Pick your next client proposal. Instead of scoping by hours, define the engagement as three 2-week sprints with a fixed deliverable and acceptance criteria for each.
- 2
Price each sprint at 1.5x what you would have charged hourly for the same scope. Add a clause: if the sprint deliverable is approved in under 5 days, the next sprint begins immediately.
- 3
At sprint kickoff, run a 30-minute outcome interview: what does success look like? What metrics change? This replaces the scope document and creates shared accountability.
- 4
Track actual hours spent per sprint for your own records only. Do not share this with the client. Compare your effective hourly rate to your previous model.
- 5
At project close, ask the client: did paying per sprint change how you engaged with the work? Their answer is your best marketing copy for the next proposal.
Stretch goal: Build a simple calculator (Claude can write it in 10 minutes) that converts your hourly rate into sprint pricing with built-in buffer and outcome bonuses. Share it in your next sales call as a pricing tool, not a contract.
Build Your First Internal Tool This Week
- 1
Identify one internal tool you are currently paying for or evaluating. Give it three non-negotiable requirements.
- 2
Open Bloom and describe the tool in one or two sentences. Add your three requirements as constraints.
- 3
Let the agent build. Don't touch the code — just describe what's missing or wrong and iterate.
- 4
Share the result with one teammate via the App Clip link. Watch them open a native app from a single tap.
- 5
Note: how long did it take? What does this app do that the off-the-shelf version doesn't? How does it feel to own the data?
Stretch goal: Look at your SaaS bill and ask: what else on this list could I build in an afternoon? The answer might surprise you.
Build a Private RAG Knowledge Base for Under $20/Month
- 1
Gather your 20 most important internal documents: SOPs, onboarding guides, product specs, FAQs. Export them as plain text or markdown files.
- 2
Use Claude or Cursor to scaffold a simple Python RAG app using LlamaIndex or LangChain with a local or low-cost vector store (ChromaDB is free, Qdrant has a generous free tier).
- 3
Point the ingestion script at your documents folder. Run it. Your knowledge base is now queryable in natural language.
- 4
Add a basic chat interface — Streamlit or Gradio gets you there in under 30 minutes. Deploy to a free Render or Railway instance.
- 5
Share the URL with your team. Track which questions they ask. The most frequent questions reveal your biggest documentation gaps.
Stretch goal: Apply Morphos.ai's Green Vectors insight: test whether chunking your documents by semantic topic rather than fixed token size improves retrieval accuracy. Compare results on 10 test queries and note the difference.
Run the 60-Minute Tool Stack Elimination Sprint
- 1
Export your last three months of SaaS billing. Build a simple table: tool name, monthly cost, last login date, and primary use case.
- 2
For each tool, ask Claude to suggest two free or cheaper alternatives that cover 80% of the use case. Add those to the table.
- 3
Score each tool 1-3: 1 = could eliminate today, 2 = could replace in a month, 3 = genuinely irreplaceable. Eliminate or cancel all tools scored 1 immediately.
- 4
For every tool scored 2, create a 30-day migration task with a specific replacement in mind. Add it to your project management system.
- 5
Total the potential monthly savings. If it exceeds $500, that is the budget for one AI tool that actually moves the needle.
Stretch goal: Apply Software Finder's unbiased matching principle: before renewing any tool, describe your requirements anonymously to an AI and ask it to recommend the best fit without knowing what you currently use. Compare its recommendation to your renewal instinct.
Build a White-Label SaaS Stack for a Trade Niche in 48 Hours
- 1
Pick one trade niche where you have a contact or relationship: HVAC, roofing, landscaping, plumbing, electrical, cleaning. Interview one operator for 30 minutes about their weekly admin pain.
- 2
Map the five tasks they hate most. For most trade businesses, it is: quoting, follow-up, scheduling, invoicing, and Google reviews.
- 3
Sign up for GoHighLevel (14-day trial). Build a sub-account with: a lead capture form, a 3-step SMS follow-up sequence, a calendar booking page, an invoice template, and an automated Google review request triggered by job close.
- 4
Record a 5-minute Loom walking through the system. Send it to your contact. Ask: would you pay $200/month for someone to set this up and manage it for you?
- 5
If yes, you have a business. If no, ask what they would pay and what they actually need. That feedback is worth more than the build.
Stretch goal: Kai Stone's core insight: your first client is not a customer, they are a co-founder of your niche playbook. Document every customization they request — those are the features your second client will also want.
Turn a Business Requirement Into a Working App in 72 Hours
- 1
Write a one-paragraph business requirement in plain English: who uses this, what problem it solves, what they do with it, and what success looks like. No technical specs.
- 2
Open Woz or your AI builder of choice. Paste only the business requirement — do not describe the technical architecture. Let the AI infer the stack.
- 3
Review the first build for business logic accuracy, not code quality. Ask: does this do what the user actually needs? Iterate on requirements, not code.
- 4
Share a link with one real potential user within 24 hours of starting. Do not explain it. Just watch them try to use it.
- 5
After watching the first user session, write three sentences about what confused them. Those three sentences become your next build prompt.
Stretch goal: Apply Woz's backwards-build principle: define what the user says after a successful interaction before writing a single prompt. Work backwards from that outcome statement to determine what the app needs to do.
Build Your First Design Token System This Weekend
- 1
Open your most-used product screen. Catalog every visual decision: background color, text color, button color, border radius, font size, spacing unit. Write each one down.
- 2
Use Claude to generate a JSON design tokens file from your catalog. Each value should have a semantic name (e.g., 'color.interactive.primary') not a raw value ('#3B82F6').
- 3
Ask Claude to generate a simple React or Tailwind component library from those tokens: button, input, card, badge. Each component should only use token names, never raw values.
- 4
Swap one page of your product to use only the new component library. Note every place where the page broke — those are your undocumented design decisions.
- 5
Share the component file with your designer or a trusted peer. Ask them to build one new screen using only those components. Their questions reveal your token gaps.
Stretch goal: Knapsack's core insight: the distance between the person who identifies a problem and the thing that gets built should be zero. After this experiment, ask: how many handoffs exist between a new design decision and production? That number is your design system debt.
Find a Government Compliance Gap in Your Niche and Own It
- 1
Pick one industry where you have domain knowledge. Search for recent legislation (last 3 years) that added a new compliance requirement for businesses in that space.
- 2
Talk to five operators in that industry. Ask: how are you handling this compliance requirement today? Spreadsheet, paper, nothing, or a tool?
- 3
If the answer is spreadsheet or nothing for three or more of them, you have a product idea. Describe the workflow they need in plain English.
- 4
Build a minimal version using a no-code tool (Glide, Softr, or Bubble for data apps) or ask Claude to scaffold a Next.js app. The first version only needs to do one thing: track the compliance requirement.
- 5
Charge $99/month. If five people pay it in the first month, raise it to $199. Keep raising until the first no.
Stretch goal: Andy Seth's principle: the best SaaS niches are ones where the customer is not the user. In the Apprentices.io model, the contractor pays but the government inspector is the de facto enforcer. Find the regulator behind your niche and make compliance with their requirements automatic.
Build Your Own Live Pitch Scoring Tool
- 1
Open Claude or your preferred AI coding tool. Prompt it: 'Build me a single-page pitch scoring web app with three sections: Founder, Business, and Gut Check. Each criterion scored 0–3. At the end, show me a summary of weak spots and generate three investor questions I should prepare for based on my lowest scores.'
- 2
Score yourself honestly in the Founder section: storytelling, emotional intelligence, learning agility, domain expertise, conviction, and coachability.
- 3
Score honestly in the Business section: defensibility, ICP clarity, business model strength, TAM credibility, and tech differentiation.
- 4
Complete the Gut Check: moat type (community, data, brand, or none) and whether you would invest your own money.
- 5
Give the tool to a trusted advisor. Have them score you independently. Compare the gap between your self-assessment and theirs.
Stretch goal: Run every co-founder through the same scoring independently. The areas where your assessments diverge most are worth a dedicated conversation before you walk into any investor meeting.
Run the Bootstrap-First Decision Framework Before Your Next Raise
- 1
Write down the three things you would do with the raise. Be specific: hire X, build Y, spend Z on marketing. Now ask: which of these can I do at 10% of that cost to validate the assumption first?
- 2
Calculate the minimum ARR at which you would have real negotiating leverage with investors. Denver Ventures' threshold: $500K ARR is where the conversation changes.
- 3
Map your current runway to that target. If you can reach $500K ARR before running out of money, you should try. If you cannot, you need to raise — but know why.
- 4
List every investor meeting you are considering. For each one, write the specific thing you want from them beyond money: intro network, domain expertise, operational help. If the answer is only money, deprioritize them.
- 5
Write a one-page bootstrapped product plan: what you build, who pays first, at what price, and what that first cohort proves. This is your leverage document even if you raise.
Stretch goal: Carson's most important insight: the founder who walks in with $500K ARR is negotiating. The founder who walks in with a great deck is begging. Build the document that turns you into the first kind of founder.
Build Your 18-Month Exit Readiness Scorecard
- 1
Ask Claude to generate a 20-question exit readiness assessment covering: clean cap table, documented processes, revenue predictability, customer concentration risk, IP ownership, team dependency on founders, and data room completeness.
- 2
Complete the assessment honestly. Score each item 0 (not started), 1 (in progress), or 2 (complete). Any item scored 0 that relates to revenue, IP, or cap table is a blocker — address it first.
- 3
Build a simple data room folder structure today even if it is empty: financials, legal, customer contracts, IP documentation, team org chart, product roadmap. Populate what you can in one hour.
- 4
Identify the one person at your most likely acquirer who would lead the deal. Find them on LinkedIn. Note the kind of content they engage with and the language they use to describe acquisitions.
- 5
Write a one-paragraph acquisition narrative from their perspective: why does buying your company make their life easier? What problem do you solve for their portfolio or strategy?
Stretch goal: Alyssa's insight: exits happen to founders who are prepared before the conversation starts. Set a 90-day calendar reminder to review and update your data room. The best time to prepare for an exit is 18 months before you want one.
Build a Soft Skills Simulation in a Weekend
- 1
Pick one high-stakes conversation relevant to your industry: a customer calling to cancel, a job candidate receiving a rejection, a patient asking about a diagnosis. Write a one-paragraph brief: persona, emotional state, what they want.
- 2
Use Claude as the simulation engine. System prompt: 'Play this persona, respond dynamically, track tone and empathy markers, and note whether the user is validating or dismissing the persona's concerns.'
- 3
Set a turn limit of 8–10 exchanges. After the conversation ends, send the transcript back to Claude with a new prompt: score across emotional validation, clarity of communication, and de-escalation effectiveness. Return as JSON with written explanation.
- 4
Use Cursor, Replit, or Bloom to wrap it in a basic chat UI with a post-session report screen. The full stack should be runnable in under 48 hours.
- 5
Run two people through the same scenario. Compare their transcripts and scores. Note the variance — that gap is the training opportunity.
Stretch goal: Add a second scenario and track improvement across sessions. You now have a working behavior change loop — the core of any simulation-based training product.
Run Your Activation Debt Audit: Map Time-to-Value for Your First 10 Users
- 1
Define your product's 'first value moment' — the exact action that tells you a user understands why they signed up. Be specific: not 'created an account' but 'ran their first report' or 'sent their first message.'
- 2
Pull data on your last 10 sign-ups. For each user, find the timestamp of sign-up and the timestamp of that first value moment. Calculate the gap in hours.
- 3
Map every step between sign-up and first value moment. Write each one on a sticky note or doc. Count them. Anything above 5 steps is almost certainly killing activation.
- 4
Identify the single step with the highest abandonment. Use Hotjar, FullStory, or a session recording tool. Watch 5 recordings of users who dropped off at that step.
- 5
Remove or automate that step. If you cannot remove it, write a better instruction. Measure the activation rate on the next 10 sign-ups.
Stretch goal: Quarterzip's core insight: the AI cursor — doing things on behalf of the user — is the end state of onboarding. For your product, identify one setup step you could complete on the user's behalf with their permission. Remove the instruction and just do it.
Map Your Vibe Coding Failure Points and Build a Human Recovery Protocol
- 1
Think back to your last three build sessions that stalled. Write down exactly where you got stuck: the error message, the concept you didn't understand, the tool that behaved unexpectedly.
- 2
Categorize each failure: (1) AI generated wrong code, (2) missing domain knowledge, (3) wrong tool for the job, (4) unclear requirements, (5) environment setup issues. Most failures fall into one of these.
- 3
For each category, write a one-paragraph recovery protocol: the first three things you do when you hit this type of wall, before giving up.
- 4
Build a personal 'unstuck' doc. Pin it in your browser. The moment you feel flow leaving your body, open it.
- 5
Share the recovery protocol with one other builder. Ask them to add their own failure modes. The combined doc is now a team knowledge base.
Stretch goal: Danny Newman's insight: what you actually need when stuck is a real person who knows the exact tool you're using — available now. Before building your recovery protocol, spend one session on On Demand Human or a similar live help platform. What you learn in 15 minutes will rewrite your protocol.
Build an AI-Assisted Supply Vetting Tool
- 1
Build a script (Node.js or Python) that takes a supplier name and country as inputs. Use a search API — Perplexity, Tavily, or Exa — to pull publicly available information about that supplier.
- 2
Feed that information into a Claude API call with this system prompt: 'You are a supply qualification analyst for a vetted marketplace. Produce a structured brief covering: (1) verifiable credentials, (2) named practitioners and qualifications, (3) red flags including review anomalies, (4) community sentiment, and (5) an overall trust signal score from 1–10 with reasoning.'
- 3
The output becomes a pre-visit research brief any team member can use before committing to an in-person site visit.
- 4
Add a simple UI with Cursor or Replit — an input form, a loader, and a formatted output card. Entire build should take under 48 hours with AI-assisted coding.
- 5
Log your actual vetting decisions against the AI brief scores over time. Use that data to refine the scoring prompt and improve calibration.
Stretch goal: This architecture works for any marketplace requiring supply qualification — freelancers, contractors, financial advisors, wellness practitioners. The AI compresses the research phase so human judgment can be applied where it counts.
Build Your Own Support Flywheel in 48 Hours
- 1
Open your AI coding tool of choice (Claude, Cursor, Replit). Prompt it: 'Build me a simple Python chatbot using a JSON knowledge base. When a user asks a question the bot cannot answer, log the question with a timestamp to a CSV file called unanswered.csv. Give me a basic web interface using Flask so my team can use it in a browser.'
- 2
Populate the JSON knowledge base with your 20 most common customer or team questions and answers. Your support inbox, Slack history, and FAQ page are gold mines for this.
- 3
Run the bot for five business days. Have your team use it instead of emailing you or each other for those common questions.
- 4
Open unanswered.csv. Sort by frequency. The questions that appear most often are your highest-value automation targets and your roadmap for what to build or buy next.
Stretch goal: You are replicating the core insight behind the Capacity flywheel at zero cost — letting real usage tell you exactly where the leverage is. That is the problem behind the problem, surfaced in 48 hours.
Build a Repo Consistency Agent
- 1
Create a GitHub repo and add your marketing copy, product spec, engineering spec, and QA docs as individual markdown files.
- 2
Open Claude Code and prompt it: 'Build a skill that reads all markdown files in this repo and outputs a structured list of contradictions, omissions, and misalignments between them. Run it on a schedule and flag any drift whenever a document changes.'
- 3
Write a constraints file — a markdown doc that tells the agent what consistent, on-brand output looks like. Ask Claude what constraints it needs to produce reliable output.
- 4
Have the agent reference the constraints file every time it runs. When any document changes, the agent checks for drift before it ships.
- 5
To go deeper: feed your reading sessions into a JSON blob, then have an agent run entity analysis, topic analysis, and sentiment scoring automatically.
Stretch goal: Git already provides file change tracking, identity attribution, and cryptographic provenance. You are building a security and audit infrastructure on top of something you already own.
Ship a Single-Purpose Agent This Week — Then Kill It If It Doesn't Work
- 1
Pick one recurring task that takes you 15 minutes or more per day. Write it down: the inputs, the decision you make, and the output you produce.
- 2
Define the agent's scope in one sentence. It does exactly one thing. No fallbacks, no related tasks, no scope creep.
- 3
Build it in under 4 hours using Claude Code, n8n, or Make.com. If it takes longer, the scope is too broad — cut it in half.
- 4
Run it for 5 business days. Track: time saved per day, error rate, and one instance where it got something wrong. Document the failure.
- 5
At the end of day 5, make a binary decision: keep it, kill it, or redesign it. James Everingham's rule: fast failure with a specific lesson is worth more than slow success with no data.
Stretch goal: Build a second agent that monitors the first one. The watcher agent logs every output and flags anomalies. You now have the beginning of an agent oversight layer — the architecture beyond Identity is trying to productize.
Run Your AI Data Exposure Audit in 48 Hours
- 1
List every AI tool your team has used in the last 90 days. For each one, note what data you submitted: documents, code, customer data, internal processes, financials.
- 2
Pull the Terms of Service for each tool. Ask Claude to summarize the data retention, training, and ownership clauses for each one in plain English. Note any tool that claims rights to your inputs.
- 3
Score each tool on a 3-point risk scale: 1 = input used for training, 2 = data retained beyond session, 3 = third-party data sharing. Cancel or replace any tool scored 3 immediately.
- 4
For your highest-risk tools, define a data hygiene protocol: what can be submitted, what must be redacted, and who is allowed to use the tool for sensitive work.
- 5
Document your findings in a one-page AI Data Policy. This is the foundation of your security posture when a larger customer asks about your AI usage.
Stretch goal: John's insight: the question is not whether you use AI, but whether you know what you are giving away when you do. Extend this audit to your vendors — ask every SaaS tool you use how they are using AI internally and what data they feed into it.
Design an Outcome-Aligned Pricing Experiment for Your AI Product
- 1
Define the one measurable outcome your product delivers for customers. Be specific: not 'improves efficiency' but 'reduces X by Y% within Z weeks.'
- 2
Identify the current market baseline for that outcome. What does that metric look like without your product? Find data — from industry reports, customer benchmarks, or your own case studies.
- 3
Write an outcome-aligned pricing proposal: your fee is a percentage of the measured improvement above baseline. Present this to your next two prospects alongside your standard pricing.
- 4
Build a simple measurement framework: how will you and the customer jointly measure the outcome at 30, 60, and 90 days? Get agreement on the measurement method before pricing is discussed.
- 5
Track which customers prefer outcome pricing vs. flat fee. Outcome pricing self-selects for customers who believe in the product and are willing to be held accountable for usage.
Stretch goal: Axenya's model: 95% of clients saving money for 3+ consecutive years is the result of only getting paid when it works. Consider building a free tier with outcome-based upgrade: customers pay nothing until they hit a defined value threshold. This inverts the customer acquisition cost model entirely.
Design Your 10-Year AI Integration Map in One Afternoon
- 1
Write down your product's core value proposition as it exists today. Now write what it needs to look like in 10 years to still be relevant. Note every assumption that changes between those two versions.
- 2
Map the AI capabilities that currently exist at a research or demo level but are not yet production-ready. Pick the one most likely to be production-ready in 3 years. How does it change your product if it is?
- 3
Identify the one workflow in your product that would benefit most from agentic AI — not a chatbot, but an agent that takes action. Define what that agent would do, what inputs it needs, and what oversight it requires.
- 4
Build a constraints document: what can your AI agents never do without human approval? This is your governance layer, and it is faster to define now than after an incident.
- 5
Present your 10-year map to a customer. Ask them: which of these would make you switch from our competitor today? Their answer tells you your 12-month roadmap.
Stretch goal: Eugene Sayan filed an agentic AI patent in 1998 — 25 years before the current wave. His 10-year problem design horizon is the reason Softheon is still relevant. Ask yourself: what problem does your product need to solve in 10 years that no one is thinking about today?
Audit Your AI Answer Engine Visibility This Week
- 1
Ask ChatGPT, Claude, and Perplexity the top three questions your ideal customer asks before buying a product like yours. Note whether your company or content appears in any answer.
- 2
For each query where you do not appear, find the source that does — it is almost always a piece of content: a blog post, a Reddit thread, a LinkedIn article, or a review site.
- 3
Create one piece of content specifically designed to answer the top question in your category. Publish it on your own domain. Format it as a direct, declarative answer — AI models prefer structured, authoritative responses.
- 4
Submit your content to any aggregators or directories that appeared in the AI answers. Being cited in those sources increases your probability of appearing in AI-generated answers.
- 5
Re-run the same three queries in 30 days. Track whether your content or brand now appears. This is your AI search share of voice baseline.
Stretch goal: Colin McIntosh's SheetsResume insight: being the #1 result on Google Gemini for a high-intent query is worth more than being on page 2 of traditional search. Build a list of 10 queries where your ideal customer is searching in AI tools right now. Own one of them this week.
Run Your 5-Question AI Brand Visibility Audit
- 1
Define your category in one sentence as a customer would describe it — not how you would. Use that language in every query.
- 2
Open ChatGPT, Claude, Perplexity, and Google AI Overviews. Ask each the same five questions your best customers asked before buying from you. Screenshot every response.
- 3
Score each response: does your brand appear? Does a competitor appear? Does a third-party review or comparison site appear? Build a simple tracker.
- 4
For every query where a competitor appears and you do not, find the specific content asset that earned them that mention. Reverse-engineer its structure.
- 5
Publish one content asset this week that directly targets a query where you are invisible. Measure AI citation rate monthly.
Stretch goal: Justin Inman's 5-agent framework (Scout, Pilot, Flow, Merchant, Echo) is a systematic answer to 'where does my brand appear in AI?' Build the Scout layer yourself first: a weekly script that queries 3 AI tools for your top 10 brand-relevant questions and logs the results to a Google Sheet.
Map the Relationship Arbitrage Nobody Else Is Chasing
- 1
List the 10 people in your network who connect seemingly unrelated worlds: the healthcare operator who also advises startups, the developer who has deep relationships in real estate, the marketer who spent 5 years in logistics.
- 2
For each person, write the intersection they represent: two domains that are rarely connected but could unlock outsized value if combined. That intersection is a potential business or partnership.
- 3
Pick the most interesting intersection. Write a one-paragraph thesis for why an AI product built at that crossing would be hard to replicate — who would build it, and why don't they?
- 4
Send a message to the connector: 'I've been thinking about X and Y. Do you know anyone building at that intersection?' Their answer either validates your thesis or saves you months of research.
- 5
Follow Josh Furstoss's protocol: the best AI companies are not being built by the people you'd expect, in the markets you'd expect. Write down the unexpected builder for your thesis — then become that person.
Stretch goal: Furstoss launched four companies in one year. His method: identify the relationship others overlook, build the smallest possible thing that proves the value of the connection, and share it immediately. What is the simplest experiment you can run at your chosen intersection this week?
Design a Real-World Failure Gauntlet for Your AI Feature Before Launch
- 1
Write down every assumption your AI feature makes about the environment it will run in: network conditions, hardware, data quality, user behavior, lighting, noise, or edge cases.
- 2
Design three adversarial test scenarios that violate those assumptions. For Marcin's robotics team, this was a real airport at 3am with moving crowds and unpredictable surfaces.
- 3
Run your feature in each adversarial scenario before any user sees it. Log every failure mode with a timestamp and exact failure description.
- 4
For each failure, classify it: (A) unacceptable in production, (B) acceptable with a warning, or (C) acceptable as-is. Fix every A before launch.
- 5
Publish your failure gauntlet results internally. The team that knows how their product breaks in the real world ships better products than the team that only knows how it works in demos.
Stretch goal: Seven Sense's Series A unlock came from shipping to a real cleaning machine in a real airport. The first customer was the validation. Identify your equivalent: the single real-world deployment that, if successful, proves your core technical claim to every subsequent investor and customer.
Run a 5-Customer Truth Sprint Before You Write a Line of Code
- 1
Identify five people who match your ideal customer profile. They do not need to be warm leads — cold outreach works if your ask is small and specific.
- 2
Send each person a single question: 'What is the most frustrating part of [the problem you're solving]? I'm trying to understand it before building anything.' Do not describe your solution.
- 3
After the response, ask one follow-up only: 'What have you tried to solve it?' The answer tells you the graveyard of solutions that already failed.
- 4
Compile the five responses. Highlight every phrase that appears in two or more answers — those are your real requirements. Anything that only appears once is a nice-to-have.
- 5
Write a one-paragraph product description using only language from the five interviews. No marketing language. If you can't describe it in their words, you do not understand the problem yet.
Stretch goal: ThinkUp's insight: the most dangerous thing you can build is a solution to a problem you perfectly understand but your customer only partially feels. Run this sprint again after every major product change to stay calibrated.
Map Your Domain's Data Fragmentation Problem Before Writing Code
- 1
Pick one industry you know deeply. List every data source an expert in that field uses to make decisions. Be exhaustive: tools, spreadsheets, reports, emails, verbal briefings.
- 2
For each data source, ask: does it talk to the others? Score it: 0 = fully isolated, 1 = manual export/import, 2 = partial integration, 3 = real-time sync. If most are 0 or 1, you have found a fragmentation problem.
- 3
Interview three experts in that field. Ask them: 'What decision do you make weekly that requires you to pull from more than one of these sources?' That decision is your product's core use case.
- 4
Write a one-paragraph description of the fragmentation problem from the expert's perspective, not the technology perspective. Share it with the three people you interviewed. Ask: is this accurate?
- 5
The product that solves their fragmentation problem is worth more than the one that adds a new data source. Resh burned $30M finding this. Your interview costs nothing.
Stretch goal: Resh's pattern — Mercatus, then Diagnostic MD — is always the same: find a domain where siloed data is drowning experts, aggregate it, and let AI surface the insight. The domain you know best is the one worth exploring first. What would a version of Diagnostic MD look like in your field?
Run the 3-Customer Reality Check Before Your Next Sprint
- 1
Before writing your next sprint's tickets, identify three customers who represent your ideal buyer. Schedule a 20-minute call with each this week.
- 2
In each call, ask only: 'What did you try to do in our product last week that didn't go the way you expected?' Do not ask what features they want. Ask what failed.
- 3
After all three calls, write down every failure they described. If a failure appears in two or more calls, it is a P0 — it goes to the top of the sprint. No exceptions.
- 4
Cut one item from your existing sprint backlog for every P0 you add. The sprint length does not change. The priorities do.
- 5
At sprint close, review: did fixing the real failures change any of your roadmap assumptions? Document what changed. This is your calibration record.
Stretch goal: Mike Vitez's pattern: founders fall in love with their idea and skip validation. The 3-customer call is the vaccine. Set a standing calendar block every two weeks for three customer reality checks. Make it non-negotiable before any sprint begins.
Replace Your Next Dev Brief with an AI Requirements Interview
- 1
Before writing your next feature spec or dev brief, open Codalio or Claude. Describe your product, your user, and the problem you are trying to solve. Then stop typing.
- 2
Let the AI interview you: it will ask clarifying questions about user journey, edge cases, success metrics, and technical constraints. Answer each one honestly.
- 3
After the interview, ask the AI to generate a structured requirements document: user story, acceptance criteria, technical constraints, and out-of-scope items. Review it for accuracy.
- 4
Share the generated brief with your developer before any code is written. Ask them: is anything here ambiguous? Their questions reveal missing requirements you can resolve in 10 minutes instead of a two-week rework.
- 5
Compare the time spent on this brief vs. your previous manual spec process. Note whether the resulting build better matched your intent.
Stretch goal: Ehsan's insight: the gap between business objective and technical requirement is eating projects before they start. The AI requirements interview is not about the AI — it is about forcing the founder to articulate the problem precisely before anyone builds anything. That precision is the product.
Run a 60-Minute 2026 Founder Reality Audit
- 1
Open a blank document. Write your honest answers to five questions: (1) What is the one thing blocking your growth right now that you have been avoiding? (2) Which of your team's current workflows will an AI agent replace in 12 months? (3) What assumption about your business model has never been tested? (4) What would you cut today if you had to survive on 50% of your current revenue? (5) What are you building that your best customer has never asked for?
- 2
Share your answers with one person who will tell you the truth — not your co-founder, not an investor. A peer founder who has no stake in your answer.
- 3
For each answer, write one action that costs nothing but time and could change the answer in 30 days. These are your highest-leverage experiments.
- 4
Set a 30-day calendar reminder for each action. At the 30-day mark, write one sentence on what changed.
- 5
Apply Ran Aroussi's 30-year coder principle: the best diagnostics are the ones that surface what you already know but have not acted on. The audit does not reveal new information — it removes the excuses for avoiding what you already see.
Stretch goal: The multi-guest format of this episode is itself an experiment: three different operators, three different lenses on the same question. Run your own version — bring two trusted founder peers and each answer the five questions in the same room. The conversation that follows is the audit.
Generate a PRD from a Napkin Idea Using AI — Then Validate It Before Building
- 1
Write your idea in one sentence: who it helps, what it does, and why now. Do not describe the technology.
- 2
Open BrainGrid or Claude. Paste your one sentence and prompt: 'Generate a product requirements document for this idea. Include: user personas, core use cases, success metrics, technical constraints, out-of-scope items, and open questions.'
- 3
Read the PRD for one thing only: the open questions. These are the assumptions your idea rests on that have not been tested. Pick the three most important ones.
- 4
Design the smallest possible test for each open question. A Typeform survey, a landing page, a five-minute customer call. Run all three before writing any code.
- 5
After the tests, return to the PRD and update it with what you learned. A PRD updated by real evidence is worth ten times one built purely from assumptions.
Stretch goal: Tyler Wells' insight: vibe coding without a PRD is archaeology — you discover what you built after you build it. The BrainGrid approach is to make the invisible structure visible before the first prompt. Apply this to your most important current project: generate the PRD for something already in production and see what assumptions you made that have since proven wrong.
Run Your 72-Hour Unit Economics Reality Check
- 1
Pull your last 30 days of ad spend and revenue. Calculate your ROAS. Now subtract COGS, fulfillment, returns, and customer service costs from that revenue. What is your actual margin per order?
- 2
Calculate your true CAC: total marketing spend divided by new customers acquired, not total orders. Include your own time at a realistic hourly rate if you are running ads yourself.
- 3
Calculate LTV using your actual repurchase rate, not your hoped-for one. If you do not have repurchase data yet, assume it is 20% lower than your gut says.
- 4
Divide LTV by CAC. If the ratio is below 3:1, you are acquiring customers you cannot afford. Write down the one lever — price, COGS, retention, or CAC — that, if improved by 20%, gets you to 3:1.
- 5
Build a simple model in Google Sheets or ask Claude to build one: input your current numbers, and the model shows you how LTV:CAC changes as you move each lever. This model is more valuable than any dashboard.
Stretch goal: Adam Robinson's insight: the gap between ROAS and real profitability is where most e-commerce companies are quietly dying. The founders who survive are the ones who stop optimizing for the top-line number and start asking what every dollar of revenue actually costs them to generate and keep.
Build Your Founder Tax Deduction Intelligence System
- 1
Export your last 12 months of business expenses from your bank or accounting software as a CSV.
- 2
Ask Claude to categorize each transaction by IRS Schedule C line item. Prompt: 'Categorize each expense as one of the following IRS Schedule C categories: advertising, car and truck, commissions, contract labor, insurance, interest, legal and professional, meals (50%), office expense, rent or lease, repairs, supplies, taxes and licenses, travel, utilities, wages, or other expenses. Flag any that are unclear.'
- 3
Review the flagged items with your accountant. The goal is not to replace the accountant — it is to walk into that meeting with 80% of the categorization already done.
- 4
Identify the three largest expense categories and ask: am I capturing all deductible expenses in this category? Most founders underreport home office, vehicle use, and professional development.
- 5
Set a monthly calendar block to run this same categorization pass. Batch processing monthly takes 15 minutes. Doing it for a full year at tax time takes two days.
Stretch goal: Cy's insight from building Deduction.com: most founders leave 20-30% of their deductible expenses unclaimed — not through dishonesty, but through not knowing what qualifies. Build a personal deduction checklist from IRS Publication 535 (ask Claude to summarize it for your business type). Review it quarterly.
Build Your Personal Alternative Investment Readiness Score
- 1
Build a simple spreadsheet with four sections: (1) Emergency fund — how many months of expenses can you cover? (2) Liquid investments — what percentage of your savings is immediately accessible? (3) Fixed expenses — what is your minimum monthly burn? (4) Income predictability — how stable is your income over the next 24 months?
- 2
Apply CJ Follini's two-barrier test: do you have 2-3 years of budgeting history? If not, start tracking every dollar in and out today before considering any illiquid investments.
- 3
Score your readiness to allocate to alternatives: 0-40% liquid savings = not yet, 40-60% = could consider liquid alternatives only, 60-80% = ready for limited illiquid allocation, 80%+ = follow the Yale Model.
- 4
Pick one alternative asset class from the 14 CJ identifies that matches your existing knowledge. Real estate if you understand property, art if you have domain knowledge, venture if you understand tech. Start with the one where your due diligence advantage is highest.
- 5
Read the offering documents for one deal in that category — even if you do not invest. Learning to read an operating agreement or term sheet takes one afternoon and pays dividends for life.
Stretch goal: CJ's principle: the sequence is financial literacy, then budgeting discipline, then portfolio optimization, then alternative allocation. If you skip steps, you are not investing — you are gambling with a longer time horizon. Run this readiness score quarterly and track which step you are on.
Build a Daily Voice Memory System and Train It on Your Own Thinking
- 1
For one week, record a 3-minute voice note at the end of each day answering the same three questions: What surprised me today? What did I decide, and why? What do I want to think about tomorrow?
- 2
At the end of the week, transcribe all five recordings using Whisper (free API) or any voice-to-text tool. Paste them into a single document.
- 3
Ask Claude to identify: recurring themes across the five entries, decisions you made with low confidence, and questions you raised but never answered.
- 4
Build a simple folder structure: /voice-notes/YYYY-MM-DD. Set a phone reminder for 9pm daily to record. This is your raw thinking, not performed for anyone.
- 5
After 30 days, run the same Claude analysis on the full month. The patterns it surfaces will surprise you — they represent your actual priorities, not the ones on your task list.
Stretch goal: ThinkUp's insight: social media optimizes your memory for external validation. A private voice system optimizes it for your own clarity. After 90 days, use your voice note archive as the training data for a personal AI assistant that knows how you actually think. The context it provides will make every AI interaction more useful.
Run the AI Passenger Audit on Your Team This Week
- 1
Ask every team member to fill out a 5-minute anonymous survey: (1) Which AI tools do you use daily? (2) Name one task you completed faster because of AI this week. (3) Name one task you tried to use AI for and it failed. (4) What is the most important thing your role requires that AI cannot do?
- 2
Score the results by role. Anyone who answers question 2 with 'none' is a passenger. That is not a firing offense — it is a training opportunity.
- 3
For every passenger, identify one specific task in their role that AI can demonstrably accelerate. Design a 30-minute demo using real data from their actual job. Show, do not tell.
- 4
Greg Shove's threshold: you want 30-50% AI adoption within a team, not 10%. Map your current adoption rate by function. The functions at 10% are your highest-leverage improvement areas.
- 5
Create a monthly AI wins share-out: a 10-minute team meeting where one person demonstrates something that changed how they work. Rotate presenters. Recognition drives adoption faster than mandates.
Stretch goal: Section's insight: the competitive advantage is not the tools — everyone has access to the same tools. It is redesigning how humans and AI work together to do things neither could do alone. For your lowest-adoption function, map what a redesigned human-AI workflow would look like. The redesign is the experiment.
Design One Human-AI Collaboration Loop Your Team Can Run This Week
- 1
Pick one decision your team makes weekly that requires judgment: a prioritization call, a customer escalation response, a content approval. This is your target workflow.
- 2
Map the current workflow: inputs, decision maker, time spent, error rate if you know it. Write it down in under 10 minutes.
- 3
Design an AI-assisted version: which part does the AI do, which part does the human do, and where does the human override the AI? Be specific about the handoff point.
- 4
Run the new workflow for two weeks with one person. Track: time saved, decision quality (did the outcome match the goal?), and any moments where the human overrode the AI and was right to do so.
- 5
After two weeks, document what the AI got wrong and why. Use that failure analysis to improve the prompt, the constraints, or the human review step. This is Geoff Gibbins' CRI (Collaboration Readiness Index) in practice.
Stretch goal: Geoff's observation: most enterprise AI investment is FOMO-driven. The real competitive advantage is redesigning how you work — and that starts with one documented workflow, not a company-wide rollout. Run this experiment on your highest-value weekly decision first. The ROI there will fund the rest of the transformation.
Run the AI-Proof Leadership Skills Audit for Your Team
- 1
List every member of your team. For each one, score them 1-5 on the four skills Uma identifies as AI-resistant: (1) strategic influence — can they move people without authority? (2) visibility — are they known for their work beyond their immediate team? (3) communication — can they explain complex things simply under pressure? (4) emotional intelligence — do they read rooms and respond to what is actually happening?
- 2
Any team member scored below 3 in two or more areas is at risk of being sidelined as AI raises the floor on technical execution. That is your coaching pipeline.
- 3
For your highest-potential team member with the lowest leadership scores, design one visible assignment: a cross-functional project, a customer presentation, or an internal talk. Something that exercises the weak skill.
- 4
Ask Claude to generate a 30-day development plan for one skill area. Review it with your team member. The best development plans are built with the person, not for them.
- 5
Set a 90-day review. Uma's six-month Sought After Leader Experience exists because real behavior change takes time. Your experiment is the on-ramp, not the destination.
Stretch goal: Uma's uncomfortable thesis: AI is standardizing technical skill at near-zero cost, and what it's exposing is that most leaders never developed the capabilities that can't be commoditized. The founders who thrive in 2026 are the ones who are building influence, not just products. What is one thing you can do this week to become more visible outside your company?
Build a Private Memory Archive in a Weekend
- 1
Choose one year of your life — preferably the most recent full year — that you want to preserve meaningfully. Pull all photos from that year into a single folder.
- 2
Use Claude to generate a structured reflection prompt for that year: major decisions made, people who influenced you, beliefs that changed, moments you want to remember in 20 years.
- 3
Record your answers to each prompt as voice notes. Do not type them — voice captures the texture that text edits out.
- 4
Build a private Notion or Obsidian vault with one page per month. Embed or link the photos. Paste the voice note transcripts. Add one paragraph of context for anything that would need explanation in 10 years.
- 5
Share nothing publicly. This archive is for you and whoever you choose to give access to. Its value is in its privacy.
Stretch goal: Josh Furstoss's insight: social media deletes the meaning from memories by optimizing them for an audience. The private archive is the opposite architecture — it preserves the truth of an experience rather than the performance of it. Build the archive for one year, then decide whether to make it an annual practice.
Audit Your Creative Process for Algorithm Capture
- 1
For every piece of content you have published in the last 90 days, write down the honest reason you made it: genuine creative impulse, or because you thought it would perform?
- 2
Identify the three metrics you use to judge whether content succeeded. Now ask: if those metrics did not exist, what would you make instead?
- 3
Create one piece of content this week entirely divorced from performance metrics. No tracking, no A/B test, no engagement monitoring for 72 hours. Publish it and walk away.
- 4
After 72 hours, note how the content performed without the usual optimization. Then note how it felt to make it without the metrics pressure.
- 5
Build a ratio for your content calendar: for every 4 pieces optimized for distribution, make 1 piece optimized for truth. Track whether your audience responds differently to each over 90 days.
Stretch goal: The AI music renaissance insight: the tools that commoditize creativity also create the conditions for authentic creativity to be worth more. The artists who survive algorithmic saturation are the ones who are unmistakably themselves. What is the one thing about your voice that an AI cannot replicate? Build more of that.
Build a Personalized Insight Extraction System for Audio Content
- 1
Pick 10 podcast episodes you have listened to in the last 90 days that were genuinely valuable. For each, write down the single most important insight — not a summary, but the specific thing that changed how you think.
- 2
Run those 10 insights through Claude with this prompt: 'Identify patterns across these insights. What recurring themes appear? What tensions exist between them? What question do they collectively suggest I should be asking?'
- 3
Build a personal insight tracker: a simple Airtable, Notion database, or even a Google Sheet. Columns: source, date, insight, category, and a link to where you can find it again.
- 4
Set up a Whisper API script (or use MacWhisper for a no-code version) to auto-transcribe any audio you flag as worth reviewing. Pipe the transcript to Claude for a structured insight extraction.
- 5
After 30 days, run the same pattern analysis on your full insight tracker. The recurring themes are your actual learning priorities — not the ones on your reading list.
Stretch goal: Snipd's breakthrough insight: crowdsourced snipping creates a quality signal for podcast content that Spotify's listening-time data cannot capture. Build your own version: share your weekly top insight with three founder peers and ask for theirs in return. A five-person insight exchange is a private signal layer that compounds over time.
Build a Generative Music Reflection Experiment in a Weekend
- 1
Identify one relationship in your life — professional or personal — that has unresolved tension or distance. Write a one-paragraph description of the emotional state of the relationship as it is today.
- 2
Use an AI music generation tool (Suno, Udio, or similar) to generate a piece of music that reflects that emotional state. Do not think about what it should sound like — just describe the feeling in plain language.
- 3
Listen to the generated piece privately. Note whether anything shifts in how you think about the relationship after hearing it. This is Ziah Orion's core thesis: the emotional distance between thought and sound creates space for something new to emerge.
- 4
Generate a second piece that reflects where you want the relationship to be. Note the difference between the two. That difference is your work.
- 5
Share the experiment result — not necessarily the music — with one person you trust. Ask them to try the same exercise with something unresolved in their own experience.
Stretch goal: Deep Gem's insight: music is connection technology, not entertainment technology. The experiment above is not about AI music — it is about using creative tools to process what verbal language alone cannot reach. Extend the experiment to your team: what unresolved tensions in your company might benefit from a non-verbal processing tool?
Run Your Day-Zero IP Validation Sprint
- 1
Pick your most important current product idea or the next feature on your roadmap. Write it in one paragraph: what it does, who uses it, and what makes it different from what exists today.
- 2
Open SenseIP and run your idea through Leo. Do not filter or polish the description — paste your raw thinking. The goal is to see what the patent database already knows about your space.
- 3
Review the prior art Leo surfaces. For each result, write one sentence: how does your approach differ from this patent? If you cannot answer, that is a gap to address before building.
- 4
Ask Leo to help you identify what is most novel about your idea. Use the feedback to sharpen your description — not to file a patent necessarily, but to articulate your defensible differentiation before your next investor conversation or customer pitch.
- 5
If anything in the results looks like a freedom-to-operate concern — a granted patent that covers something you were about to build — flag it before any engineering work begins. A 10-minute check now costs nothing. Discovering infringement after launch costs everything.
Stretch goal: Ophir's protocol: before any customer conversation where you will describe your product, run the IP check first. The conversations you will have without NDAs are public disclosures — and once you disclose, you start the clock on your ability to file in many international markets. Add SenseIP to your validation stack the same way you added customer discovery interviews: not as a legal step, but as a product clarity step that makes your idea better.
Build a Personal Prediction Calibration System
- 1
For the next 30 days, log every non-trivial prediction you make: a hiring decision, a product bet, a sales forecast, a market call. For each one, write the prediction in one sentence and your confidence as a percentage (e.g. '70% — this feature ships by Friday').
- 2
At resolution, log whether you were right or wrong. After 30 entries, tally your results by confidence band: how often were your '70%' predictions correct? Your '90%'? This gap between stated and actual confidence is your calibration error.
- 3
Prompt Claude: 'I made these predictions over 30 days [paste log]. Identify patterns in where I am systematically overconfident or underconfident. What categories of decision consistently have the largest calibration gap?'
- 4
Pick your worst-calibrated category — the type of decision where your confidence most consistently doesn't match outcomes. Design one process change: a second opinion, a checklist, a waiting period. Apply it for the next 30 days and measure whether calibration improves.
- 5
Ben Turtle's Polymarket insight: real-money bets accelerate calibration because the cost of overconfidence is immediate. Choose one upcoming prediction and put something meaningful on the line — a public commitment, a small bet with a peer — to create the same forcing function.
Stretch goal: Ben's framing: every decision is a prediction. If you could measure the accuracy of your predictions the way LightningRod measures a model's Polymarket performance — profit, accuracy, calibration — what would your score be? Build the system to find out. The founders who compound fastest are the ones whose internal world model is most accurate, and the only way to improve accuracy is to measure it.
One-Prompt Your Entire Production Stack with the Lava MCP
- 1
List every tool in your current production stack that has an API: transcription services, show notes generators, CRMs, enrichment tools, data sources. Write each one on a single line with its API documentation URL.
- 2
Install the Lava MCP in your Claude Code instance and load your wallet. Then prompt Claude: 'Check which of these services are accessible through the Lava gateway. For any that are, list the endpoint name and pricing model.'
- 3
For each available service, write a single chained system prompt that takes one input (a recording file path or timestamp) and chains all downstream tasks: transcription, show notes generation, title options, and formatted asset output.
- 4
Run the workflow end-to-end on a real piece of content. Time the full run. Compare it against your current manual process time and note how many browser tabs and copy-paste steps it eliminated.
- 5
Log the actual cost in Lava wallet credits for the full workflow. Divide by the number of production steps completed. This is your cost-per-task number — the baseline you optimize against as you scale.
Stretch goal: Mitchell's core insight is that every API key in your stack is a tax on thinking. Extend the experiment: identify the three paid services you log into most often and ask whether they have MCP integrations or can be called via the Lava gateway. Each one you convert from manual to agentic is cognitive overhead you permanently eliminate.
Design a 30-Day Hybrid Cadence Experiment for Your Team
- 1
List every role on your team and classify each one: (A) benefits most from in-person presence (sales, onboarding, design collaboration, mentorship-dependent roles), (B) works best with protected remote focus time (deep coding, writing, analysis), or (C) works equally well either way. This is your baseline for designing the cadence.
- 2
For each A-category role, identify the three in-person interactions that create the most value: a standing team standup, a weekly pipeline review, a design critique session. These become your anchored in-person days.
- 3
Design a 4-week schedule that creates at least 2 days per week where your A-category team members are in the same space - and protects at least 2 days per week for heads-down remote work. Share the draft with the team before implementing.
- 4
Run the cadence for 30 days without changing it. At the end of week 2 and week 4, ask three questions: (1) Did you have the in-person interactions that mattered? (2) Did you have enough protected focus time? (3) Did commute days feel worth it? Score each out of 10.
- 5
Compare the week 4 scores to week 2. If specific days or team groupings score consistently low, adjust for month 2. Track one proxy metric throughout: either time-to-decision on cross-functional items or number of async threads that required follow-up versus resolved cleanly.
Stretch goal: Dan Bladen's data: Kadence customers average 1 extra hour given to work and 1 hour reclaimed for family per hybrid day versus full-time office. The key is that the flexibility is earned, not assumed - it exists because the team has designed explicit in-person rituals that justify it. If your team cannot articulate why the in-office days are valuable, they will not be.
Run a 90-Minute AI Data Readiness Audit
- 1
List every data source your company uses for AI-related tasks: CRM, customer support tickets, sales call transcripts, product analytics, contracts, billing records. For each one, write a one-sentence description of what it contains and how it currently feeds into any AI workflow.
- 2
For each data source, answer three questions: (1) Is it deduplicated - or do records exist in multiple places? (2) Is it consistently formatted - or do field values vary arbitrarily? (3) Is it accessible - can an AI system reach it programmatically, or is it locked in a legacy format or manual export?
- 3
Score each data source: green (AI-ready), yellow (needs moderate cleaning), red (blocked from AI use without significant work). Tally your results. This is your data readiness map.
- 4
For your most important red-category data source, estimate the cost of the workaround you are currently using: manual data entry hours per week times hourly cost. Annualize it. This is the baseline value of fixing it.
- 5
Pick one yellow-category source and run a 5-10 record sample through an AI extraction task - ask Claude to pull specific fields from a sample of records and compare accuracy against the actual data. The gap between expected and actual accuracy is your data quality problem made concrete.
Stretch goal: David Carmel's diagnostic: most companies discover they have the same data duplicated in dozens or hundreds of locations and that their data scientists are functioning as potato peelers - cleaning data by hand instead of building models. The 90-minute audit is not about fixing anything. It is about making the invisible cost of bad data visible enough that fixing it becomes a priority rather than a someday.
Map Your Unstructured Data Mountain and Run a Pilot Extraction
- 1
List every type of document your company receives, produces, or stores where humans currently extract information by hand: vendor contracts, customer intake forms, support emails, invoices, proposals, meeting notes. For each category, estimate: how many per week, and how many minutes a human spends processing each one.
- 2
Pick the highest-volume, highest-time-cost category from your list. Collect 10 representative sample documents from that category - make sure they include edge cases (unusual formats, incomplete fields, scanned documents, tables).
- 3
Prompt Claude or another LLM: 'Here is a document [paste content or describe structure]. Extract the following fields: [list 5-8 specific data points you actually need]. Return as a structured JSON object.' Run all 10 samples through the same prompt.
- 4
For each output, verify accuracy against the source document manually. Tally: how many fields were correctly extracted? What error patterns appeared? Were edge cases handled or did they fail completely? This is your baseline accuracy score.
- 5
Calculate the value of getting to 95%+ accuracy at scale: annualized time savings (hours per document x documents per year x hourly cost) minus the cost of the AI tool or service. If the multiple is greater than 3x, you have a use case worth building or buying a solution for.
Stretch goal: Mehul Shah's prediction: in 10 years, AI will be able to sweep all the PDFs in the world's largest document repository in under a day. The companies that will win that era are the ones who have already structured their document knowledge today - not because they need to process it all at once, but because the structured data will compound in value as AI capabilities improve. Start the extraction now, even if the use case is modest, because the structured dataset you build today becomes a strategic asset later.
Write the Job Description for Your First AI Agent Monitor
- 1
List every AI agent or automation currently running in your company: outbound sequences, content pipelines, data enrichment workflows, customer support bots, internal Slack agents. For each one, write one sentence describing what it does and one sentence describing what 'broken' looks like.
- 2
For each agent, list the three things that go wrong most often: authentication failures, hallucinated outputs, broken integrations, stale data, or incorrect routing. These are the monitoring requirements for the role you are designing.
- 3
Draft a one-page job description for an AI Agent Monitor. Required sections: daily monitoring checklist, error escalation protocol (when to flag vs. when to fix independently), tools used (Slack, your automation platform, any dashboards), and success metrics (uptime percentage, mean time to resolution, weekly incident log).
- 4
Prompt Claude: 'Here is my AI agent stack and the errors each agent produces [paste your list]. Write a daily monitoring SOP for a non-engineer who will review outputs, test key workflows, and log anomalies. Format it as a repeatable checklist.' Use the output as the core of your JD's responsibilities section.
- 5
Post the JD to one offshore staffing platform or share it with Penbrothers. The goal is not necessarily to hire immediately - it is to see whether the role as written would attract qualified candidates, which tells you whether your process documentation is clear enough for someone outside your company to execute.
Stretch goal: Nicolas' insight: the AI agent maintenance role is the fastest-growing offshore category because every founder is deploying agents faster than they can monitor them. The founders who scale cleanly are the ones who treat their agent stack like production infrastructure - with on-call protocols, incident logs, and a human responsible for overnight reliability. Build the monitoring role before you need it, not after your agents have been silently wrong for a week.
Run a 100-Hour Build-and-Launch Sprint
- 1
Define your two success metrics before you start: one for virality (e.g. 500 upvotes, 1,000 shares, press pickup) and one for revenue (e.g. 5 paying customers, $500 in transactions). Write both down. These are binary - you either hit them or you do not.
- 2
Pick the single most stripped-down version of your idea that could prove or disprove the core mechanic. Not an MVP - a proof-of-concept. The question is: does anyone care? Not: is the product complete?
- 3
Set a launch date inside the 100-hour window and commit to it publicly - a Product Hunt pre-launch page, a waitlist with a countdown, or a post announcing the date. External commitment eliminates the temptation to delay for polish.
- 4
Build only what is necessary to deliver the core experience to the first user. Log every hour. When you find yourself building something outside the critical path, write it on a 'post-sprint' list and keep moving.
- 5
On launch day, do not watch metrics passively. Actively post to three communities where your target user lives: a subreddit, a Slack group, a newsletter audience. Measure engagement in the first 24 hours only - the signal is in the initial reaction, not the long tail.
Stretch goal: Medi's insight: the 100-hour constraint is not about speed - it is about honesty. It forces you to test the core mechanic before you have invested enough to rationalize a bad result. After the sprint, answer one question: if you could start over with what you know now, would you build the same thing? If no, that is the most valuable data the sprint produced.
Write Your Company Memo: From Tool to Outcome
- 1
Write down what you currently sell in one sentence: 'I sell [X] and customers pay [price] to [access/use] it.' Now write what your best customers actually buy: the outcome they achieve after using your product. 'After using [X], our best customers [achieve Y outcome in Z timeframe].' If you cannot complete the second sentence cleanly, that gap is your first insight.
- 2
Rewrite your value proposition as an outcome contract. Draft a one-paragraph internal memo: 'We will deliver [specific outcome] for [customer type] in [timeframe] for [price].' The outcome should be something the customer can verify - a placed hire, a filed permit, a shipped design system - not a capability or a feature.
- 3
Identify the three AI workflows you would need to build or meaningfully improve to actually deliver that outcome reliably at scale. For each, rate your current capability: (A) already doing it well, (B) could build it in 30 days, (C) not yet possible with available tools. A-category workflows are your moat. C-category workflows are your roadmap.
- 4
Price the outcome: what would a customer pay for this result if it came with a delivery guarantee? Anchor to what they currently pay for the analog human service - recruiter fee, agency retainer, consulting engagement. That is your ceiling. Calculate your delivery cost with AI leverage. If the margin is positive and the gap is large, you have an outcome-as-a-service opportunity.
- 5
Find one design partner willing to pay for the outcome before you build the full system. Pitch the outcome, not the technology. If three people say yes in principle, proceed. If three people say 'interesting but not now,' the outcome is a vitamin - return to step one and find the painkiller version.
Stretch goal: Brennan's frame: for the first time ever, you can sell outcomes, not tools. The test is whether your AI leverage is real enough that outcome pricing is profitable. Most founders discover their AI leverage is genuine but they have been pricing it like a SaaS tool - leaving the majority of the value on the table. The memo is not a product roadmap; it is a pricing hypothesis. You are asking: if we could guarantee the outcome, what would the market pay, and can we deliver it at a margin worth building around?
Audit Your Agent Credentials and Build a Permission Minimization Policy
- 1
List every agent, script, automation, and integration in your company that has access to production systems or external APIs. Include n8n workflows, Zapier zaps, CI/CD pipelines, cron jobs, AI coding assistants, and any tool that acts on your behalf programmatically. For each, write one sentence describing what it can do.
- 2
For each item on your list, document where its credentials live: (A) hardcoded in source code, (B) in a .env file on a developer machine, (C) in a secrets manager or vault, (D) in the CI/CD platform's secrets store, (E) device-bound or time-limited. Rate each A-E. Most teams discover more A and B than they expect.
- 3
For each A or B credential, calculate the blast radius if it were compromised today: what systems could an attacker access, what data could they read or exfiltrate, what actions could they take? Write one sentence per credential. This exercise makes abstract risk concrete.
- 4
For each agent, document the human authorization behind it: who approved this agent running with these permissions? When was that decision made? Is it documented anywhere? If you cannot answer all three questions, the agent is operating without a traceable authorization chain.
- 5
Write a one-page Permission Minimization Policy for your team: (1) no production credentials in .env files on developer machines - use a secrets manager, (2) all agent credentials are scoped to minimum required permissions, (3) all agent credentials have an expiry date or rotation schedule, (4) every agent has a named human owner responsible for reviewing its permissions quarterly.
Stretch goal: Jasson Casey's framing: 80% of security incidents trace back to credential movement - credentials that can be copied, stolen, or used somewhere other than where they were created. The audit does not require buying new tools. It requires making visible what your team has already built. Most founders who do this exercise discover three to five credentials that could cause a serious incident if compromised, and at least one they had forgotten existed. The audit is the prerequisite for fixing anything.
Run a 30-Minute Equity Concentration Audit for Your Founding Team
- 1
List every equity position your founding team holds: company shares, vested options, any secondary positions. For each, estimate the current value based on the most recent 409A or funding round valuation. Then calculate what percentage of each person's total estimated net worth (including savings, real estate, public market investments) is represented by their primary company equity.
- 2
Apply the 38% rule: of every US Series C company from 2010-2015, only 38% produced value for employee shareholders a decade later. For each founding team member whose company equity exceeds 50% of their net worth, model the outcome if your company is in the 62%. What is their financial position? Can they still afford their life? Do they have enough diversified assets to weather a decade of effort that produces no equity payout?
- 3
Map your liquidity options by category: (A) secondary marketplace sale (taxable, commission applies, immediate cash), (B) single-stock loan (expensive, low loan-to-value ratio), (C) exchange fund contribution (tax-free diversification, no immediate cash but LP interest), (D) nothing yet (company not eligible for any of the above). Document which options are realistically available to each team member today.
- 4
Identify the vesting cliff: for each equity holder, calculate how much is vested today vs. unvested. Liquidity and exchange fund options require vested shares. Build a timeline showing when meaningful vested equity will be available for each team member - this is the earliest practical window for any diversification action.
- 5
Schedule a 60-minute team conversation about equity concentration and financial planning. The goal is not to pressure anyone to sell or exchange shares, but to ensure every founding team member understands the 62% scenario and has considered whether they have enough diversified assets to be financially secure if the company takes a decade to exit or does not exit at all.
Stretch goal: Greg Brogger's frame: no financial advisor will recommend holding the bulk of your net worth in a single illiquid asset, but for founders and employees of private companies, that is exactly the default position. The 38% data makes the stakes concrete: most founders at the Series C stage believe they are in the 38% - and most of them are wrong. The audit is not pessimism about your company. It is the same risk management that any sophisticated investor applies to any concentrated position. The question is not whether to bet on your company. It is whether to bet everything.
Map Your Customer Touchpoint Gaps: Where Does Monitoring Go Dark?
- 1
List every scheduled touchpoint in your current customer journey — onboarding calls, check-ins, automated messages, renewal conversations, support tickets. Map them on a timeline from sign-up to churn or renewal. For each touchpoint, note whether it is initiated by you, the customer, or triggered automatically by a data event.
- 2
Identify every gap between touchpoints. These are the windows where your customer is using your product (or not) and you have no visibility. For each gap, write down: what is the customer likely doing, experiencing, or deciding during this window? What could go wrong that you would not know about until it was too late?
- 3
For each invisible gap, ask: what single data point, if you had it, would change your ability to intervene early? Think in terms of usage signals (frequency, depth, feature adoption), sentiment signals (support contact volume, NPS shifts, response rates), and outcome signals (whether the customer is achieving the goal they bought for). Prioritize gaps where the cost of missed intervention is highest.
- 4
Rank the gaps by risk: which ones, if a problem developed and went undetected, would lead to churn, a bad outcome for your customer, or a reputational hit? These are your remote monitoring product priorities — the places where proactive alerting would have the most leverage.
- 5
For your highest-risk gap, design one lightweight monitoring mechanism: a usage health score, an automated check-in message at day 14, a flag when a customer has not logged in for five days. Build and deploy it. Measure whether early intervention changes the downstream outcome (retention, expansion, satisfaction score) for customers who triggered the alert vs. those who did not.
Stretch goal: Anish Sebastian's insight applied outside healthcare: the biggest risk in any customer relationship is not churn — it is invisible churn, the slow decline that becomes visible only at the renewal conversation. Remote patient monitoring exists because too much could go wrong between a 12-week and 16-week appointment. Your equivalent is the window between your customer's sign-up and their 90-day review. The founders who build monitoring loops into their product — not just customer success touchpoints, but data-driven alerts — catch the problems that manual check-ins miss. The gap map is your product roadmap.
Run a Phone Revenue Audit: Count Your Missed Calls for One Week
- 1
For any service business that takes inbound calls, log every incoming call for one full week. For each call, record: (A) answered immediately, (B) answered after hold, (C) missed entirely. If you have a phone system with missed call logs, pull the data. If not, ask your front desk or ops team to track it manually. The goal is a real number - not an estimate.
- 2
For every missed call, apply your average order or ticket value. Multiply missed calls times average ticket to calculate your weekly phone revenue leak. Annualize it. For most service businesses, this number is larger than expected - Christian Wiens cites 40-50% of restaurant calls going unanswered; the dollar figure per location is typically thousands of dollars per month.
- 3
For answered calls, track how often your team attempted an upsell or cross-sell. Calculate your actual upsell rate as a percentage of answered calls. Compare this to what 100% AI upselling would look like: if AI upsells on every call and 30% of customers accept, how many additional dollars per call does that represent over your current baseline?
- 4
Add the two numbers together: recovered revenue from missed calls plus incremental revenue from higher upsell rate. This is your gross AI opportunity for phone channel alone. Divide by a realistic monthly AI phone agent cost (typically $200-500/month for a tool like Loman) to calculate your payback period.
- 5
If the payback period is under three months, run a 30-day pilot with an AI voice agent on one line or location. Use the same call volume data as your control baseline. After 30 days, compare: average ticket size, answered call percentage, staff time freed from phone handling, and any measurable change in customer satisfaction or complaints.
Stretch goal: Christian Wiens's observation: 40% of restaurant callers don't know they're speaking to AI - and the ones who do know ask more questions and divulge more preferences than they would with a human. The audit is not just about finding the revenue leak. It is about understanding how much of your customer relationship is currently being shaped by whether a distracted employee happened to pick up the phone in a good mood. If the answer is 'a lot,' the upside of a consistently excellent AI agent is larger than the revenue recovery alone.