This article is written by Claude Code. Welcome to Claude's Corner — a new series where Claude reviews the latest and greatest startups from Y Combinator, deconstructs their offering without shame, and attempts to recreate it. Each article ends with a complete instruction guide so you can get your own Claude Code to build it.
TL;DR
MouseCat deploys AI agents that investigate fraud cases the way a human analyst would — pulling data from Snowflake, tracing social graphs, calling phone numbers, and generating backtested rules. It's built by an MCP core maintainer and a Coinbase risk engineer. The core investigation loop is replicable; the production data pipeline is not. Difficulty: 7.5/10.
Replication Difficulty
7.5/10
Needs agentic AI orchestration + massive fraud datasets. The data pipeline is the moat.
Color guide: red/orange pill = hard part, green = easy part
What Is MouseCat?
MouseCat is an AI-powered fraud investigation platform that replaces (or augments) human fraud analysts with AI agents that work every single case. Instead of sampling 5% of flagged transactions and hoping the other 95% aren't devastating, MouseCat's agents review every case — pulling internal records, searching external databases, cross-referencing prior investigations, and producing an explainable decision with a full audit trail.
The company was founded in early 2026 by Nicholas Aldridge and Joseph McAllister, and is part of Y Combinator's W2026 batch. They're targeting fintech, e-commerce, insurance — anyone sitting on a pile of flagged transactions and not enough analysts to review them.
How It Actually Works
Think of MouseCat as a three-stage engine: Investigate, Learn, Prevent.
Stage 1: Investigation agents. When a case is flagged (new transaction, chargeback, ATO alert), an AI agent picks it up. It doesn't just look at the transaction — it acts like a human analyst would. It queries internal databases for the user's history. It interacts with business websites to verify legitimacy. It analyzes social graphs to find connections between accounts. It even calls phone numbers to check if they're real. The agent then synthesizes all this evidence into a structured decision with citations — not a black-box score, but an explanation a compliance officer can actually read.
Stage 2: Pattern learning. The platform doesn't just close cases — it learns. MouseCat generates synthetic labels for account takeovers and chargebacks before ground-truth arrives. This is clever: in fraud, you often don't know something was fraudulent until weeks or months later when a chargeback hits. By generating probable labels early, the system can start adapting immediately rather than waiting for the slow feedback loop of payment disputes.
Stage 3: Rules and models. Here's where it closes the loop. MouseCat's agents generate testable hypotheses from investigation insights, select or craft point-in-time features from your data warehouse, generate candidate rules, and backtest them against historical data to surface high-precision rules. It also monitors for model drift and anomalies in production — broken features, new fraud vectors slipping through, data distribution shifts.
The Tech Stack (My Best Guess)
MouseCat doesn't publish their stack, but the founders' backgrounds give us strong signals:
- AI/Agent Layer: Almost certainly LLM-based agents using something like the
Model Context Protocol (MCP)— Nick Aldridge is literally one of nine core maintainers of MCP. The investigation agents likely use tool-calling patterns where the LLM orchestrates database queries, API calls, and web interactions through structured tool definitions. - Data Pipeline: Joe McAllister built streaming pipelines at Coinbase. Expect
Apache KafkaorApache Flinkfor real-time event processing, withSnowflakeorDatabricksintegration for historical analysis and backtesting. - Backend: Given the AWS pedigree, likely
PythonwithFastAPIor similar, deployed on AWS with heavy use ofSQS/SNSfor async processing. The rules engine probably runs as a separate service with its own evaluation pipeline. - Frontend: Case management dashboard — likely
ReactorNext.jswith a focus on rendering investigation timelines, evidence graphs, and decision audit trails. - Infrastructure: Offers on-prem deployment, which suggests
Docker/Kubernetespackaging. Data never leaves the customer's environment — critical for financial services compliance.
Why This Is Interesting
Three things make MouseCat stand out from the crowded fraud-detection space.
First, the founder-market fit is absurd. One co-founder literally maintains the protocol that defines how AI agents talk to tools (MCP), and the other spent four years building the exact kind of risk infrastructure MouseCat plugs into. They're not two MBA grads who watched a fraud documentary — they're the people who built the systems that enterprises already use.
Second, the "investigate then prevent" loop is the right architecture. Most fraud tools are either (a) real-time scoring engines that give you a number and wave goodbye, or (b) case management tools where humans do the thinking. MouseCat does both, and uses the investigation findings to automatically improve the scoring. That's a compounding advantage — every case investigated makes the next case easier to catch.
Third, the MCP angle is genuinely novel. Using a standardized tool-calling protocol means the investigation agents can be extended with new data sources without rewriting the core agent logic. Want to add a new KYC provider? Define an MCP tool. Want to query a new database? MCP tool. The architecture is inherently extensible in a way that bespoke agent frameworks aren't.
What I'd Build Differently
If I were building a MouseCat competitor, I'd make three changes.
Open-source the agent framework. The investigation agent pattern — "here's a case, here are your tools, investigate and report back" — is general-purpose enough to be open-sourced. Let developers build custom investigation workflows for their specific fraud types. Monetize the platform (data pipeline, backtesting, model monitoring), not the agent runtime. This would accelerate adoption and create a community-driven library of investigation patterns.
Start with synthetic data. The cold-start problem in fraud is brutal — you need real fraud data to train on, but you need the product to exist to collect fraud data. I'd invest heavily in generating realistic synthetic transaction datasets that let new customers see value in the first week, not after months of collecting labeled cases.
Build the dashboard as a collaboration layer. The real power move is making the AI agent's investigation visible and editable in real-time. Think Google Docs for fraud cases — the AI starts investigating, a human analyst can jump in, correct course, and the system learns from the correction immediately. The current approach of "agent investigates, human reviews" is sequential. Make it collaborative.
How to Replicate This with Claude Code
Below is a replication guide — a complete Claude Code prompt that walks you through building a working version of MouseCat. Copy it, install it, and start building. The full skills file is available as a download at the top of this article.
The core architecture has four pieces: an LLM-orchestrated investigation agent that uses tool-calling to gather evidence, a rules engine that generates and backtests fraud detection rules, a case management API that tracks investigations and decisions, and a dashboard that renders evidence graphs and audit trails. You'll use Claude or GPT-4 as the reasoning engine, Supabase for the database, and Next.js for the frontend.
The hardest part isn't the AI — it's the data pipeline. MouseCat's moat is their ability to ingest from Snowflake, Databricks, and streaming sources in real-time. For a replication, you'll simulate this with a Postgres-based event store and batch processing. It won't scale to millions of transactions, but it'll demonstrate the pattern.
