AI Companies School Like Fish
A Blue Ocean Turns Red in <18 Months
If we look at the ecosystem of AI-powered products, there’s a clear pattern of how they emerge and roll out to the world:
- Initial POC: Someone throws together a software demo – not a robust product ready for public consumption – proving a capability. Often this comes from an open source developer, academic researcher, or an R&D team at a larger company. The demo catches fire, hits the frontpage of Hacker News, and circulates through social media.
- Open Experimentation: Open source devs and projects start to experiment with the concept, adding support for the feature to their framework or shipping usable software. This is a Cambrian Era, when lots of variants hit Github and get kicked around.
- Fast-Mover Launch: Eventually, a fast moving company brings a product to market. This could be a start up built around the core idea or it could be an existing organization that quickly adds the feature or product to their offering. For the first time, people are paying (or not paying, depending on if the demand is there and the demo works in production).
- Incumbent Clone: Finally, large companies bring the product or feature to their offerings.
For example, let’s look at text-to-SQL.
In 2022, prior to ChatGPT’s launch, Immanuel Trummer published CodexDB, which translated natural language into SQL queries1. As ChatGPT juiced the AI ecosystem, text-to-SQL became an early example of a business application. LangChain and others shipped components for building and enabling text-to-SQL use cases. In short order, all the large data platforms cloned the feature, including Tableau, Snowflake, and Databricks.
Usually, this cycle happens relatively fast, in less than 18 months.
How many times has it happened? I count at least 8:
- Text-to-SQL: See above.
- Customer Service Bots: Chatbot interfaces to FAQs
- Document Q&A: Turnkey RAG applications with chatbot interfaces
- Note Taking & Summarization: Meeting transcription with extracted summaries and follow-ups.
- Search: Perplexity-style search that uses LLMs to package information gathered from web queries.
- Code Text Completion: Auto-suggest in IDEs as you edit code, powered by AI.
- Coding Agents: Tools like Cursor and Claude Code that perform whole coding tasks for you.
- Deep Research: Like search, but with longer depth, wider breadth, and more depth in pursuit of assembling a report.
- Browser Control: Browsers driven by AI to accomplish tasks the user provides.
I’m sure there’s some product archetypes we’ve missed.
What can we learn from this pattern and the way we’ve been steadily encountering new archetypes, then walking them through the process above?
- Ideas come from hackers, not customers. Few people know how to conceptualize products and cobble together unique applications with AI. This skill comes only through experience and play, and for these first few years most ideas come from the open source community in the form of demos. Not from designers, product managers, or feedback from customers. Applied AI ideas are hard, but execution is cheap. Which is a nice set up to our next take-away…
- Cloning happens faster when the model is the magic. Cloning happened in previous eras, but nowhere near as fast. Start ups would create markets, prove their worth, and only then would larger companies invest in their own teams, projects, and (often) acquisition. Today, when so much of the lift comes from the model itself, there’s little reason to wait (especially when there are few other low-hanging ideas).
- Applied AI start-ups need a niche. When big companies can enter the market in a matter of months, it’s more imperative than ever that start-ups focus on a niche. Google or OpenAI can clone your product, but they’re not nimble enough to invest in your outreach with a specific community and tailoring their product for a segment doesn’t make business sense. Most of the general-purpose RAG start-ups from 2023 have pivoted or failed, but those that focused on one sector (legal, medical, insurance, financing, etc.) are thriving.
- If you’re not niche, you better build a beachhead in <12 months. If you insist on shipping a general purpose applied AI product, and think being early to market is an advantage…well…think again. Cloning moves so fast, you better have an incredble gameplan to pull off significant market acquisition in a handful of months – which will then fuel you through user feedback, training data, and more. But unless you launch with both an incredible marketing advantage and a killer product, you’ll face incredibly tough competition once the big players enter.
The idea that fast-following occurs faster than ever, thanks to everyone having access to the same models, is related to the, “Will the model eat your stack?” problem we discussed earlier.
Considering both the rapid cloning problem and the speed of model advancements, I think every non-niche, applied AI start up needs to ask themselves two questions:
- If a better model arrives tomorrow, does your product get better or does your backend get simpler? If your product doesn’t get better, you need to rethink. A better model simplifying your backend (by reducing the complexity of your prompts, your error handling, your infra, etc.) makes your product easier to clone.
- If you are early to market with this use case, what are you going to do in a handful of months that will fend off Google/OpenAI/whomever’s entry into your market? Cursor and Perplexity are the rare examples that have managed to grow fast enough to be able to fend off larger entrants. What are you going to do, if you can’t go niche, to prepare your defenses?
-
CodexDB used OpenAI’s Codex model, published in 2021. This is not their coding tool named Codex. ↩