OpenAI launched their o3 model last week, and Tyler Cowen called it AGI. While the AGI debate is semantic, it's clear we're in the age of AGI. I share some practical observations about using o3, including its verbose nature, technical depth, and how it differs from GPT-4 in terms of engagement and dialogue style...
The cards in this feed now have a full clickable area. (Thanks for the suggestion, Tova!)
I built DistillJS to recreate specific LLM behaviors in Javascript. It works by challenging o3 to write a minified JS function that yields results broadly comparable to gpt-4o-mini's on a specific task. I just launched it (in alpha) so try it out and email feedback...
I built stringme.dev a few weeks ago to generate a plain text version of any webpage in a format optimized for LLMs. Firecrawl recently introduced their own version of basically the same tool: llmstxt.new. While there's been some traction on stringme.dev, llmstxt.new is a pretty solid solution, so stringme.dev will now redirect to llmstxt.new...
Jason Lemkin (SaaStr) and Rory O'Driscoll (Scale) went on 20VC recently and shared their observations about how product-market fit has become more transient than ever. They discuss how SaaS investing has changed, with companies now potentially losing product-market fit in weeks rather than years...
OpenAI is reportedly in talks to buy Windsurf for $3 billion, after previously considering Cursor. Building on Jake Handy's observations about Cursor (and Windsurf's) potential to expand to non-developer work, I'd argue OpenAI is bidding on Windsurf as a play for the future of work...
Here's a bunch of new data on small teams with massive revenue numbers, including tinyteams.xyz, IndieHackers' thread on one-person companies, and data on AI unicorns' revenue per employee...
ChatGPT's projects feature launched in December and offers powerful custom instructions for organizing related chats. I've been using it for meal planning and startup ideation, with project-specific instructions and persistent context across conversations...
I'm reading the book Problem Hunting by Brian Long and exploring how LLMs can make the process easier. Tools like Deep Research and Rally can help validate startup ideas faster and cheaper than traditional methods...
In a recent talk, OpenAI Chief Product Officer Kevin Weil said writing evals is becoming a core skill for PMs. This reflects both the growing importance of models in software and the changing nature of product management work, where implementation is increasingly outsourced to LLMs...
ChatGPT announced a new version of its "memory" feature that will draw from users' entire chat history. This creates significant switching costs, especially in enterprise contexts where chat history contains valuable business data...
Sahil Lavingia, the founder of Gumroad and Antiwork, is testing compelling ideas about the future of work, including open source bounties, cybernetic teammates, royalty-based funding, and flexible work arrangements...
Highly recommend the new book The Thinking Machine by Stephen Witt. I learned a lot about Jensen, Nvidia, the history of GPUs, and machine learning...
Shopify CEO Tobi Lütke recently sent a memo to his employees mandating the use of AI tools. The capitalistic logic behind this mandate is irrefutable and will become the norm, especially if economic conditions worsen in the near-term...
Josh Mohrer was on the Generative Now podcast recently and floated an exit strategy for solo founders. If you successfully bootstrap to profitability, it might be best to sell to a company that wants to go multi-product, which is happening earlier and earlier...
We're doing lead gen for Flow State sponsorships, and Clay is really good. They raised a "pre-emptive Series B expansion" at a $1.25B valuation earlier this year...
Zach Weinberg, who founded Flatiron Health and Curie Bio, floated an interesting idea for OpenAI in a podcast today: adding user-to-user messaging to ChatGPT as a form of non-monetary lock-in...
Our friends over at Flow State put out their quarterly R&D post for founding members. Highlights include surpassing 40,000 subscribers, revenue growth through their sponsor booking page, and plans for a potential indie streaming service...
Recent reports of questionable revenue practices at highly capitalized AI startups raise important questions about what constitutes sustainable growth in the AI era...
On Tuesday, Google announced Gemini 2.5, a reasoning model that is optimized for coding and has a 1 million token context window.
Cursor made it available today, so we decided to give it a shot.
We saw that for another Cursor user, Gemini 2.5 was able to solve a problem that Sonnet had gotten stuck on...
I think Microsoft has the right idea with their security agents. This could help the urgent "vibe code exploits" problem...
Fireship tackled vibe coding in a recent video. There's an urgent need for a turnkey cybersecurity agent in the age of AI-generated code...
Paul Graham has a classic essay about how to get startup ideas.
"The way to get startup ideas is not to try to think of startup ideas. It's to look for problems, preferably problems you have yourself..."
A common pain point working with code generation tools like Cursor is that with every new chat, the agent needs to scan the code base to get context. The agent acts as though it's never seen this codebase before, even if it wrote every line of it.
The best way to mitigate this is to include a product requirements doc (PRD) as a markdown file in the repo...
One of our observations at Subtle is that, enabled by LLMs, "individuals will increasingly outperform teams." A new paper from HBS, Wharton, and Proctor & Gamble tested this idea in practice.
Researchers worked with 776 employees at P&G, randomly assigning employees to work with or without AI and work alone or with a team...
There is a telltale sign of a vibe-coded landing page: header (product name, call to action), h1 greeting, h3 description, another call to action. They all look the same.
Check out some examples of vibe-coded landing pages...
Sam Lessin offered an interesting, skeptical take on Cursor and similar LLM wrappers recently: "I like [Cursor]. I use it as a product. I don't see it as a good business.... There's no lock-in.... I think the whole point of AI is the switching costs for all this stuff goes to zero."
Even though Cursor may be the fastest startup to reach $100M in ARR, it does face some existential risks...
Vibe coding is not without risks and may lead to lost work and security vulnerabilities. Here are some pointers for how to vibe code responsibly.
That said, there's also a new security opportunity in the world of superabundant software...
Today Stripe added the ability to copy the text of their docs or view the page as markdown. This is an example of a company proactively making the contents of their webpages more readily available for LLMs.
However, this is one of over 50 billion webpages. There's a tragedy of the commons problem: most webpages won't make /llms.txt or equivalent available anytime soon. In the meantime, that's what stringme.dev is for...
The main lesson from building stringme.dev was to identify the languages, frameworks, and services you'll need up front when vibe coding.
We ran into problems when we told Cursor mid-project that we wanted to use Tailwind, deploy via Vercel, or switch from TypeScript to JavaScript.
Here's how we vibe-coded stringme.dev and what we learned along the way...
Spent today iterating on stringme.dev.
Added quick URL buttons for example websites, improved copy across the site, and implemented Mozilla's readability library to strip away extraneous filler text from webpages.
Also improved text optimization for LLMs while stubbornly refusing to use an LLM for this process itself.
Check out all the updates and improvements...
Here's something new from us: stringme.dev.
This service takes a URL and generates a text string. Sounds simple. Why does it matter?
I got the idea from an Andrej Karpathy tweet about needing a service that takes a URL and returns clean text content for LLMs.
Check out the full story behind this new tool...
Simon, who created the popular Python framework Django, published a long blog post today about using LLMs in his coding workflow.
He's been using LLMs for about two years now and has some interesting insights:
Once I've completed the initial research I change modes dramatically. For production code my LLM usage is much more authoritarian: I treat it like a digital intern, hired to type code for me based on my detailed instructions.
Learn about his approach to using LLMs for research, coding, and more...
Manus is another impressive Chinese LLM product, following DeepSeek. It's similar to OpenAI's Operator product, in that it's an agent that can be assigned a complex task and then autonomously, iteratively complete it.
Sesame is an extremely impressive voice demo.
Check out more interesting products including seven39 and Mastra...
Stripe just published their annual letter.
The main point I wanted to discuss is their defense of LLM wrappers. But I couldn't help but admire their writing abilities:
For too long the crypto economy was an isolated atoll, with vibrant native customs but few exports to the rest of the world.
Clearly the Collisons write without LLMs. They raw dog the page like Mark Twain...
A little newsletter called flowstate.fm asked us for help with their sponsorship business.
They wanted to make a website where people could book weekly sponsorships directly. They also wanted some help with outbound lead generation.
The result is sponsor.flowstate.fm which we built in just a few hours. Here's how we did it...
You can thank – or blame – Roilan for this page being center aligned now!
Azeem Azhar recently summarized the before/after of commoditized intelligence (Bloomberg):
For most of history, hiring a dozen PhDs meant a massive budget and months of lead time. Today, a few keystrokes in a chatbot summon that brainpower in seconds.
In light of this sudden appearance of free intelligence, Azhar continues, there's an important question:
The question facing individuals and organizations alike is: What will you do when intelligence itself is suddenly ubiquitous and practically free?
The way I think about this is that the traditional company is being unbundled...
When you're working for a company, your compensation is typically a salary and maybe a bonus and/or stock options, with performance reviews determining your success. But in one person teams, the only priority is customer service – compensation is exactly company revenue, determined entirely by customers rather than managers...