<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Peter Strauss</title>
    <description>The latest articles on Forem by Peter Strauss (@strauss).</description>
    <link>https://forem.com/strauss</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F120555%2Fe44b56e3-9adf-4174-b83a-313fca3e4ec8.jpg</url>
      <title>Forem: Peter Strauss</title>
      <link>https://forem.com/strauss</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/strauss"/>
    <language>en</language>
    <item>
      <title>Open Source Stars Are Not Revenue, You Need a Commercial Bridge</title>
      <dc:creator>Peter Strauss</dc:creator>
      <pubDate>Fri, 17 Apr 2026 11:40:08 +0000</pubDate>
      <link>https://forem.com/strauss/open-source-stars-are-not-revenue-you-need-a-commercial-bridge-d3e</link>
      <guid>https://forem.com/strauss/open-source-stars-are-not-revenue-you-need-a-commercial-bridge-d3e</guid>
      <description>&lt;p&gt;Open source companies are rich in attention and poor in monetization. &lt;/p&gt;

&lt;p&gt;The repo grows. Stars go up. Contributors show up. The community gets louder. &lt;/p&gt;

&lt;p&gt;Everyone feels like momentum is real, and in one sense, it is. But if that attention never crosses the bridge into paid value, the business can stay strangely fragile for a long time.&lt;/p&gt;

&lt;p&gt;I think this is one of the most common GTM blind spots in open source. Founders assume growth in the community automatically means growth in the company. Sometimes it does. A lot of the time, it just means the market likes the project more than it understands the commercial reason to pay for it.&lt;/p&gt;

&lt;p&gt;That is the trap.&lt;/p&gt;

&lt;h2&gt;
  
  
  The market says community matters — but not in the way people think
&lt;/h2&gt;

&lt;p&gt;The Linux Foundation’s &lt;a href="https://www.linuxfoundation.org/research/2025-state-of-commercial-open-source" rel="noopener noreferrer"&gt;2025 State of Commercial Open Source report&lt;/a&gt; is one of the clearest recent signals here. It shows that commercial open source companies consistently outperform closed-source peers in valuations, funding speed, and liquidity outcomes, especially in infrastructure software. It also says something even more important: &lt;strong&gt;strong community health is closely linked to higher company valuations&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;That sounds bullish for open source, and it is.&lt;/p&gt;

&lt;p&gt;But it is very easy to hear the wrong lesson.&lt;/p&gt;

&lt;p&gt;The right lesson is not:&lt;br&gt;
“community equals revenue.”&lt;/p&gt;

&lt;p&gt;The right lesson is:&lt;br&gt;
“community creates leverage — if the company knows how to convert community value into commercial value.”&lt;/p&gt;

&lt;p&gt;That is a much harder and more useful truth.&lt;/p&gt;

&lt;p&gt;Ubuntu’s &lt;a href="https://ubuntu.com/engage/world-of-open-source-global-2025" rel="noopener noreferrer"&gt;2025 global open source report&lt;/a&gt; helps sharpen this. It says &lt;strong&gt;83% of organizations see open source as valuable to their future&lt;/strong&gt;, and &lt;strong&gt;46% reported increased business value from open source year over year&lt;/strong&gt;. At the same time, the same research warns that adoption is outpacing maturity, especially in governance and operational discipline.&lt;/p&gt;

&lt;p&gt;That is the pattern I keep seeing.&lt;/p&gt;




&lt;p&gt;What are the top 1% of GTM operators doing differently? We deconstruct their playbooks every week. Don’t miss the next one—subscribe at &lt;a href="https://www.gtm.news/" rel="noopener noreferrer"&gt;GTM.NEWS&lt;/a&gt;.&lt;/p&gt;




&lt;p&gt;Open source gets distribution first.&lt;br&gt;
The business model comes later.&lt;br&gt;
And if “later” lasts too long, the company gets trapped in a weird middle ground: popular enough to attract users, but too vague commercially to convert them cleanly.&lt;/p&gt;

&lt;h2&gt;
  
  
  The harsh truth
&lt;/h2&gt;

&lt;p&gt;Usage is not monetization.&lt;/p&gt;

&lt;p&gt;Community is not monetization.&lt;/p&gt;

&lt;p&gt;Stars are definitely not monetization.&lt;/p&gt;

&lt;p&gt;Those things can support a business. They do not automatically become one.&lt;/p&gt;

&lt;p&gt;A lot of founders avoid this because the transition feels uncomfortable. The community likes the project partly because it feels open, accessible, and developer-first. The moment the company talks more directly about money, support, enterprise packaging, or commercial boundaries, it worries about backlash.&lt;/p&gt;

&lt;p&gt;That fear is real.&lt;/p&gt;

&lt;p&gt;But the worse risk is building a beloved project with no clear economic engine.&lt;/p&gt;

&lt;p&gt;Because once that happens, the company starts making bad GTM decisions:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;vague enterprise pages&lt;/li&gt;
&lt;li&gt;weak upgrade paths&lt;/li&gt;
&lt;li&gt;unclear feature gating&lt;/li&gt;
&lt;li&gt;random sales outreach&lt;/li&gt;
&lt;li&gt;awkward “contact us” monetization&lt;/li&gt;
&lt;li&gt;community growth treated like a revenue strategy&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is not a bridge.&lt;br&gt;
That is wishful thinking.&lt;/p&gt;

&lt;h2&gt;
  
  
  My rule: identify the moment free use becomes expensive
&lt;/h2&gt;

&lt;p&gt;This is the cleanest commercial frame I know for open source GTM.&lt;/p&gt;

&lt;p&gt;Do not start with:&lt;br&gt;
“What can we charge for?”&lt;/p&gt;

&lt;p&gt;Start with:&lt;br&gt;
&lt;strong&gt;When does free usage stop being enough for a serious team?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;That is where the commercial bridge lives.&lt;/p&gt;

&lt;p&gt;Usually the trigger shows up in one of five places:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;governance&lt;/li&gt;
&lt;li&gt;reliability&lt;/li&gt;
&lt;li&gt;support&lt;/li&gt;
&lt;li&gt;scale&lt;/li&gt;
&lt;li&gt;security or compliance&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The open source product wins the developer.&lt;br&gt;
The commercial product wins the organization.&lt;/p&gt;

&lt;p&gt;That is the distinction.&lt;/p&gt;

&lt;h2&gt;
  
  
  The practical fix: build a three-step commercial bridge
&lt;/h2&gt;

&lt;p&gt;If I were helping an open source company clean up its GTM, this is the system I would build.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Define the free-to-paid trigger clearly
&lt;/h3&gt;

&lt;p&gt;This is not a pricing exercise first.&lt;br&gt;
It is an operating transition.&lt;/p&gt;

&lt;p&gt;Ask:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;when does the tool become business-critical?&lt;/li&gt;
&lt;li&gt;when does the team need controls?&lt;/li&gt;
&lt;li&gt;when does uptime or support start mattering?&lt;/li&gt;
&lt;li&gt;when does security review become real?&lt;/li&gt;
&lt;li&gt;when does “we can manage this ourselves” stop being true?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you cannot point to that transition cleanly, the sales motion will stay fuzzy.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Package the paid layer around organizational pain
&lt;/h3&gt;

&lt;p&gt;This is where a lot of companies get too clever and not useful enough.&lt;/p&gt;

&lt;p&gt;Do not sell vague “enterprise value.”&lt;/p&gt;

&lt;p&gt;Sell the things that remove pain for the organization:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;SSO / access control&lt;/li&gt;
&lt;li&gt;auditability&lt;/li&gt;
&lt;li&gt;support SLAs&lt;/li&gt;
&lt;li&gt;compliance posture&lt;/li&gt;
&lt;li&gt;policy enforcement&lt;/li&gt;
&lt;li&gt;deployment flexibility&lt;/li&gt;
&lt;li&gt;admin and governance tooling&lt;/li&gt;
&lt;li&gt;faster implementation or migration help&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The developer might love the open source project.&lt;br&gt;
The buyer pays when the company makes organizational pain smaller.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Route community signals into product-qualified accounts
&lt;/h3&gt;

&lt;p&gt;Community should not feed a vanity dashboard.&lt;br&gt;
It should feed the commercial motion.&lt;/p&gt;

&lt;p&gt;I would track:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;repeated usage from the same company domains&lt;/li&gt;
&lt;li&gt;contributors from target accounts&lt;/li&gt;
&lt;li&gt;GitHub actions or repo activity that suggest production use&lt;/li&gt;
&lt;li&gt;docs traffic from likely enterprise environments&lt;/li&gt;
&lt;li&gt;community questions that signal rollout, governance, or scale pain&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is how open source attention becomes GTM intelligence.&lt;/p&gt;

&lt;h2&gt;
  
  
  A worked example
&lt;/h2&gt;

&lt;p&gt;Imagine an open source observability tool.&lt;/p&gt;

&lt;p&gt;The weak monetization path says:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;free repo&lt;/li&gt;
&lt;li&gt;lots of stars&lt;/li&gt;
&lt;li&gt;community Slack&lt;/li&gt;
&lt;li&gt;vague enterprise page&lt;/li&gt;
&lt;li&gt;“contact sales” buried in the nav&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The stronger path says:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;free core remains genuinely useful&lt;/li&gt;
&lt;li&gt;clear page explaining when teams outgrow self-managed use&lt;/li&gt;
&lt;li&gt;enterprise product built around access control, support, auditability, and scale&lt;/li&gt;
&lt;li&gt;proof showing what changes when the tool becomes critical infrastructure&lt;/li&gt;
&lt;li&gt;community and product signals routed into outbound or lifecycle plays&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Now the company is not trying to “monetize the community” in some vague sense.&lt;/p&gt;

&lt;p&gt;It is helping serious users recognize when the commercial layer makes sense.&lt;/p&gt;

&lt;p&gt;That is much cleaner.&lt;/p&gt;

&lt;h2&gt;
  
  
  What to measure
&lt;/h2&gt;

&lt;p&gt;I would track:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;percentage of community or OSS users converting to team or enterprise use&lt;/li&gt;
&lt;li&gt;time from first repo interaction to commercial inquiry&lt;/li&gt;
&lt;li&gt;top free-to-paid trigger by segment&lt;/li&gt;
&lt;li&gt;domain concentration in product usage and community activity&lt;/li&gt;
&lt;li&gt;expansion rate among accounts that started in OSS&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Those numbers tell you whether the bridge exists or whether the company is still standing on one side hoping money appears on the other.&lt;/p&gt;

&lt;h2&gt;
  
  
  My practical take
&lt;/h2&gt;

&lt;p&gt;One of the more useful truths in open source GTM is that community is not the finish line.&lt;/p&gt;

&lt;p&gt;It is the beginning of leverage.&lt;/p&gt;

&lt;p&gt;The business wins when it can preserve community energy while giving organizations a very clear reason to pay when the stakes get higher.&lt;/p&gt;

&lt;p&gt;That means:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;know when free stops being enough&lt;/li&gt;
&lt;li&gt;package around organizational pain&lt;/li&gt;
&lt;li&gt;route OSS signals into GTM&lt;/li&gt;
&lt;li&gt;and stop treating stars like a business model&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Because open source can absolutely create great companies.&lt;/p&gt;

&lt;p&gt;But only when the bridge from project love to commercial value is built on purpose.&lt;/p&gt;

</description>
      <category>opensource</category>
      <category>gtm</category>
      <category>gtmstrategy</category>
    </item>
    <item>
      <title>Time-to-First-Success Is Your Real Acquisition Funnel</title>
      <dc:creator>Peter Strauss</dc:creator>
      <pubDate>Thu, 16 Apr 2026 09:04:38 +0000</pubDate>
      <link>https://forem.com/strauss/time-to-first-success-is-your-real-acquisition-funnel-3g0c</link>
      <guid>https://forem.com/strauss/time-to-first-success-is-your-real-acquisition-funnel-3g0c</guid>
      <description>&lt;p&gt;A lot of devtool companies think they have an acquisition problem when they really have an activation problem. They fight for the click, celebrate the signup, and then quietly lose the user in the first 15 minutes. &lt;/p&gt;

&lt;p&gt;That is the trap. &lt;/p&gt;

&lt;p&gt;The market calls it “top of funnel,” but the builder’s version is simpler: if the developer does not reach a real first win fast enough, you did not really acquire them.&lt;/p&gt;

&lt;p&gt;I think this is one of the biggest GTM mistakes in developer businesses because it hides inside decent-looking growth metrics. Traffic can be up. Signups can look healthy. Docs can get views. Community can be active. And still the business can feel weirdly sticky and slower to grow than it should, because the product is asking the user to do too much work before they feel any payoff.&lt;/p&gt;

&lt;p&gt;That is not just a product issue.&lt;/p&gt;

&lt;p&gt;That is a &lt;a href="https://www.gtm.news/" rel="noopener noreferrer"&gt;GTM&lt;/a&gt; issue.&lt;/p&gt;

&lt;h2&gt;
  
  
  The data already points in one direction
&lt;/h2&gt;

&lt;p&gt;The strongest recent developer-adoption research says the same thing very plainly: adoption is not mainly a content problem. It is a product-experience problem.&lt;/p&gt;

&lt;p&gt;In &lt;a href="https://instruqt.com/state-of-developer-adoption" rel="noopener noreferrer"&gt;Instruqt’s State of Developer Adoption 2025&lt;/a&gt;, developer GTM teams still rely heavily on written documentation, but hands-on, real-world training is ranked as the most effective way to drive adoption at &lt;strong&gt;42.6%&lt;/strong&gt;, ahead of step-by-step documentation at &lt;strong&gt;39%&lt;/strong&gt;. Yet fewer than one-third of organizations are actively investing in interactive labs today. The same research says &lt;strong&gt;57.3%&lt;/strong&gt; of organizations track product usage as their primary adoption metric, and nearly &lt;strong&gt;60%&lt;/strong&gt; report that it takes &lt;strong&gt;one to three months&lt;/strong&gt; for developers to fully adopt new software.&lt;/p&gt;

&lt;p&gt;That is a huge clue.&lt;/p&gt;

&lt;p&gt;If the market says hands-on experience is what actually drives adoption, but most teams still lean hardest on static docs and then wait one to three months for usage to stabilize, the gap is not in awareness. The gap is in how quickly the product helps the user succeed.&lt;/p&gt;

&lt;p&gt;Atlassian’s 2025 DevEx work says the same thing from a different angle. In &lt;a href="https://www.atlassian.com/blog/developer/developer-experience-report-2025" rel="noopener noreferrer"&gt;its 2025 State of Developer Experience report&lt;/a&gt;, almost all developers say AI is saving them time, but &lt;strong&gt;50% still lose 10+ hours a week&lt;/strong&gt; to non-coding work and &lt;strong&gt;90% lose 6+ hours or more&lt;/strong&gt;. Their biggest friction points are &lt;strong&gt;finding information, adapting new technology, and context switching between tools&lt;/strong&gt;. Atlassian also says developers spend only &lt;strong&gt;16% of their time coding&lt;/strong&gt;, which is a really useful reminder that the thing slowing adoption is not usually the code itself. It is the friction around the code.&lt;/p&gt;

&lt;p&gt;That matters a lot for devtools founders.&lt;/p&gt;

&lt;p&gt;Because if your product asks the user to search, interpret, adapt, switch, configure, and guess before they get a real success moment, you are not just creating friction. You are making the acquisition more expensive than your dashboard admits.&lt;/p&gt;

&lt;h2&gt;
  
  
  The harsh truth
&lt;/h2&gt;

&lt;p&gt;A signup is not proof of demand.&lt;/p&gt;

&lt;p&gt;A signup is proof of curiosity.&lt;/p&gt;

&lt;p&gt;That sounds obvious.&lt;br&gt;
A lot of companies still operate like it is the same thing.&lt;/p&gt;

&lt;p&gt;Curiosity signs up because:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;the idea sounds promising&lt;/li&gt;
&lt;li&gt;the docs look interesting&lt;/li&gt;
&lt;li&gt;the product got shared on X or GitHub&lt;/li&gt;
&lt;li&gt;the buyer wants to compare options&lt;/li&gt;
&lt;li&gt;the developer wants to test whether this could save time later&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Demand shows up when the developer gets a working result and thinks:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;“Okay, this is actually useful.”&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;That is a very different moment.&lt;/p&gt;

&lt;p&gt;And it usually happens much later than the marketing team wants to believe.&lt;/p&gt;

&lt;h2&gt;
  
  
  My rule: acquisition ends at first success, not first signup
&lt;/h2&gt;

&lt;p&gt;This is the cleanest operating rule I know for developer GTM.&lt;/p&gt;

&lt;p&gt;Do not ask:&lt;br&gt;
&lt;strong&gt;How many signups did we get?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Ask:&lt;br&gt;
&lt;strong&gt;How many users reached a meaningful first success quickly enough to want the second step?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;That immediately changes what you build, what you measure, and where you spend time.&lt;/p&gt;

&lt;p&gt;Because once first success becomes the metric, all the usual debates start looking different.&lt;/p&gt;

&lt;p&gt;You stop obsessing over:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;homepage tweaks&lt;/li&gt;
&lt;li&gt;shallow lead counts&lt;/li&gt;
&lt;li&gt;vanity community growth&lt;/li&gt;
&lt;li&gt;generic doc traffic&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And you start obsessing over:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;how long setup takes&lt;/li&gt;
&lt;li&gt;where users get stuck&lt;/li&gt;
&lt;li&gt;what information they search for first&lt;/li&gt;
&lt;li&gt;how many steps it takes to get one real working outcome&lt;/li&gt;
&lt;li&gt;how quickly the user can see the product in their own reality&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is a much better GTM lens for developer businesses.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why first success matters so much
&lt;/h2&gt;

&lt;p&gt;There are three reasons this lever is stronger than it looks.&lt;/p&gt;

&lt;h3&gt;
  
  
  1. It collapses the gap between product and GTM
&lt;/h3&gt;

&lt;p&gt;For most B2B categories, marketing and product can still pretend to be separate functions for a while.&lt;/p&gt;

&lt;p&gt;Developer products do not have that luxury.&lt;/p&gt;

&lt;p&gt;If the developer cannot understand, test, and validate the value quickly, the business does not just have a product problem. It has a conversion problem, a trust problem, and a retention problem all at the same time.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. It reduces how much selling has to happen later
&lt;/h3&gt;

&lt;p&gt;When users reach a real first success quickly, a lot of downstream GTM gets easier:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;docs feel more useful&lt;/li&gt;
&lt;li&gt;support load drops&lt;/li&gt;
&lt;li&gt;community explanations get cleaner&lt;/li&gt;
&lt;li&gt;team invites happen more naturally&lt;/li&gt;
&lt;li&gt;enterprise conversations start from evidence, not theory&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is a big deal.&lt;/p&gt;

&lt;p&gt;The easier the first win is to reach, the less human effort you need to “sell” the product later.&lt;/p&gt;




&lt;p&gt;Enjoyed this deep dive? Get more actionable Go-To-Market insights delivered straight to your inbox. Join the community and sign up at &lt;a href="https://www.gtm.news/" rel="noopener noreferrer"&gt;GTM.NEWS&lt;/a&gt; today.&lt;/p&gt;




&lt;h3&gt;
  
  
  3. It creates a cleaner signal for who is truly activated
&lt;/h3&gt;

&lt;p&gt;Instruqt’s data is useful here again. If most teams are using product usage as the primary adoption metric, then the smarter question is not “did they use the product?” It is “did they reach the first behavior that predicts meaningful usage?” That is a much sharper activation signal than simply counting accounts created.&lt;/p&gt;

&lt;h2&gt;
  
  
  The practical fix: build a first-15-minutes activation path
&lt;/h2&gt;

&lt;p&gt;If I were fixing this for a developer product this week, I would not start by writing more docs.&lt;/p&gt;

&lt;p&gt;I would start by designing the shortest path to one undeniable win.&lt;/p&gt;

&lt;p&gt;Here is the framework I would use.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Define the first success moment
&lt;/h3&gt;

&lt;p&gt;This is the most important question in the whole article.&lt;/p&gt;

&lt;p&gt;What is the first moment where the user can honestly say:&lt;br&gt;
&lt;strong&gt;“It works.”&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Not “I understand the product.”&lt;br&gt;
Not “the dashboard loaded.”&lt;br&gt;
Not “the environment is configured.”&lt;/p&gt;

&lt;p&gt;A real win.&lt;/p&gt;

&lt;p&gt;Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;first successful API call&lt;/li&gt;
&lt;li&gt;first live deployment&lt;/li&gt;
&lt;li&gt;first test generated and passing&lt;/li&gt;
&lt;li&gt;first working integration with an existing tool&lt;/li&gt;
&lt;li&gt;first useful alert or automation firing in production or staging&lt;/li&gt;
&lt;li&gt;first code issue caught or fixed in a realistic workflow&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That moment needs to be concrete, visible, and meaningful.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Strip the path down to the minimum useful steps
&lt;/h3&gt;

&lt;p&gt;Once you know the first win, remove everything that is not necessary to get there.&lt;/p&gt;

&lt;p&gt;I would map the current path like this:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;signup&lt;/li&gt;
&lt;li&gt;verify email&lt;/li&gt;
&lt;li&gt;create workspace&lt;/li&gt;
&lt;li&gt;choose use case&lt;/li&gt;
&lt;li&gt;install package&lt;/li&gt;
&lt;li&gt;connect repo&lt;/li&gt;
&lt;li&gt;configure permissions&lt;/li&gt;
&lt;li&gt;read docs&lt;/li&gt;
&lt;li&gt;write code&lt;/li&gt;
&lt;li&gt;test output&lt;/li&gt;
&lt;li&gt;debug setup&lt;/li&gt;
&lt;li&gt;finally see result&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;That is already too much for many products.&lt;/p&gt;

&lt;p&gt;The better question is:&lt;br&gt;
&lt;strong&gt;What can we pre-configure, automate, sandbox, template, or defer until after the first win?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This is where a lot of adoption gets rescued.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Create one golden path per ICP
&lt;/h3&gt;

&lt;p&gt;Do not make one generic onboarding path for “developers.”&lt;/p&gt;

&lt;p&gt;That is lazy and usually weak.&lt;/p&gt;

&lt;p&gt;Create one shortest-path experience for each core use case or buyer type.&lt;/p&gt;

&lt;p&gt;For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;solo developer evaluating in a sandbox&lt;/li&gt;
&lt;li&gt;startup engineer integrating with a real repo&lt;/li&gt;
&lt;li&gt;enterprise evaluator testing security and workflow fit&lt;/li&gt;
&lt;li&gt;DevOps lead validating rollout across environments&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Different users need different first wins.&lt;br&gt;
Treating them the same slows everyone down.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Build one fallback path for failure
&lt;/h3&gt;

&lt;p&gt;This is one of the little tricks more experienced operators use.&lt;/p&gt;

&lt;p&gt;Most onboarding paths are designed for success.&lt;br&gt;
A lot of developer trust is actually won in failure.&lt;/p&gt;

&lt;p&gt;When setup breaks, the user should not have to improvise the next move.&lt;/p&gt;

&lt;p&gt;Give them:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;one fast troubleshooting page&lt;/li&gt;
&lt;li&gt;one known-good sample project&lt;/li&gt;
&lt;li&gt;one way to test in a sandbox&lt;/li&gt;
&lt;li&gt;one clear “here is where people usually get stuck” guide&lt;/li&gt;
&lt;li&gt;one fast route to support or community help&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That makes the product feel more mature immediately.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Measure time-to-first-success directly
&lt;/h3&gt;

&lt;p&gt;I would track:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;median time from signup to first success&lt;/li&gt;
&lt;li&gt;percentage of users reaching first success in 15 minutes, 1 hour, and 24 hours&lt;/li&gt;
&lt;li&gt;top drop-off steps before first success&lt;/li&gt;
&lt;li&gt;most common setup failures&lt;/li&gt;
&lt;li&gt;second-step behavior after first success&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That last one matters a lot.&lt;/p&gt;

&lt;p&gt;Because a first success that does not lead to deeper usage might still be too shallow. You want the first success to create momentum, not just a temporary smile.&lt;/p&gt;

&lt;h2&gt;
  
  
  A worked example
&lt;/h2&gt;

&lt;p&gt;Let’s say you run a developer observability product.&lt;/p&gt;

&lt;p&gt;The weak GTM story says:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;drive traffic to docs&lt;/li&gt;
&lt;li&gt;offer a free trial&lt;/li&gt;
&lt;li&gt;let users connect their stack&lt;/li&gt;
&lt;li&gt;hope they reach value after instrumentation&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That sounds normal.&lt;br&gt;
It is also a little cruel.&lt;/p&gt;

&lt;p&gt;The better version says:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;pick one high-pain use case, like “find the root cause of a slow endpoint”&lt;/li&gt;
&lt;li&gt;give the user a sample environment or staged sandbox&lt;/li&gt;
&lt;li&gt;show one issue being found and explained in under 15 minutes&lt;/li&gt;
&lt;li&gt;then guide them into connecting their real environment&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Now the first win comes before the heavy lift.&lt;/p&gt;

&lt;p&gt;That changes the emotional arc completely.&lt;/p&gt;

&lt;p&gt;Instead of:&lt;br&gt;
“this looks complicated”&lt;/p&gt;

&lt;p&gt;the user thinks:&lt;br&gt;
“okay, this helps — now I’m willing to do the setup.”&lt;/p&gt;

&lt;p&gt;That is a much stronger growth motion.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where AI helps
&lt;/h2&gt;

&lt;p&gt;This is one of the best areas to use AI productively.&lt;/p&gt;

&lt;p&gt;Use AI to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;identify where users get stuck in docs and onboarding&lt;/li&gt;
&lt;li&gt;cluster failed setup flows&lt;/li&gt;
&lt;li&gt;personalize the first-path instructions by stack&lt;/li&gt;
&lt;li&gt;generate better in-product troubleshooting guidance&lt;/li&gt;
&lt;li&gt;summarize the fastest route to success based on user context&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;But the key is the same as everywhere else:&lt;br&gt;
AI should reduce friction, not add another layer of vague possibility.&lt;/p&gt;

&lt;p&gt;A bloated AI assistant inside onboarding does not save you if the path to first value is still too long.&lt;/p&gt;

&lt;h2&gt;
  
  
  What I would do this quarter
&lt;/h2&gt;

&lt;p&gt;I would run a 30-day activation audit.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Watch 10 real users try to reach first success.&lt;/li&gt;
&lt;li&gt;Time every step.&lt;/li&gt;
&lt;li&gt;Mark every point where they search, switch tools, or ask, “What do I do now?”&lt;/li&gt;
&lt;li&gt;Cut at least one major step before the first win.&lt;/li&gt;
&lt;li&gt;Build one golden path and one fallback path.&lt;/li&gt;
&lt;li&gt;Make time-to-first-success a company-level GTM metric, not just a product metric.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;That is a very practical way to turn adoption into a growth lever.&lt;/p&gt;

&lt;h2&gt;
  
  
  My practical take
&lt;/h2&gt;

&lt;p&gt;One of the more useful truths in developer GTM is that product acquisition is often won or lost after the signup.&lt;/p&gt;

&lt;p&gt;That is the part many teams still underweight.&lt;/p&gt;

&lt;p&gt;They spend heavily to create interest, then quietly ask the developer to do too much work before the product proves itself. In a world where developers are already overloaded, already switching tools, and already losing time to friction, that is a very expensive mistake.&lt;/p&gt;

&lt;p&gt;The good news is that this is fixable.&lt;/p&gt;

&lt;p&gt;You do not need more hype.&lt;br&gt;
You need a cleaner first win.&lt;/p&gt;

&lt;p&gt;And once the first win gets faster, a lot of the rest of GTM starts working better too:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;activation improves&lt;/li&gt;
&lt;li&gt;support gets lighter&lt;/li&gt;
&lt;li&gt;sales gets cleaner signals&lt;/li&gt;
&lt;li&gt;community becomes more useful&lt;/li&gt;
&lt;li&gt;and the product starts feeling easier to believe in&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is what a real acquisition funnel looks like for developer products.&lt;/p&gt;

&lt;p&gt;It does not end at signup.&lt;/p&gt;

&lt;p&gt;It ends when the user succeeds.&lt;/p&gt;

</description>
      <category>saas</category>
      <category>development</category>
      <category>softwaredevelopment</category>
      <category>startup</category>
    </item>
    <item>
      <title>Stop Building AI Features Nobody Asked For</title>
      <dc:creator>Peter Strauss</dc:creator>
      <pubDate>Wed, 15 Apr 2026 08:40:03 +0000</pubDate>
      <link>https://forem.com/strauss/stop-building-ai-features-nobody-asked-for-koa</link>
      <guid>https://forem.com/strauss/stop-building-ai-features-nobody-asked-for-koa</guid>
      <description>&lt;p&gt;The easiest way to waste six months in 2026 is to build an AI feature that gets a great demo reaction and a weak buying reaction. &lt;/p&gt;

&lt;p&gt;People will tell you it’s cool. &lt;/p&gt;

&lt;p&gt;They will not tell you it is urgent. &lt;/p&gt;

&lt;p&gt;And unless you learn to hear the difference early, you can burn a lot of engineering time polishing curiosity instead of solving pain.&lt;/p&gt;

&lt;p&gt;I think this is one of the biggest traps for builders right now, especially developer founders. AI makes prototyping fast, which feels like an advantage. It is an advantage. But it also makes it dangerously easy to ship the wrong thing at high speed. When build cost drops, discipline has to go up.&lt;/p&gt;

&lt;h2&gt;
  
  
  The market is rewarding faster validation, not just faster building
&lt;/h2&gt;

&lt;p&gt;McKinsey’s latest &lt;a href="https://www.mckinsey.com/capabilities/business-building/our-insights/how-to-build-businesses-faster-and-better-with-ai" rel="noopener noreferrer"&gt;work on AI-first venture building&lt;/a&gt; says AI is materially compressing venture timelines, reducing time to validation, and raising output per person and per dollar. It also makes a very important distinction that I think a lot of founders miss: the winners are not just using AI to generate more ideas or code faster. They are validating more ideas earlier and scaling the winners sooner.&lt;/p&gt;

&lt;p&gt;That matters because most early-stage product mistakes are not engineering mistakes. They are market mistakes.&lt;/p&gt;

&lt;p&gt;The classic version of this is already familiar. In &lt;a href="https://www.cbinsights.com/research/report/startup-failure-reasons-top/" rel="noopener noreferrer"&gt;CB Insights’ long-running startup failure analysis&lt;/a&gt;, lack of market need remains one of the biggest reasons companies fail. AI does not remove that risk. It actually makes it easier to hide for a while, because you can produce a lot of convincing surface area before you ever prove someone will pay.&lt;/p&gt;

&lt;p&gt;That is the trap.&lt;/p&gt;

&lt;h2&gt;
  
  
  The harsh truth builders need to hear
&lt;/h2&gt;

&lt;p&gt;A lot of AI features are built because the founder can imagine the future, not because the customer is desperate in the present.&lt;/p&gt;

&lt;p&gt;That sounds harsh. It is also useful.&lt;/p&gt;

&lt;p&gt;There are usually three reasons teams build AI features too early:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;the market expects an AI story&lt;/li&gt;
&lt;li&gt;the prototype looks impressive fast&lt;/li&gt;
&lt;li&gt;the founder wants to keep up with competitors&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All three are understandable. None of them is evidence of demand.&lt;/p&gt;

&lt;p&gt;For a builder, that is one of the harder mindset shifts: being able to build something quickly is not a reason to build it now. The real question is whether the workflow is painful enough, frequent enough, and expensive enough that a customer will change behavior or budget for it.&lt;/p&gt;

&lt;p&gt;If the answer is vague, you are probably building a feature for applause.&lt;/p&gt;

&lt;h2&gt;
  
  
  My rule: sell the workflow before you automate the workflow
&lt;/h2&gt;

&lt;p&gt;This is the cleanest principle I know for avoiding AI feature waste.&lt;/p&gt;

&lt;p&gt;Do not ask:&lt;br&gt;
&lt;strong&gt;Can we build this?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Ask:&lt;br&gt;
&lt;strong&gt;Can we sell the outcome this workflow would create?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;That subtle change does a lot of work. It forces you to think in commercial terms before you get seduced by technical possibility.&lt;/p&gt;

&lt;p&gt;For developer tools, the workflow usually matters more than the model. Customers are not paying for “AI.” They are paying for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;faster code review&lt;/li&gt;
&lt;li&gt;safer releases&lt;/li&gt;
&lt;li&gt;easier API integration&lt;/li&gt;
&lt;li&gt;less debugging&lt;/li&gt;
&lt;li&gt;better test coverage&lt;/li&gt;
&lt;li&gt;faster onboarding for new engineers&lt;/li&gt;
&lt;li&gt;fewer support escalations&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is where the demand lives.&lt;/p&gt;

&lt;h2&gt;
  
  
  The practical fix: run a 5-customer AI validation sprint
&lt;/h2&gt;

&lt;p&gt;If I were running a devtools startup right now, this is the exact process I would use before shipping a serious AI feature.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 1: Name one painful workflow
&lt;/h3&gt;

&lt;p&gt;Not a category.&lt;br&gt;
Not a vision deck.&lt;br&gt;
One painful workflow.&lt;/p&gt;

&lt;p&gt;Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;generating test cases for flaky endpoints&lt;/li&gt;
&lt;li&gt;writing migration notes during releases&lt;/li&gt;
&lt;li&gt;triaging production incidents&lt;/li&gt;
&lt;li&gt;summarizing code-review feedback for junior developers&lt;/li&gt;
&lt;li&gt;creating API integration snippets from messy internal systems&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Good workflows have three traits:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;they happen often&lt;/li&gt;
&lt;li&gt;they are expensive or annoying&lt;/li&gt;
&lt;li&gt;they already make someone feel real pain&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Step 2: Write the offer before writing the feature
&lt;/h3&gt;

&lt;p&gt;This is where most teams skip the hard part.&lt;/p&gt;

&lt;p&gt;Write one page that says:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;who this is for&lt;/li&gt;
&lt;li&gt;what painful workflow it fixes&lt;/li&gt;
&lt;li&gt;what measurable outcome improves&lt;/li&gt;
&lt;li&gt;how quickly value should show up&lt;/li&gt;
&lt;li&gt;what it costs&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you cannot explain the commercial value in plain language, the feature is still too fuzzy.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Pre-sell to five target users
&lt;/h3&gt;

&lt;p&gt;Do not ask, “Would you use this?”&lt;br&gt;
That question creates fake optimism.&lt;/p&gt;

&lt;p&gt;Ask for one of these instead:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;a paid pilot&lt;/li&gt;
&lt;li&gt;a design partnership with defined success criteria&lt;/li&gt;
&lt;li&gt;access to real workflow data&lt;/li&gt;
&lt;li&gt;a commitment to trial the feature when it reaches a specific outcome&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The goal is not broad feedback.&lt;br&gt;
The goal is proof of seriousness.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: Deliver manually or semi-manually first
&lt;/h3&gt;

&lt;p&gt;This is the part builders hate and smart operators love.&lt;/p&gt;

&lt;p&gt;If the workflow matters, do some of it manually first. Use prompt chains, scripts, internal tools, or even human-in-the-loop delivery. You will learn:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;where context breaks&lt;/li&gt;
&lt;li&gt;what “good enough” actually means&lt;/li&gt;
&lt;li&gt;where the user still wants control&lt;/li&gt;
&lt;li&gt;which output format people trust&lt;/li&gt;
&lt;li&gt;what edge cases kill the experience&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That learning is worth a lot more than an elegant prototype built in isolation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Productize only what repeats
&lt;/h3&gt;

&lt;p&gt;Once you see the same workflow, correction pattern, and desired outcome repeating across users, then build.&lt;/p&gt;

&lt;p&gt;Not before.&lt;/p&gt;

&lt;p&gt;Build the repeated value.&lt;br&gt;
Not the imagined value.&lt;/p&gt;

&lt;h2&gt;
  
  
  A worked example
&lt;/h2&gt;

&lt;p&gt;Imagine a small startup building for backend teams. The founder wants to launch an AI feature that “explains pull requests automatically.” It sounds useful. Maybe even obviously useful. But that is still too abstract.&lt;/p&gt;

&lt;p&gt;The better move is to narrow it.&lt;/p&gt;

&lt;p&gt;Pick one workflow:&lt;br&gt;
&lt;strong&gt;help engineering managers review large PRs faster without missing risk&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Now pre-sell that outcome:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;reduce review time on large PRs&lt;/li&gt;
&lt;li&gt;surface risky file changes&lt;/li&gt;
&lt;li&gt;summarize architectural tradeoffs&lt;/li&gt;
&lt;li&gt;generate a comment-ready explanation the manager can trust&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Then run five pilots.&lt;br&gt;
Maybe two customers care mostly about security-sensitive diffs.&lt;br&gt;
Maybe one team wants onboarding context for newer reviewers.&lt;br&gt;
Maybe another team does not need summaries at all — they need change-risk ranking.&lt;/p&gt;

&lt;p&gt;Now the feature has shape.&lt;br&gt;
Not because the founder guessed better.&lt;br&gt;
Because the market taught the product what to become.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where AI really helps in this process
&lt;/h2&gt;

&lt;p&gt;This is the optimistic part.&lt;/p&gt;

&lt;p&gt;AI is fantastic for speeding up validation work if you use it correctly. McKinsey is right that AI can help teams generate, test, and refine ideas faster, run mini marketing experiments, and pressure-test concepts before full commitment. That does not replace customer conversations. It makes the early learning loop cheaper and faster.&lt;/p&gt;

&lt;p&gt;Use AI to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;map competing products and positioning&lt;/li&gt;
&lt;li&gt;generate landing page variants for smoke tests&lt;/li&gt;
&lt;li&gt;synthesize interview notes&lt;/li&gt;
&lt;li&gt;build lightweight prototypes for pilot users&lt;/li&gt;
&lt;li&gt;summarize repeated objections&lt;/li&gt;
&lt;li&gt;cluster workflow patterns across early testers&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That is real leverage.&lt;/p&gt;

&lt;p&gt;What I would not do is confuse internal enthusiasm with external proof.&lt;/p&gt;

&lt;h2&gt;
  
  
  My practical take
&lt;/h2&gt;

&lt;p&gt;One of the quiet business truths in the AI era is that it is now easier than ever to overbuild.&lt;/p&gt;

&lt;p&gt;That means good founders need a stronger filter, not just faster hands.&lt;/p&gt;

&lt;p&gt;The filter is simple:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;is the workflow painful?&lt;/li&gt;
&lt;li&gt;is the buyer specific?&lt;/li&gt;
&lt;li&gt;is the outcome measurable?&lt;/li&gt;
&lt;li&gt;has someone serious committed?&lt;/li&gt;
&lt;li&gt;did the same need repeat across multiple real users?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If the answer is no, slow down.&lt;br&gt;
If the answer is yes, then build aggressively.&lt;/p&gt;

&lt;p&gt;Because the real edge is not shipping more AI features.&lt;/p&gt;

&lt;p&gt;It is shipping fewer, better ones — the ones customers were already trying to buy before the code was finished.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>softwaredevelopment</category>
      <category>startup</category>
    </item>
  </channel>
</rss>
