<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Rewrite the Roadmap]]></title><description><![CDATA[The practical guides I wish I had growing up as a product manager.]]></description><link>https://www.rewritetheroadmap.com</link><image><url>https://substackcdn.com/image/fetch/$s_!Ba9y!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e0ec836-04a1-41a1-b871-8d7705033b30_256x256.png</url><title>Rewrite the Roadmap</title><link>https://www.rewritetheroadmap.com</link></image><generator>Substack</generator><lastBuildDate>Wed, 13 May 2026 11:23:07 GMT</lastBuildDate><atom:link href="https://www.rewritetheroadmap.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Luis Camara]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rewritetheroadmap@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rewritetheroadmap@substack.com]]></itunes:email><itunes:name><![CDATA[Luis Camara]]></itunes:name></itunes:owner><itunes:author><![CDATA[Luis Camara]]></itunes:author><googleplay:owner><![CDATA[rewritetheroadmap@substack.com]]></googleplay:owner><googleplay:email><![CDATA[rewritetheroadmap@substack.com]]></googleplay:email><googleplay:author><![CDATA[Luis Camara]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[How to: Define project scope]]></title><description><![CDATA[Finding your project's goldilocks zone]]></description><link>https://www.rewritetheroadmap.com/p/how-to-define-project-scope</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-define-project-scope</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Fri, 21 Feb 2025 03:44:06 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8f22f78c-f71a-4336-9de7-2c5f16d6bdd9_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I've seen countless projects fail before they even begin, and more often than not, the culprit is poor scoping. Projects where the scope is too large make it difficult to deliver meaningful value quickly, while projects where the scope is too narrow often fail to address the core problem effectively. The key to accurate scope is finding the middle ground that maximizes value while maintaining focus.</p><p>This article focuses on finding the right scope for your projects. It won&#8217;t go into detail on how to make trade-offs or the mechanics of refining scope. Instead, it will dive into what makes for good scope and how to evaluate if you&#8217;re correctly weighing the variables.</p><div><hr></div><h2>Scope = Value + Timeline</h2><p>Scope defines the volume and depth of solutions that are to be built to achieve an outcome. Scope is not something you want to define in full upfront; instead, it is a range of acceptable solutions that is continuously refined until you&#8217;ve reached a deliverable.</p><p><strong>At its core, scope is defined by two key questions:</strong></p><ul><li><p><strong>What specific value are we trying to create?</strong></p></li><li><p><strong>How long/much is the business willing to invest into extracting this value? </strong></p></li></ul><p>These two questions anchor the entire process of determining scope. The Goldilocks zone for scope is the place where your solution extracted an outcome while staying within the business&#8217; tolerated investment level. When we correctly answer these two questions we end up with projects that are likely to succeed. Over-scoped or under-scoped projects are often projects that fail to effectively answer one or both questions. </p><h3>Value</h3><p>Starting with value, value is the benefit to be gained for both the customer and the business as a result of pursuing the project. In other words, it&#8217;s the reason we&#8217;re doing the project. </p><p>One of the most important insights I've gained about value is that it&#8217;s easy to mistake function for value. The failure to understand this distinction can doom your project before it begins. Let me illustrate with a series of examples:</p><ul><li><p>Example 1</p><ul><li><p>Function: "I want to put water in a cup so I can drink it"</p></li><li><p>Value: "I want to feel hydrated"</p></li></ul></li><li><p>Example 2</p><ul><li><p>Function: "We need a messaging system with channels and direct messages"</p></li><li><p>Value: "Teams need to communicate effectively (whether that&#8217;s in a 1-to-1, 1-to-many or many-to-many setting) so that they can collaborate effectively"</p></li></ul></li><li><p>Example 3</p><ul><li><p>Function: "We need to display 20 different metrics with filtering options"</p></li><li><p>Value: "Analysts need to quickly gain insight into issues so that they can make data-driven decisions quickly"</p></li></ul></li></ul><p>This distinction might seem obvious, but it's crucial. <strong>When we focus on function, we lock ourselves into a specific solution.</strong> <strong>When we focus on value, we open ourselves to innovative approaches we might have otherwise missed.</strong></p><p>Because scope is a range of acceptable options, we look for value to be precise in terms of benefit but ambiguous in terms of implementation, as this allows teams the freedom to explore solutions until they find the right one.</p><p>The first key to finding your project&#8217;s scope is ensuring that you are rock solid on the value you&#8217;re creating and that you&#8217;ve decoupled the function from the value as much as you can.</p><h3>Timeline</h3><p><strong>Timeline is how long the business is willing to fund the value extraction</strong>. Every business has a patience threshold - even when they say, "Take as long as you need." they actually mean, &#8220;Take as long as you need within reason.&#8221; I try to convert this appetite level into a range of time explicitly. For example: </p><ul><li><p>"We need it ASAP" = Days-3 weeks</p></li><li><p>"Take the time to do it right" = 1-2 months</p></li><li><p>"Whatever it takes" = 4 months maximum</p></li></ul><p>While most people hate having timeline awareness, I&#8217;ve found that these constraints are net-helpful because they remove options and force you and the team to really focus on what is required to hit the goal and what is not. </p><p><strong>Note: </strong>A timeline is not the same as a deadline; timelines are guidance, and deadlines are restrictions; thinking of timelines is a better mechanism for teams to process scoping decisions as deadlines overpower value; you would rather the two be seen as linked, but where necessary, value always beats timeline.</p><p></p><h4>Value is the target, timeline is the appetite</h4><p>While value provides you with a clear target, a timeline tells you how long you have to hit it. A team with clear value targets will often hit these targets but take far too long to do so, a team with timeline constraints will hit a timeline but miss its target, a team with both will find its way to a successful feature. </p><p>It is essential that a team understands the answer to these questions and is bought in. I&#8217;ve found that conflict during projects is often tied to a misalignment around these questions or a lack of explicit answers. </p><div><hr></div><h2>Finding scope in practice</h2><p><a href="https://www.rewritetheroadmap.com/p/how-to-identify-and-answer-critical-questions?r=3du6f&amp;utm_campaign=post&amp;utm_medium=web&amp;showWelcomeOnShare=false">You will find a more in-depth article about the process of refining scope here</a>, but at a high level, this is how I go about this process:</p><p>I always start by translating my project goals into simple, high-level user stories that focus on value rather than implementation. I then work with others/leadership to understand the acceptable investment level and develop my sense of what that looks like in terms of timeframe and how flexible it is. This gives me a clear picture of what value we&#8217;re trying to create and how long we&#8217;ll have to get there.</p><p>I will then take that context and bring it to the team, explaining my reasoning and beliefs in the process. This sets up the foundation for team exploration of a potential solution, which usually follows a general path of:</p><ol><li><p>Open ideation around identifying possible solution paths</p></li><li><p>Evaluation of options against value and time constraints</p></li><li><p>Picking the strongest starting point solution that delivers the core value</p></li><li><p>Exploring that starting pointing and developing it into a lo-fi solution that validates if the path is worth continuing</p></li><li><p>As we gain confidence that the path works, prioritizing the list of enhancements needed to achieve a release.</p></li></ol><p>Fundamentally, the process of refining scope is focused on exploring paths and quickly and effectively identifying the best one. By doing this, you permit yourself and the team to make scoping decisions along the way as you gain more information as opposed to all upfront (which often leads you to bloating releases and then having to cut scope at the last mile). This incremental path allows your scope to gradually get narrower and narrower until you&#8217;re left with the ultimate vision of what your solution should be.</p><h4>Find scope along the way, nail down scope at the end</h4><p>It is my belief that scope should be finalized as late as possible while being defined as gradually as possible. In an ideal world you&#8217;re getting to 100% final scope at about 80-90% completion of the project. This allows you and the team to continue being flexible and making decisions about what matters late into the project.</p><p>Now this doesn&#8217;t mean you have defined scope and are cutting at the end, it means that you have options and are choosing which ones make sense within the release and which ones don&#8217;t. To achieve this outcome, here are some practical tips I've found valuable:</p><ol><li><p><strong>Always start smaller than you think you need to.</strong></p><ol><li><p>The smaller the starting point, the easier it is to move and make decisions, as there will be less tying you down. Always start in the order of most important and add along the way. It&#8217;s always easier to add scope than it is to remove it.</p></li></ol></li><li><p><strong>Add scope incrementally as you confirm progress.</strong></p><ol><li><p>Oftentimes, teams make decisions before needing to - this leads to bloat. Allow yourself the pain of needing something, as that pain is the clearest guide you&#8217;ll have for scope. Focus on making incremental decisions that expand the depth and volume of the solution as these decisions become necessities or forks in the road. </p></li></ol></li><li><p><strong>Never delete, just backlog and prioritize for follow-up releases</strong></p><ol><li><p>Teams will often fight tooth and nail for scope because they fear this is the only chance they&#8217;ll have to build it. Create a culture where each release is not an end but instead is a stepping stone, this will allow the team to make scope decisions more comfortably. Question every "must-have" feature against your value definition and keep a list of &#8220;nice-to-have&#8221; features for future releases. </p></li></ol></li><li><p><strong>Make scope decisions (and their stakes) shared across the team</strong></p><ol><li><p>The more the team feels responsible and empowered to achieve value within the timeline, the better the decision-making will be, and the easier it will be to find the right solution for your problem. As much as you can avoid being the sole individual responsible for scoping decisions. </p></li></ol></li></ol><p></p><div><hr></div><h2>Parting thoughts:</h2><p>Managing scope is one of those things in product management that will never stop being a challenge, as each project has its own world of scoping decisions. While a framework can help you more consistently navigate this opportunity, be ready for each challenge to have its own unique quirks and necessities. </p><p>Over time, finding the Goldilocks zone for scope becomes more natural. The key is reminding yourself to always stay focused on value while maintaining the flexibility to adjust as you learn.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Get people to believe in product management]]></title><description><![CDATA[Building authority when everyone else has hard skills and you have... making slides]]></description><link>https://www.rewritetheroadmap.com/p/how-to-get-people-to-believe-in-product-management</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-get-people-to-believe-in-product-management</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Tue, 11 Feb 2025 04:25:07 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/21e5e98b-2dc8-4ea1-9990-b08964864f77_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Product management can feel like a tough sell. Engineers and designers build tangible things, while we appear to mainly make tickets and run meetings. But here's the truth: product management isn't just about presentations and JIRA tickets. It's about orchestrating value creation and getting everyone to elevate their game so that they can achieve something beyond expectations.</p><p>Early in my career, I inherited a team that had been burned by previous product initiatives. Features would mysteriously vanish from roadmaps, requirements would shift without explanation, and promises would evaporate under pressure from leadership. The result? A deeply skeptical group that viewed product management as an unnecessary layer of bureaucracy.</p><p>Sound familiar? Many PMs face similar challenges. The key to overcoming this skepticism lies not in defending the role's existence but in demonstrating its value through consistent, principled action. Here's how to do it.</p><div><hr></div><h2>The 7 Principles of Building Product Management Authority</h2><p></p><h2>1. Build Trust Through Predictability</h2><p>The foundation of product management authority isn't about being liked - it's about being trusted. I learned this lesson the hard way, having spent my first years oscillating between trying to be everyone's friend and clinical detachment. Neither worked. What did work was establishing a consistent, predictable approach to:</p><ul><li><p>Decision-making: Using clear frameworks that teams could understand and anticipate allowed teams to predict where I&#8217;d stand and adjust accordingly.</p></li><li><p>Communication: Maintaining transparency about what's changing and why and updating teams regularly helped teams focus on their work.</p></li><li><p>Commitment management: Following through on promises, no matter how small, even when the cost to me was high, gave teams faith in my word and its weight.</p></li></ul><p>I found that when teams can predict your behavior and methods, they become easier to work with. This predictability removes ambiguity and builds trust, which evolves into respect. Teams quickly lose respect for managers whose actions seem arbitrary - they'll always prefer someone predictable over someone unpredictable, even if they disagree more often with the predictable one.</p><p></p><h2>2. Master the Art of Collaboration</h2><p>Product management isn't about being the smartest person in the room - it's about orchestrating collective intelligence. The faster you develop your ability to get people to work together effectively, the faster the team will understand the value of product management. Here's what effective collaborative leadership looks like in practice:</p><ul><li><p>There are forums for input about projects and decisions and those forums feel appropriate and heard</p></li><li><p>Discussions feel free-flowing and shared across many team members as opposed to dictated by the PM</p></li><li><p>Team members understand and trust specific individuals to navigate challenges within their domain and communicate their processes with the team</p></li><li><p>There is a strong sense of accountability across the team without fear of repercussion or blame</p></li></ul><p>Every team member should have input on product value; the key is finding the structure and moments to make that collaboration happen. An effective PM will quickly show their value through their ability to make teams work better together. </p><p></p><h2>3. Demonstrate Unwavering Resilience</h2><p>For teams to believe in product management, they need to see that our principles hold up under pressure. If product management betrays its principles at the first encounter with friction, then it will fail to inspire trust in the team. Withstanding pressure generally means:</p><ul><li><p>Maintaining consistent processes even when facing setbacks</p></li><li><p>Using failures as learning opportunities rather than excuses</p></li><li><p>Pushing through organizational resistance with principled persistence</p></li><li><p>Showing that setbacks strengthen rather than weaken our approach</p></li></ul><p>When teams see that you remain steady and focused on improvement regardless of circumstances, they begin to trust that they can adapt to your leadership style and methodology. </p><p>When teams believe their PM will not waver and will keep pushing to improve things, they will accept that reality and adapt to it. By being unrelenting, you&#8217;re sending a message to your team that they can safely adjust to you and that they can trust that you will be consistent regardless of what comes your way.</p><p></p><h2>4. Champion the Long-term Vision</h2><p>One of product management's most valuable contributions is maintaining focus on the bigger picture. In a detail-oriented environment, having an individual who can broaden the perspective is often very helpful in improving team performance. This is something technical teams often realize and seek out. Specifically, owning and articulating your product&#8217;s vision impacts the team in different ways:</p><ul><li><p>It provides context for decisions and trade-offs</p></li><li><p>Teams are motivated by understanding their work's impact</p></li><li><p>It creates a shared sense of direction</p></li><li><p>It helps teams understand the "why" behind decisions</p></li><li><p>It empowers teams to filter and debate decisions through a shared lense</p></li></ul><p>Most individuals have a strong inner desire to be part of something bigger than themselves. When teams believe their mission is worthy, they will find ways to elevate themselves to meet their goals. </p><p>At its best, product management is about getting people to see the potential in something and believing they can meet it together. By providing this value to your team, you will indirectly illustrate to the team that product management brings an important intangible variable to the work, and they will value it more as a a result.</p><p></p><h2>5. Develop Craft Knowledge</h2><p>Nothing undermines PM credibility faster than strong opinions paired with shallow understanding. While having opinions is core to the role, speaking confidently without understanding fundamentals undermines credibility. </p><p>You're often working with people whose entire career focuses on their craft - they need to see you value and respect it to accept your input as valid. Here there are several ways to get better:</p><ul><li><p>Invest time in understanding technical fundamentals through courses, books, projects </p></li><li><p>Learn continuously from team members' expertise by proactively asking questions, especially when you feel those questions are dumb</p></li><li><p>Seek to participate in projects that challenge your technical knowledge in new ways</p></li></ul><p>The key here is that you accept that the subject matter you work with is a world in itself and that you have a commitment to being knowledgeable about that world and respecting it for what it is. Your world can be software, hardware, or anything else. The key is that you commit yourself to becoming someone who understands the inner workings of that world and is constantly trying to build a better understanding of how they can be better at the creating process.</p><p></p><h2>6. Deliver Measurable Results</h2><p>Product management ultimately depends on results. To believe in it, teams need to see tangible, differentiated value from your approach. When teams win, ideas become easy to implement and buy into; when teams struggle to produce value, teams start isolating and fighting against their structure.</p><p>There are many, many ways to get better at delivering results (many of my past articles address this problem directly). However, there are some easy strategies you can pursue to highlight how practices are tied to results:</p><ul><li><p>Implement value forecasting and project retrospectives to identify how/if practices are leading to more consistent value delivery</p></li><li><p>Show your team how proper project scoring leads to better prioritization and decision-making through examples and work sessions</p></li><li><p>Connect product management practices to concrete results in tangible ways through stories and scenarios</p></li></ul><p>Remember: you're a business person first, equipped with tools and perspective to drive value creation through collaborative leadership. Your actions and, most importantly, your results will determine whether others see product management as essential or expendable. By making product management practices well understood and tying those practices to tangible results, you can help build a shared belief in the value of your role. </p><p></p><h2>7. Become a Product Evangelist</h2><p>Effective product management requires more than process expertise - it demands genuine belief in your product's potential impact. Being a hardcore fan of your product helps teams become bigger fans of you. This is because authentic passion for your product:</p><ul><li><p>Reveals to the team that you care about what is being built and that you will care for the work they do</p></li><li><p>Inspires teams to push beyond immediate obstacles and, in doing so, builds a connection with the team</p></li><li><p>Creates emotional investment in outcomes, which often fuels creativity and innovation</p></li><li><p>Builds resilience during challenging periods and creates a &#8220;we&#8217;re in this together&#8221; mindset</p></li></ul><p>The industry has many mercenary PMs - those motivated primarily by compensation and advancement. Truly effective product leadership requires being an evangelist who can inspire others even when sacrifices are needed. Your genuine belief in your product's impact and approach to building it can spark similar dedication in others.</p><p>Authenticity matters here; if teams believe in your commitment and passion, they will be infected by it and carry it on. Passion is a fuel source for creativity and collaboration, and that fuel will open the team up to product management practices as they will relate these to a desire to make better things.</p><p></p><div><hr></div><h2>Parting thoughts:</h2><p>Building belief in product management isn't a one-time achievement. It's built and rebuilt through daily actions, decisions, and interactions. Teams may be initially skeptical of product management, but consistent application of these principles can transform that skepticism into trust and eventually into belief.</p><p>I've found that organizations don't really change their mind about product management through theoretical arguments. They change their mind through experiencing effective product management in action. Each time you demonstrate trustworthiness, show deep craft knowledge, and maintain long-term vision while delivering near-term results, you're not just doing your job - you're showing what product management at its best can be.</p><p>The future of product management isn't determined by industry trends or thought leaders, It's determined by how we practice it every day. Focus on embodying these principles consistently, and the value of product management will become self-evident to your organization.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Identify feature defining questions]]></title><description><![CDATA[The art of asking less, shipping more]]></description><link>https://www.rewritetheroadmap.com/p/how-to-identify-and-answer-critical-questions</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-identify-and-answer-critical-questions</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sat, 21 Dec 2024 17:26:12 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/432ef853-137b-46c7-8642-9ec06228e6cd_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The success of any product project hinges on two timelines: how long it takes to figure out what to build and how long it takes to build it. Most teams obsess over the latter while underestimating the former. </p><p>Every product team faces an endless flood of questions, problems, and possibilities competing for attention. Yet, in my experience, successful releases ultimately come down to answering just 1-3 critical questions. Get these right, and everything else falls into place. Get them wrong, and no amount of perfect execution can save you.</p><p>This article will show you how to identify these critical questions, align your team around them, answer them decisively, and use those answers to drive successful and faster execution. We'll cover:</p><ul><li><p>What critical questions are and why they matter</p></li><li><p>The three types of critical questions you'll encounter (value, solution, technical)</p></li><li><p>How to find critical questions as a team</p></li><li><p>A practical framework for answering these questions quickly</p></li><li><p>How to know when you've found your answers</p></li><li><p>Moving from answers to execution</p></li><li><p>Getting organizational buy-in for this approach</p></li></ul><div><hr></div><h2>What is a critical question:</h2><p>A critical question is the make-or-break challenge that determines a feature release's success or failure. While releases are complex, their outcomes usually hinge on answering just 1-3 fundamental questions correctly.</p><p>This insight has reshaped how I approach releases. Instead of trying to solve everything at once, I work with the team to build a release strategy around answering these critical questions quickly and decisively. By doing so, I&#8217;ve found you can derisk the entire release early and consistently drive value. Once you have these answers, they become the foundation for all other decisions, accelerating development and increasing your chances of success.</p><p>Critical questions typically fall into three primary categories:</p><h3><strong>1. Value extraction: Is it worth it?</strong> </h3><p>This is the most common type of critical question. It asks whether your project's core premise is valid and if you can deliver meaningful value to users through it. The answer might be 'no'&#8212;and that's valuable to learn early as it allows you to either pivot toward real value or stop the project before heavy investment.</p><p>Projects often fail not from poor execution, but because they conflict with user realities, expectations, or needs. Solving this question allows you to prove a release&#8217;s worth without the cost of a full cycle behind it.</p><p><strong>Example variations of the question:</strong></p><ul><li><p>Will users consider this worth their time?</p></li><li><p>Is the core assumption the solution is built on a true assumption?</p></li><li><p>Will the value be immediately obvious to users?</p></li><li><p>Can we get this feature sold to the right people?</p></li><li><p>Can we get people to experience valuable results every time they use it?</p></li><li><p>Can we maintain enough quality content to power the experience?</p></li><li><p>Will ongoing support costs outweigh the value delivered?</p></li><li><p>Is now the right time to push this feature through?</p></li></ul><p> </p><h3><strong>2. Solution shape: How will it work?</strong></h3><p>This type of question comes up when your solution's form is so crucial that getting it wrong would prevent users from experiencing any value. This becomes critical in projects where workflow integration or user friction can make or break success.</p><p><strong>Example variations of the question:</strong></p><ul><li><p>Will this fit seamlessly into existing user workflows?</p></li><li><p>Can users complete key tasks consistently and efficiently?</p></li><li><p>Can we minimize friction in critical paths?</p></li><li><p>Can we make the experience engaging and feel rewarding at key steps?</p></li><li><p>Will this coexist with related features and systems?</p></li><li><p>Do we have the right framework for how this experience should work?</p></li></ul><p></p><h3><strong>3. Technical approach: Can we build it?</strong></h3><p>Here, we&#8217;re trying to determine whether a technological decision completely defines the solution and makes it possible or impossible. This question often appears when the solution that is desired is at the cutting edge or when there is a significant constraint that may make the solution&#8217;s cost, reliability, or capability untenable.</p><p><strong>Example variations of the question:</strong></p><ul><li><p>Can we make this feature really fast and performant?</p></li></ul><ul><li><p>Can we scale up this feature to the demands of our user base effectively?</p></li><li><p>Is the technology we&#8217;re using to build this feature mature enough?</p></li><li><p>Can we support this feature with our existing team&#8217;s knowledge and skill sets?</p></li><li><p>Are we going to regret this approach in 6-12 months?</p></li></ul><div><hr></div><h2>Finding critical questions as a team</h2><p>Now that we&#8217;ve covered the common types of critical questions we can run through how to identify them within a release and go about solving them. </p><p>Identifying critical questions isn't a solo sport&#8212;it requires your entire team's perspective and insight. Before diving in, ensure your team has a shared understanding of five key areas:</p><ul><li><p><strong>The Opportunity:</strong> What specific challenge are we addressing?</p></li><li><p><strong>The Context:</strong> Why is this worth solving, and why are we uniquely positioned to solve it?</p></li><li><p><strong>The Value:</strong> What concrete benefits will our customers experience?</p></li><li><p><strong>The Usage Scenarios:</strong> In what real-world situations will this solution create value?</p></li><li><p><strong>The Business Constraints:</strong> What practical limitations around cost, integration, and scale must we consider?</p></li></ul><p>Really take the time to validate and refine these knowledge areas with the team as nailing this background will improve your odds of identifying the critical question on the first try.</p><h3>Crafting the question:</h3><p>Once your team has built this shared context, you can begin crafting your critical question. While you can have up to three, aim for one powerful question that captures your core challenge. Your critical question should be:</p><ul><li><p><strong>Concise:</strong> Expressed in one or two clear sentences</p></li><li><p><strong>Specific:</strong> Focused on a single-core challenge</p></li><li><p><strong>Actionable:</strong> Something you can validate through experimentation</p></li></ul><p>Have a session with your team framed around identifying what matters bring with you your thoughts on what the question(s) is and why you feel that way. Then, ask the team to express their own thoughts. If you don&#8217;t have alignment, continue to push through and debate ideas until you have it. Do not compromise halfway, as this often ends up being ineffective; instead of splitting the difference, empower people to go and make something to illustrate their point, or if that&#8217;s not possible, pick an initial approach and keep open the promise of reviewing the need to pursue the alternate direction if needed.</p><p>There are often strong opinions and lengthy debates, which means there&#8217;s passion behind the project, which is always valuable. Don&#8217;t feel the need to suppress that energy. Instead, focus on channeling that passion in the direction of &#8220;let&#8217;s prove this out&#8221; and give the most passionate people the opportunity to make something (if feasible). If two team members have very strong competing ideas, have them prototype their approaches and let those approaches speak for themselves. You&#8217;re looking to cultivate &#8220;a best idea wins&#8221; environment, not a &#8220;best argument wins&#8221; environment. If you have team members who are more silent or don&#8217;t enjoy participating in public debates, try to have private conversions or have them write their thoughts down first and bring their ideas to the table. Focus on getting everyone involved as this will give you the most angles to evaluate.</p><p>Also, don&#8217;t worry about getting the question perfect immediately, and remind the team that evolution will happen. Trust your team's instincts&#8212;you're better off starting with a good enough question than getting stuck in endless debate. You can always refine your question as you learn more. As long as the team feels you&#8217;re generally in the ballpark, you can probably get started.</p><div><hr></div><h2>Answering the critical questions</h2><p>Once you've identified your critical questions, it's time to find answers. But here's a crucial mindset shift from most project approaches: you are not pursuing perfection or conducting exhaustive research. Your goal is to answer these questions as quickly and efficiently as possible.</p><p>Teams often fall into a trap at this stage. They get caught in endless cycles of debate and persuasion, spending weeks or months discussing theoretical solutions. Meetings pile up. Documents grow longer. Everyone has an opinion to share. Before you know it, you've spent more time talking about the solution than actually exploring it.</p><p>But here's the uncomfortable truth: you can't argue your way to a great product. No amount of passionate debate or perfectly crafted presentations will tell you if you're on the right track. The only way to genuinely answer your critical questions is to put something real in front of stakeholders (and often users) and have the solution prove its worth. This means creating tangible artifacts early&#8212;prototypes, mockups, proofs-of-concept, or simplified solutions that test your core assumptions. These don't need to be perfect; they just need to be real enough to generate authentic reactions and insights. </p><h3>Create, Play, Discuss, Evaluate (Repeat)</h3><p>The heart of answering critical questions lies in rapid cycles of experimentation and learning. At its core, this cycle has four key elements:</p><ol><li><p><strong>Create something tangible</strong>, even if it's rough. Shoot to build a proof-of-concept or work through a very basic implementation of the thing. The key is that it's something people can react to, not just discuss theoretically. </p></li><li><p><strong>Play with what you've created</strong>. Get it in front of people&#8212;team members, users, stakeholders&#8212;and let them interact with it naturally. Give them alone time with it in the order of days. This hands-on interaction often reveals insights that no amount of explanation could uncover.</p></li><li><p><strong>Discuss what you've learned</strong>, but keep it focused. What surprised people? What excited them? Where did they struggle? Is this looking viable? These real-world observations are worth more than hours of theoretical debate.</p></li><li><p><strong>Evaluate honestly whether you're getting closer</strong> to answering your critical question. Is your confidence growing or wavering? Are you seeing patterns in the feedback? Does the team's energy and excitement build with each iteration?</p></li><li><p><strong>Take all the new knowledge gained and apply it </strong>to create a better version of what you had before. Keep going until you have something the team really believes in.</p></li></ol><p>What makes this cycle powerful is how it transforms abstract questions into concrete learning. Instead of endlessly debating possibilities, you're constantly creating and testing hypotheses. The solution becomes a living, evolving thing that the entire team can see, touch, and improve.</p><p>This approach also tends to energize teams. There's something inherently motivating about seeing your ideas take shape and evolve based on real feedback. Each cycle builds momentum, replacing the fatigue of endless discussion with the excitement of discovery.</p><p><strong>Remember:</strong> be ruthless about throwing away anything that doesn't directly help answer your critical question. There will be plenty of time later to polish features and add capabilities. Right now, your only goal is to find that answer as efficiently as possible. Every piece of effort should be laser-focused on building confidence in your direction.<br></p><h3>Knowing when you found the answer: The 70-80% Confidence Rule</h3><p>After several cycles of creation and learning, how do you know when you've sufficiently answered your critical question? In my experience, you're looking for 70-80% confidence in your answer. But this isn't about gut feelings or arbitrary thresholds- it's about recognizing specific signals that indicate you're ready to move forward.</p><p><strong>Signs of 70-80% Confidence:</strong></p><ul><li><p>Your proof-of-concepts validate core assumptions about the solution.</p></li><li><p>Team members consistently tell the same story about how the solution will work and why.</p></li><li><p>Stakeholders show genuine enthusiasm for what they&#8217;re interacting with.</p></li><li><p>User feedback consistently validates your core assumptions and approach.</p></li><li><p>You can point to specific, realistic scenarios where your solution delivers clear value.</p></li><li><p>Everyone feels confident that the questions you came into this process with are answered.</p></li></ul><p><strong>Warning Signs You're Not There Yet:</strong></p><ul><li><p>Team members describe fundamentally different solutions when asked about the approach.</p></li><li><p>Key team members express lingering doubts about the current direction.</p></li><li><p>Fundamental debates about the approach keep resurfacing.</p></li><li><p>User feedback remains mixed or unclear about the core value.</p></li><li><p>The team struggles to articulate how users will experience value.</p></li><li><p>Your team isn&#8217;t aligned about the shape of the questions being answered or about how well you answered them.</p></li></ul><p>Once you reach the 70% confidence threshold, project execution often becomes almost intuitive. The team shares a strong, unified understanding of where you're going and why. Decisions become clearer because you have a solid foundation against which to evaluate them. You can move faster because you're no longer questioning the fundamentals.</p><p>Getting to 70-80% confidence doesn't mean having everything figured out. You'll still have open questions and details to resolve. Trust that if you've followed the process and seen these positive signals, you've built the foundation you need for successful execution and move on through the rest of the project.</p><div><hr></div><h2>From Answer to Execution: Riding the Momentum</h2><p>When you start seeing promising signals, it's time to shift gears. The momentum you've built to answer your critical question is valuable. Use it. Continue exposing work and gathering feedback, but now with a broader lens. Your incremental builds can tackle bigger chunks of the solution, always staying grounded in the answers you&#8217;ve arrived at.</p><p>Execution brings its own set of complex problems to solve&#8212;but here's the difference: you're now solving them from a position of strength. Your team shares a deep understanding of the core value being delivered and a clear vision of where you're headed.</p><p>Keep these principles in mind as you move forward:</p><ul><li><p>Stay iterative: Break down the remaining work into smaller, focused challenges.</p></li><li><p>Maintain momentum: Keep the team moving and building, and use exposure to the product to generate this momentum.</p></li><li><p>Stay grounded: Regularly reconnect with your core answers and validate what you&#8217;re doing against them.</p></li><li><p>Remain flexible: Be ready to adapt as you learn more, possibly even revisiting your original answers if needed.</p></li><li><p>Keep learning: Continue gathering feedback and seek to increase the scale and depth of exposure for your work-in-progress solution.</p></li></ul><p>Is this approach easy to execute? Not at all. Constant discipline is required to maintain focus and make tough choices. You'll face pressure to chase interesting but peripheral problems. You'll be tempted to revisit settled questions. You'll want to optimize details before their time.</p><p>But if you can maintain your focus on what truly matters&#8212;if you can keep cutting through the noise and stay grounded in your core answers&#8212;you'll find yourself building products that not only work but also truly resonate with users.</p><div><hr></div><h3>Case Study: Building Social Proof Through Search</h3><p>The following case study shows how this framework played out in a real project. It illustrates how focusing on critical questions helped us build a valuable search feature efficiently.<br></p><h4><strong>The Context:</strong></h4><p><strong>The Opportunity:</strong> We have a huge network of users but people who don't know us are often skeptical of the size of that network, and, as a result, they're skeptical of the value of being part of that network.</p><p><strong>The Context:</strong> This problem is worth solving as it will improve the odds that users sign up and join the network. We're uniquely positioned to solve this problem as we have the strongest network and very clean/complete user data for members of our network.</p><p><strong>The Value:</strong> Users with this concern will feel that they're getting honest, transparent answers as they'll be able to find people they know and validate that the network is worth joining.</p><p><strong>The Usage Scenarios:</strong> When users come to our website from Google searches or any other entry-point, they're able to undertake an in-product search for someone they know.</p><p><strong>The Business Constraints:</strong> We need to support search behavior at our current traffic volume and beyond. We need to ensure that we're delivering valuable results each time and minimizing the frequency of poor/few results. We need to update our underlying data frequently to ensure searches show updated results.<br></p><h4>Finding the Question: </h4><p>From here, the team had a long discussion where different stakeholders pursued investigations in their domain. Design spent time thinking about how this experience integrates with our site, engineering looked at different options for building the search and prioritizing results, and product spent time mapping the most common search patterns likely to be reproduced on the site.</p><p>The team then identified one core question: </p><ul><li><p><strong>Can we deliver accurate, valuable search results at high rates?</strong></p></li></ul><p>This was the critical question that defined the release&#8217;s odds of success. For the feature to work, search results needed to be consistently valuable. <br></p><h4>Answering the Question: </h4><p>The team found that the most effective way to answer the question was to:</p><ul><li><p>Build a dummy page with an unpolished-looking search function</p></li><li><p>Tie that search function to our production DB so people can search within the actual user DB that real users would use</p></li><li><p>Focus primarily on the search prioritization logic and test that logic for specific search patterns we identified</p></li></ul><p>The team spent about five cycles just iterating on search prioritization logic and ended up with something that very effectively solved the real search cases we believed users would undertake. We also opened up the search prototype to other people at the company, and these people tested it and started using it to solve their own separate use cases.</p><p>Eventually, the team decided the search logic was strong. It wasn't perfect, but pretty effective and strong enough to advance. From here, the team moved to the execution phase, breaking down the remaining release into a sequence of items (building the final UI, building remaining backend functionality, and integrating it onto different pages).</p><p>This case study shows how focusing on one critical question - can we deliver valuable search results consistently - helped us validate our core assumption before investing in the full feature build. By using a stripped-down prototype, we could iterate quickly on what mattered most while avoiding getting stuck on less important details early on.</p><div><hr></div><h2>Getting organizational buy-in for this approach</h2><p>Teams often resist changing their workflows, even when current processes aren't working well. At first glance, the critical questions approach might seem inefficient since you're not planning the full solution from day one. Here's how I've successfully sold this methodology and gotten it adopted:</p><p><strong>Start with pain points</strong>: <br>Most teams have vivid memories of building the wrong thing&#8212;projects that consumed months of effort only to miss the mark with users. Use these examples to illustrate the cost of moving forward without answering fundamental questions first. Make it clear that focusing on critical questions early dramatically reduces the risk of these expensive mistakes.</p><p><strong>Emphasize better decision-making:</strong><br>The critical questions approach gives teams tangible evidence about whether a solution works before major investment. This concrete understanding leads to faster, more confident decisions about scope and direction compared to trying to plan everything upfront.</p><p><strong>Prove it through practice:<br></strong>Rather than pushing for widespread adoption, start with a single project. Use this as your proof of concept, regularly sharing progress and wins with stakeholders. Let the results speak for themselves&#8212;teams that see this approach in action quickly recognize its value.</p><p>While this methodology might not feel necessary for teams that are already shipping quickly and effectively, it's particularly powerful for organizations stuck in slow, waterfall-style processes. It can provide a practical path to more efficient, focused product development while reducing the risk of building the wrong thing.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Balance your roadmap]]></title><description><![CDATA[Rocks, pebbles, and sand.]]></description><link>https://www.rewritetheroadmap.com/p/how-to-balance-your-roadmap</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-balance-your-roadmap</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sun, 24 Nov 2024 19:09:24 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2442c44d-dfad-4f36-830d-3f5fc752a541_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>An imbalanced roadmap is a roadmap that is overinvested into a single type of approach to succeed. This is a dangerous kind of roadmap because it strikes a single strategy and, as a result, lives and dies by that strategy. </p><p>On the other hand, a balanced roadmap pursues multiple different strategies (often concurrently) in an attempt to improve its odds of succeeding. This article isn&#8217;t about the content of the roadmap but instead focuses on its form. Specifically, it focuses on the sizes of projects to be pursued and the quantity of each type. </p><p>This article makes the case for a more balanced roadmap-building strategy as well as how to pull it off.</p><div><hr></div><h2><strong>Balancing the roadmap: rocks, pebbles and sand</strong></h2><p>You might have heard of the empty jar concept before, I&#8217;ll extend it here to fit the context of a product roadmap. Think of your roadmap as an empty jar that you have to fill. To fill it you have <strong>rocks</strong>, <strong>pebbles</strong>, and <strong>sand</strong>. </p><ul><li><p><strong>Rocks: </strong>These are your large projects, usually new features, revamping old ones, or tackling large tech debt projects. Usually around 1-3 months of work.</p></li><li><p><strong>Pebbles: </strong>These are your midsized projects, usually small features, feature improvements, performance enhancements, and other mid-sized iterations. Usually around 2-3 weeks of work.</p></li><li><p><strong>Sand: </strong>These are your small projects, usually smaller feature improvements, adjustments, and other small iterations. Usually around 1-1.5 weeks of work. </p></li></ul><p>I draw the line at rocks as anything larger is very prone to failing. A project longer than 3 months to me is a mountain, and you might as well break it down into rocks, pebbles, and sand. What this model implies is also that everything, at the end of the day, is the same matter, just broken down to a size that maximizes your ability to extract value and get that thing done.</p><p>Given these elements, what you want to do is give yourself the opportunity to take big swings but derisk those projects with smaller projects and iterations.  The big swing gets you the home run, but by pursuing some smaller projects, you&#8217;re still liable to get a single or double per plate appearance even when you miss a big swing.</p><p></p><h3>Rocks</h3><p>The roadmap can only really start with the anchor pieces and these anchor pieces are your rocks. If you start by filling the jar with sand and pebbles, you&#8217;ll quickly run out of room for the rocks. You&#8217;ll then find yourself emptying the jar to get the rocks in. If you put the rocks into the jar first, you can fill the remaining empty space with pebbles and sand. </p><p><strong>Commit to your large projects first but make room for the rest.</strong> If you&#8217;re in a rock-obsessed culture, start by trying to create a commitment of only 50-65% of your time spent on rocks with the rest spread across pebbles and sand.<strong> </strong>The roadmap will need all three kinds of work to achieve balance just as the jar will need all three elements to be full.</p><p></p><h4>Don&#8217;t fall into the rock trap:</h4><p>There is often a lot of temptation to only try to fit rocks into the jar but you&#8217;ll find that the jar simply can&#8217;t fit too many rocks and that you&#8217;ll feel that a lot was left on the table. </p><p>In most people, there is a natural bias towards large-scale projects. There are a few reasons behind this:</p><ul><li><p>Large-scale projects are generally easier to visualize, describe, and arrive at, </p></li><li><p>They don&#8217;t force you to make tradeoffs (when the scope is large enough, it&#8217;s always easy to just add another thing to make the project make more sense), </p></li><li><p>It&#8217;s easy to picture the outcome of a large-scale project and even easier to imagine the upside being transformative to the business.</p></li></ul><p>Unfortunately, large projects cost a lot and take a long time to achieve (which reduces your ability to pursue other projects), given the scale, they require making lots of assumptions, which creates lots of room for error. </p><p>Large projects are worth pursuing but spending 12 months on 3-4 projects for most teams is a terrible idea as you&#8217;re really hoping those 3-4 things hit big and are giving yourself very little room for iteration. When you don&#8217;t hit big the downside can be catastrophic.</p><p></p><h3>Pebbles and Sand:</h3><p>You want to have a lot of room for sand and pebbles for a simple reason: <strong>The more shots you take, the higher the projected point total you&#8217;ll have</strong>. This doesn&#8217;t mean doing a bunch of crappy work, just that when all things are the same, you&#8217;re better off doing more than less. The easiest way to do this is to do more mid-to-small sized things.</p><p>This will require a big commitment from your organization: <strong>The entire system depends on you having room for small and mid-sized work.</strong> I&#8217;ve found this buy-in to be the single biggest challenge to roadmap building as most people, even when they say they want iteration, naturally only want large-scale projects. However, the sand and pebbles are what will give your product the ability to evolve quickly and react to opportunities and feedback and, as a result, are essential.</p><p>To make this even harder there is a really important commitment you&#8217;ll have to secure:<strong> when it comes to sand and pebbles, don&#8217;t lock those projects in ahead of time. </strong>You just want to commit to your rocks and get commitment to an open road of other things that will be picked from the highest-ranking items in the sand and pebble queues when the time comes. This allows you to be reactive and iterate based on what&#8217;s happening. By not planning ahead of time you allow yourself the flexibility to respond to today&#8217;s problems, not yesteryear&#8217;s ideas. </p><div><hr></div><h2><strong>Sustainable jar-filling practices  </strong></h2><h3><strong>Focus on the goal, not the implementation:</strong></h3><p>When someone reads the words that describe or identify your project remember that in those words is coded a set of ideas and that that set of ideas can be different from person to person. This can lead to disconnects about what the project is, what it&#8217;s trying to achieve, and its overall value. This disconnect hurts your ability to get projects done. </p><p>Try to define and identify your projects as much as you can by the goal, problem, or use case that it is setting out to address (vs by a code name or feature/implementation details). Ultimately, the team, through working on the project, may arrive at a completely different solution than what was imagined at the beginning of the project. By avoiding a focus on implementation, you give people the ability to hone in on the why of a project, not the how. This also ensures that anyone who comes upon that project has a higher likelihood of understanding the value it is looking to extract. </p><p></p><h3><strong>Create lots and lots of optionality:</strong></h3><p>The best way to improve your roadmap&#8217;s likelihood of success is to give yourself a lot of optionality with your investments. Optionality (in the roadmap-building context) is the ability to control investments gradually and to decide to quickly double down or cut back when things aren&#8217;t working. </p><p><strong>To put this idea into the context of rocks, pebbles, and sand:</strong> Sand and pebbles give you a lot of optionality because the investment period is shorter, whereas rocks lock you into investments for longer.</p><p>You&#8217;ll find that it is much easier to do multiple pebble and sand projects back-to-back within a certain area than it is to do a rock and follow that up with sand and pebbles. This is because teams lose enthusiasm and momentum the longer they work on something. The joy of shipping is the one thing that can refuel teams in their pursuit of a long-term solution to a problem. As a result, roadmaps that pursue a lot of smaller-scaled iterations have the ability to stay committed to a problem for longer than roadmaps that are looking to solve the same problem with one large release. </p><h4>More is more:</h4><p>It&#8217;s better to do a single project at a time with quality than two with no quality.&nbsp;However, it&#8217;s often better to do 2 projects, each at 80% of their potential, than 1 project at 100% of its potential. Ultimately, you want to increase the volume of shipping until you hit a wall with your team&#8217;s ability to still deliver with quality. Sacrifice perfection (note perfection, NOT quality, as quality is essential) in lieu of velocity and volume. You can always improve upon things gradually but the door often closes if the investment proves to be too long.</p><p>The key to success with this strategy is to be able to tackle multiple items in parallel; the more concurrent initiatives your team can pursue in parallel, the easier it is to slot smaller-sized work. I try to have at minimum two independent projects in progress at all times for any given team; ultimately, this depends on the team, the team size, and your ability to support parallel work with quality. </p><p></p><h3><strong>Break things down:</strong></h3><p>You have to start by decomposing your roadmap before projects start. I&#8217;ve often found that most rocks are just collections of pebbles, and most pebbles can be made into sand. By tying a potential project to a size that best reflects the business&#8217; appetite to solve that problem, you are starting projects with a healthy constraint. This honesty allows teams to accept the limitations of a project and work within those limitations. As a result, any change in the shape of a project is additive.  </p><p>The more you shift things down a tier, the better your odds get as well. Often times downshifting is really just removing the parts of a project that have risk or are less proven and turning those into their own smaller stake endeavors. </p><p>Give yourself and the team the true constraints for a project and you&#8217;ll often find that this will improve the likelihood that that project gets picked up as well as the effectiveness with which a team can solve that particular problem.</p><h4>The mental and emotional value of breaking things down:</h4><p>Roadmaps composed of only large projects are naturally prone to lower vibrations. Large projects tend to be late, large projects tend to require cutting, large projects are left often unfulfilled in order to meet a release date. A large-project-oriented roadmap is a roadmap built through <strong>subtraction</strong>. You are constantly subtracting your way to a solution. Teams come up with complete solutions that are then cut, at best, early in the process and, at worst, close to a release. This is both inefficient as well as draining for teams. Editing is a key part of  the process, and all editing is about sacrificing, but healthy editing feels like getting more for less, whereas unhealthy editing feels like being conned out of something. </p><p>In an environment that is more properly balanced, you can begin shifting towards an&nbsp;<strong>addition-</strong>based roadmap. Here, you&#8217;re often finding ways to start small, and over the course of a project, as a result of enthusiasm, discovery, or any other insight, you end up adding to your work. Addition is a highly enjoyable experience like receiving a gift; you did not expect it, but you&#8217;re glad it came your way. Editing in this context feels like trading for something better and it&#8217;s the most enjoyable kind of editing.  </p><p></p><h3><strong>Adjust the bar to entry to match the investment:</strong></h3><p>Teams often apply consistent rules to every kind of project they pursue. These can be incredibly rigorous validation processes to gut-driven gamble-based approaches. </p><p>The problem with this is not in the process itself but in the fact that a one-size-fits-all mindset creates an implicit bias towards projects of a certain kind. For example, In a highly rigorous scoring culture, only rock projects are pursued because sand and pebble projects are often too fuzzy (or worse their perceived value is too low to warrant all the effort it takes to score). </p><p>To avoid this, I adjust the bar to entry for projects based on size:</p><ul><li><p>Sand projects only require philosophical alignment and some key stakeholders liking the idea.</p></li><li><p>Pebble projects can also get through with just philosophical alignment and stakeholders liking the idea. However, we can ask for scoring for a pebble if there is contention.</p></li><li><p>Rock projects require scoring unless they play a key strategic role for the business or have a really high degree of belief from key stakeholders. </p></li></ul><p>Ultimately, the bar you create is up to you, but the important idea is that you adjust it deliberately to incentivize the kind of balance you&#8217;re seeking.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Gather product feedback at scale ]]></title><description><![CDATA[Building a feedback factory]]></description><link>https://www.rewritetheroadmap.com/p/how-to-gather-product-feedback</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-gather-product-feedback</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sun, 10 Nov 2024 18:42:15 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/25e1e5b0-5602-40a2-ba6e-f273a7337f70_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Strong roadmaps are a result of properly assessing a great deal of options and identifying the ones that best fit your constraints and goals. The more information you have, the more options you&#8217;ll be able to identify in your roadmap. </p><p>The challenge with information is that it (1) requires regular collection, (2) requires proper cleaning to be useful, and (3) both collecting and cleaning are time-consuming.</p><p> This article goes into how to acquire information on a regular basis and how you can attempt to do so efficiently.</p><blockquote><p>&#8220;The universe is only as large as our perception of it. When we cultivate our awareness, we are expanding the universe. This expands the scope, not just of the material at our disposal to create from, but of the life we get to live.&#8221;</p><p>&#8212; Rick Rubin, The Creative Act: A Way of Being</p></blockquote><p></p><div><hr></div><h2>The feedback factory  </h2><p>In most businesses, there is an abundance of customer contact. The reality is that, in many cases, the daily contact your business has with customers covers a very broad and very deep slice of the customer experience and perception of your product. It includes:</p><ul><li><p>Product data that is created by users using the product,</p></li><li><p>Support interactions between your team and customers, </p></li><li><p>Interactions your sales team has with customers, </p></li><li><p>Reviews, NPS surveys, social media interactions, </p></li><li><p>In-person events.</p></li></ul><p>In addition to this contact, there are also specific efforts you can undertake to gather additional information, including customer interviews, surveys, analysis, usability tests, experiments, and gradual feature roll-outs, among other things. </p><p>The main challenge is that many times this rich stream of feedback never makes it to your desk, and as a result, you&#8217;re left with incomplete information and blindspots about your product and business. </p><p>The question is, then, how do you sustainably set your team up to gather information at scale, process that information, and have it feed your decision-making? The answer is in building a feedback factory.</p><div><hr></div><h2>Building the feedback factory</h2><p>A feedback factory is a model for systematizing the routine acquisition and processing of feedback at scale. To build the feedback factory you have to assemble the three core systems and ingrain these systems into your culture. These three systems are:</p><ol><li><p><strong>Collection</strong></p></li><li><p><strong>Cataloguing</strong></p></li><li><p><strong>Socialization</strong></p></li></ol><div><hr></div><h3><br>Collection</h3><p>A key idea behind building a feedback factory is to try to expand the ways you can capture valuable customer input. By tapping into all sources of contact with customers you&#8217;ll be able to collect more feedback than you thought was possible.</p><p>Collection includes any acquisition of information about your customer, business, market, or products, and the storage of this information. There are 4 core mechanics that make up a great collection system:</p><p></p><h4><strong>The feedback portal</strong></h4><p>You want to create a space where both team members and customers can log any feedback or insights. The feedback collection portal setup is simple: you want one or more forms that team members or customers can fill out that links to a database where you store all logged feedback. You then want to permeate the usage of this form to everyone at your company and use it whenever there is applicable customer contact, whether it be used on calls, emails, ticket reviews, or even integrated directly into your products.</p><p>Ultimately, by expanding this approach to every customer contact across your business, you&#8217;ll end up building a small window into those interactions, and through that window will come meaningful nuggets from which you can build projects. This will become the most abundant form of feedback that you have stored. It&#8217;ll be composed of short messages about your product experience but collected at high levels of volume allowing you to effectively quantify and qualify the intensity of problems and opportunities.</p><p></p><h4><strong>Contact with customer-facing team leads</strong></h4><p>Connecting with team leads, I&#8217;ve found, is a great way to get high-level feedback, but it is also how you build a shared commitment to feedback collection. This is because team leads are the ones who can keep individuals outside of your team committed to capturing and logging feedback. </p><p>You&#8217;ll want a weekly or bi-weekly sync dedicated to their domain area where you can learn from their customer contact and understanding of the business line. I&#8217;ve found this recurring cadence strengthens the bond between teams and gives you the ability to quickly course correct if feedback collection is lacking.</p><p></p><h4><strong>Research integration into the software development lifecycle</strong></h4><p>Once you move onto a world where you&#8217;re trying to solve a specific problem you need to collect information in more targeted ways. The key here is integrating gathering feedback and information into the software development process.</p><p>Specifically, it means building the time and appetite to do validation/discovery activities at the beginning, middle, and end of projects. Not all projects need research, and when they do, they might only need it at certain stages. However, without the ability to dig in during a project, you&#8217;ll often find that you&#8217;re missing the necessary information to make correct scoping and problem-solving decisions.</p><p>Practically speaking, this means determining what kind of validation (if any) your approach to a project needs, what information you&#8217;re missing that could enable better decisions, and deploying efforts designed around capturing this information. </p><p>This concept is an article on its own, but some examples of this are interviewing customers before a project starts, doing usability sessions, and running early access programs (to name a few). These exercises can be critical to your decision making process and make up a key part of collecting feedback.</p><p></p><h4><strong>Reviewing data on a regular cadence</strong></h4><p>It&#8217;s easy to think of feedback as verbal customer communication, however, any customer reaction is feedback. I believe data often reflects customer sentiment, preference, concerns, and confusion about the product. Data is a form of non-verbal feedback and is just as important as feedback delivered verbally by customers.</p><p>You want to track key metrics from your product experience on an ongoing basis so that you can develop a more cohesive understanding of the area you manage. I look at certain datasets daily (often every few hours), certain data sets weekly, and certain data sets monthly. By building these regular checking windows you&#8217;ll end up building a sophisticated understanding of the area you manage and will be spurred to undertake investigations that can generate valuable projects. </p><p></p><h4>Helping others understand what feedback is:</h4><p>I&#8217;ve found that people often discount information that is valuable feedback because they don&#8217;t see it as feedback. It sounds obvious, but it&#8217;s worth remembering: how is someone supposed to collect feedback if they don&#8217;t fully know what feedback is? </p><p>I like to describe feedback in the following way:</p><ul><li><p><strong>Feedback is any information pertaining to how customers experience your business. </strong>Fundamentally, any experience that a customer expresses about your product, services, value, and so forth is feedback. Confusion about your product is feedback. Anger, love, and anything in between is feedback. </p></li></ul><p>As you can surmise, almost every single customer interaction contains feedback. This can be overwhelming for people in customer-facing roles, but the more they buy into this idea and the more motivated they are about that information being used, the better off your business will be in terms of its ability to make decisions. </p><div><hr></div><h3>Cataloging </h3><p>Ok, so you have tons of information now. New problem, how do you make that information accessible weeks, months, or even years after it was originally captured? Well, the answer is painful but necessary - you need to catalog information and create and maintain libraries. </p><p>There are four keys to doing this:</p><p></p><h4><strong>Put as much as you can into the feedback portal.</strong> </h4><p>The more you concentrate your knowledge base in a single place the less things are likely to get lost. The feedback portal is one such place and it has the distinct advantage that the information making it into the feedback portal is getting there via form. </p><p>Make sure that your forms are proactively capturing information that can be used to catalog insights. Specifically, ask form fillers to provide user type, feedback type, and product area along with the feedback itself. Doing so will save you hours of feedback re-processing. </p><p></p><h4><strong>Create re-usable knowledge hubs.</strong></h4><p>As you build more and more feedback materials (interviews, write-ups, analysis, etc) the challenge is how to make it intuitive for people to find this information and remember to store any new information in that same place. It is essential for a sustainable feedback factory that the information you&#8217;re collecting does not get scattered. There are a few key ways you can group these materials effectively. </p><ol><li><p><strong>Group by research activity type:</strong> Here, you&#8217;re grouping information by the type of feedback-gathering activity that was used instead of the content area. Think about a hub where you can find sections containing all of your interviews, all of your surveys, all of your usability sessions, etc. </p></li><li><p><strong>Group by specific projects or features:</strong> Here you&#8217;re creating a hubs specific to projects or features. This allows anyone who is looking for information about that specific feature or use case to easily find previous research efforts around that area. </p></li><li><p><strong>Group by user types:</strong> Here&#8217;s you&#8217;re grouping materials and information by specific user types that you work with allowing anyone to search for information by that user type.</p></li></ol><p>Ultimately the type of aggregation you do depends on your business, what information is most valuable, and how you often go about solving problems. That being said, the key is that you are grouping information in some way so that it is easier for others to find it and so that people always know where to store additional information. </p><p></p><h4><strong>Try to consolidate your knowledge to a small number of resources</strong></h4><p>The more spread out your information is the less likely you are to use it and the more likely you are to forget you have it. As much as you can, try to keep a small number of places where information lives as this will help create sustainable patterns around collecting, cataloging, and retrieving information.</p><p><strong>Try to have 3-4 main structures where your feedback lives with key differentiation between them.</strong> In my experience, a model where you have the feedback portal (for short-form insight), knowledge hubs (for theme specific longer form insight), and metric trackers (with data on a specific domain area) has been an effective practice as I always know where to find insights based on the kind of information I&#8217;m looking for.</p><p></p><h4><strong>Routinely clean and reorganize insights</strong></h4><p>Maintenance is an essential component of a productive feedback-collection machine. You need to allocate time on a regular basis to undergo feedback cleaning, reorganization, cataloging, etc efforts. I often schedule deep cleans during slow times during the year and find that this practice yields cleaner spaces but often also yields new insights as I&#8217;m looking at feedback with a fresh perspective.</p><div><hr></div><h3>Socialization</h3><p>The final piece to this system is the socialization of feedback, its use, and embedding it in your culture. This is the heart of the system, and without it, the rest of the feedback factory will fall apart. </p><p>Simply put, the feedback factory exists to help make better product decisions, if it is not clear that this is the case then the factory itself will cease production as individuals will stop being motivated to collect feedback. </p><p>Socialization is the routine broadcasting of the usage of collected feedback to power the roadmap. For me, this works in a few ways:</p><ul><li><p><strong>I routinely make projects in the backlog of the roadmap based on feedback items that were logged and tie the project to those specific items. </strong>This allows others to see feedback be used to foment actual projects and see in a very practical way me making use of their feedback.</p></li><li><p><strong>I</strong> <strong>routinely make use of captured feedback to help drive the product development process and onboard team members onto a project.</strong> This sets up product development teams to expect and seek out feedback to make decisions and allows them to speak about their projects using the language of our customers.</p></li><li><p><strong>I routinely highlight the specific knowledge items that helped drive the execution of the roadmap project as a project comes to a release. </strong>We often follow up with customers and team members directly about how their feedback was used as we deliver a project. Team members and customers alike find this to be valuable and feel that this shows that their opinions are listened to.</p></li><li><p><strong>I</strong> <strong>routinely acknowledge the usefulness of feedback in large company or team-wide meetings.</strong> I specifically highlight the impact certain logged feedback had on our ability to make good decisions, avoid bad ones, and evaluate the full set of possibilities. I also deliberately call out specific individuals who have done a great job at gathering feedback as this (1) rewards them publicly and (2) creates an incentive for others to collect feedback. </p></li></ul><p>By consistently making use of feedback you&#8217;ll find it clearer to define what you need in terms of future feedback as well as motivate those capturing it. The more explicit the connection of feedback to decision-making, the higher the buy-in level for participants in the feedback process.</p><div><hr></div><h2>Parting thoughts:</h2><p>Any roadmap-building process is based on information. The less information you have the weaker your decision set is. The more information you have, the more options are available to you, and as a result, the more likely you are to arrive at the right option. </p><p>To ensure you have information, you need a company-wide buy-in to the importance of collecting feedback. To pull this off, you&#8217;ll need really good working relationships with other teams, and they&#8217;ll need to trust you in your ability to execute and make use of the information they&#8217;re collecting. You&#8217;ll also need to build a system where you can collect and catalog feedback at scale and keep people motivated to support that system. </p><p>In a functioning organization the whole business is bringing you the information you need to make good decisions, leaving you with a better foundation from which to make these decisions and putting the onus on your ability to evaluate and prioritize. This structure has been essential to my roadmap for many years, and hopefully, it will be to yours too.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Run a standup]]></title><description><![CDATA[A guide to making everyone's least favorite meeting not terrible.]]></description><link>https://www.rewritetheroadmap.com/p/how-to-run-a-standup</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-run-a-standup</guid><pubDate>Sat, 12 Oct 2024 19:14:08 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/788444df-470e-412f-b0da-e95a2a9b3d69_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I have run more standups than I can count. For a long time, I ran them pretty poorly. Think eyes glazed over, people listing out tasks with no response, constantly re-asking questions because people weren&#8217;t paying attention, no one wanted to be there, just an overall tough hang.</p><p>I firmly believe that when used correctly, standups help projects stay on track and help identify issues long before they escalate. I&#8217;ve spent a long time trying to figure out with others how to make standups valuable and purposeful. I believe the current formula I use works pretty well, so I want to share it with you. </p><p>This article covers what I believe is the foundation of a productive standup.</p><div><hr></div><h2>The key to a great standup is touchpoint-spotting. </h2><p>Standups should focus on one thing and one thing only: <strong>Identify any necessary touchpoints in the workflow to drive higher release quality.</strong></p><p>By focusing on identifying anything that needs to be discussed to improve the likelihood of a successful release (whether that discussion happens during a standup, in a separate meeting, or asynchronously after a review), the standup becomes a routine checkpoint that enables high-leverage conversations to happen.</p><p>What are those touchpoints? Here&#8217;s a list of frequent touchpoints I&#8217;ve identified during my standups:</p><ul><li><p>Identifying blockers, </p></li><li><p>Requesting feedback, </p></li><li><p>Asking others to review progress, </p></li><li><p>Demo stuff, </p></li><li><p>Identify the need for and schedule pairing, </p></li><li><p>Build alignment on something,</p></li><li><p>Requesting context on something, </p></li><li><p>Align on a deployment/release approach,</p></li><li><p>Align on a testing approach,</p></li><li><p>Align on priorities,</p></li><li><p>Talk over a timeline and the likelihood of meeting it/the need to adjust it,</p></li><li><p>Go over testing results. </p></li><li><p>Asking for help, </p></li><li><p>Anything else where contact helps improve the likelihood of a successful release.</p></li></ul><p>As you can tell from the above Standups are NOT recitations of what you did. if we don&#8217;t have any of the above touchpoints to talk through, then the standup ends then and there. Standups are about necessary touchpoints and necessary touchpoints only, if there&#8217;s no necessary touchpoint then there&#8217;s no need for the standup.</p><p></p><h3>Standup Mechanics:</h3><p>Now, there are a lot of touchpoints listed above, but the reality is that covering those touchpoints in standup is pretty simple. I follow a straightforward model built on a loop between touchpoint-spotting and next-step identification. </p><p><strong>Touchpoint-spotting mechanics:</strong></p><ul><li><p>I&#8217;ll start standup by asking the team if anyone has anything they want to connect on; if they do, we start there (unless I have something I believe is critical). If there&#8217;s nothing, I&#8217;ll start. </p></li><li><p>When I start, I&#8217;ll often pull up our work board so that everyone can focus on a shared visual document of our current work. I&#8217;ll work backward from the work that&#8217;s closest to releasing to the work that&#8217;s furthest from releasing.</p><ul><li><p>I&#8217;ve found this ordering significantly improves a team's release rate. It creates urgency around getting stuff out and establishes an implicit mental priority for work (first what&#8217;s close, then what&#8217;s next). </p></li><li><p>Again, if there&#8217;s something critical, however, I&#8217;ll prioritize that first, or if there&#8217;s a meaningful discussion to be had, I'll prioritize that, too. </p></li></ul></li><li><p>As I cover the board, others and I call out any work items about which we have questions, concerns, or information that needs sharing with others. Most of the time, this leads to deeper conversations. </p></li><li><p>We don&#8217;t go over everything we&#8217;re working on, instead, we deliberately focus on the things that people actually feel the need to talk about. This makes the standup efficient and valuable. Again, if there&#8217;s nothing that qualifies then we end the standup early.</p></li></ul><p><strong>Next step identification mechanics:</strong></p><ul><li><p>Often, as we identify these touchpoints, long discussions ensue. The key to a time-effective standup is identifying the things that should discussed fully during the standup from the things that need additional time outside of it. </p></li><li><p>When starting these discussions, it&#8217;s okay to let things breathe a bit if people are making progress and the progress is valuable and relevant to everyone. Also, letting topics breathe gives you the ability to identify whether we should fully discuss something or if it&#8217;s something to be covered later. This decision is mostly subjective, and it&#8217;s your job to figure out consistently what should be closed during that time slot and what should be moved. </p></li><li><p>If  you realize you&#8217;re running out of time or if there&#8217;s no clear end in sight, politely let people know that this item might just need a follow-up as we&#8217;re running out of time to cover everything. </p></li></ul><p><strong>Note: Keep your average to 15 minutes:</strong></p><p>You can sometimes run over, but you don&#8217;t want standups to become planning meetings, as people will start losing their appetite and endurance for them. You want them to primarily serve as ways to identify touchpoints, resolve the ones that can be quickly addressed in the standup, and schedule and align on any touchpoints that need their own space. </p><p>Standups should take an average of 15 minutes. If they&#8217;re consistently under or over 15 minutes, the team either isn&#8217;t surfacing enough items, or you&#8217;re not effectively identifying and scheduling follow-ups to items that surfaced and are instead just handling them within the meeting. Some days will be light, and some days will be heavy, but in the aggregate, your team should be spending about 15 minutes a day on touchpoint identification and assessment. </p><div><hr></div><h2>Closing Notes:</h2><p>Everyone hates meetings, and rightly so; meetings can often devolve into a weird kind of theatre where everyone pretends that what&#8217;s being talked about matters while everyone knows it doesn&#8217;t. </p><p>Standups are not performative, they are not an exercise in bureaucracy, and they are not needed if they&#8217;re not needed. If there is nothing to talk about, then at best, standups are just a way to see everyone each morning. If there&#8217;s a reason a touchpoint can help make things flow smoothly then it&#8217;s a place where you can have that conversation.</p><p>Standups, like other process tools, can be effective when used correctly and detrimental when not. To make a standup valuable, focus on creating a meeting built around identifying needed touchpoints. By doing this, you may find that the standup can aid flow, create a forum for team members where they can be vocal about projects, and, most importantly, have a tool that can be used to identify if projects are starting to get off track.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Write a weekly report about your roadmap]]></title><description><![CDATA[Weather reports for TLDR.]]></description><link>https://www.rewritetheroadmap.com/p/how-to-write-a-weekly-report</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-write-a-weekly-report</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Tue, 08 Oct 2024 04:01:50 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c1204ce9-da00-4d3a-a0f2-a475d4e605b7_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It&#8217;s not easy to write a weekly report (and even harder to stick to it). It can also feel sort of useless to do it as people often won&#8217;t read what you write. However, when deployed effectively reports can have a number of positive effects including:</p><ul><li><p>Interactions with stakeholders become more productive as stakeholders are consistently well-informed about the state of things.</p></li><li><p>Meeting times and meeting volume are cut down since you need to do less context-building during meetings and instead can focus more on solutions. </p></li><li><p>Stakeholders start addressing issues much earlier than before and voicing their opinions more consistently.</p></li><li><p>You&#8217;ll gain a better understanding of your product and the ebbs and flows of your projects, which will allow you to improve your practices.</p></li></ul><p>This article covers the current model I use (which was workshopped by the PM team at SportsRecruits and through the feedback of many teammates), why I&#8217;ve found it effective, and how I&#8217;ve tried to keep it valuable week after week. </p><div><hr></div><h2>Weather Reports:</h2><p>Weather reports are simple; they&#8217;re an update on how things are going in your domain, clearly highlighting which areas have clear skies and which areas have stormy weather. </p><p>The report is concise and uses clear imagery (through weather emojis) to help others quickly screen and prioritize information.</p><p>By doing this, weather reports allow teams to quickly gather key information about your roadmap, ask questions, dig deeper, and, most importantly, act quickly. </p><h3><strong>Weather Report Dos:</strong></h3><ol><li><p><strong>Focus on the key things:</strong> Highlight the top 5 things on your roadmap that you want stakeholders to be aware of and potentially involved in.</p></li><li><p><strong>Be concise: </strong>The more words, the harder it is to find what matters. Brevity is essential for the report to be read by others.</p></li><li><p><strong>This is where discussions start: </strong>Think of the report as the starting point for discussions, work sessions, and other connection points and not as the endpoint.</p></li><li><p><strong>Emojis for focus:</strong> Use weather emojis at the beginning of each bullet to highlight how you&#8217;re feeling and make your report easy to skim:</p><ol><li><p>&#9728;&#65039;&nbsp;&#8221;Sunny&#8221; This is going great!</p></li><li><p>&#127780;&#65039;&nbsp;&#9925; &#127781;&#65039;&nbsp;&#8221;Partly Cloudy&#8221; There are aspects of this that are going well, and there are aspects that I&#8217;m unsure about because they need definition or have the potential to become issues.</p></li><li><p>&#9729;&#65039;&nbsp;&#8221;Cloudy&#8221; There are a lot of aspects that I&#8217;m unsure about.</p></li><li><p>&#127783;&#65039;&nbsp;&#8221;Rainy&#8221; There are issues here. We should be able to resolve them, but I want to point them out.</p></li><li><p>&#9928;&#65039;&nbsp;&#8221;Storming&#8221; There are some serious issues here. The team is going to need some help to resolve these issues.</p></li><li><p>&#127786;&#65039;&nbsp;&#8221;Tornado&#8221; Oh shit! We need to figure this out.</p></li></ol></li><li><p><strong>Keep a cadence: </strong>Publish these weekly so that everyone builds a routine of checking these for information.</p></li><li><p><strong>Make it interactive: </strong>Include videos, links, and demos to make the team's progress tangible and ideally interactive for anyone wanting to learn more.</p></li><li><p><strong>Ask for feedback: </strong>Make sure to ask for feedback on the value of your reports and what others would like to see in them. This will help you refine your writing over time.<br></p></li></ol><h3><strong>Weather Report Don&#8217;ts:</strong></h3><ol><li><p><strong>Oversell or overpolish your update:</strong> You should report on the ground truth, especially if it&#8217;s challenging, so that others can step in and help. If you hide a problem, it&#8217;s on you if it gets worse.</p></li><li><p><strong>Skip weeks regularly</strong>: The more consistent you are, the likelier issues or opportunities will be spotted. It&#8217;s okay to skip it sometimes if you&#8217;re totally overwhelmed, but the report should come first in most weeks, given its impact on others.</p></li><li><p><strong>Overwrite</strong>. The more information in a report, the more that information turns to noise. There&#8217;s a sweet spot between too much and not enough. You&#8217;re trying to find that sweet spot every week.</p></li><li><p><strong>Overwork and don&#8217;t do it if you and others are not experiencing value</strong>. Ultimately, every workflow that fails fails because it requires too much from people or because there&#8217;s no belief that it is valuable. Pursue reporting with the mindset of creating value but doing so efficiently.</p></li></ol><div><hr></div><h2>Example weather report:</h2><h3>&#127783;&#65039;<strong> Outage related to transcoding bug</strong></h3><ul><li><p><strong>What happened:</strong> </p><ul><li><p>The transcoding queue didn't correctly prioritize jobs for about 3 hours, from 11:30 to 14:30 EST.</p></li><li><p>This led to around 526 video files getting stuck in the queue without finishing, affecting 256 users.</p></li><li><p>This issue was a result of the recent change to the prioritization approach for transcoding.</p></li></ul></li><li><p><strong>What we&#8217;ve done since and what&#8217;s next:</strong></p><ul><li><p>We have since processed these jobs and implemented a failsafe for future failures. </p></li><li><p>However, this failsafe will not work in the long run, so we&#8217;re going to schedule work to better address the issue and its underlying causes.</p></li></ul></li><li><p><strong>Additional Links:</strong></p><ul><li><p>For a technical breakdown of what happened and what the failsafe is see this link: {{link}}<br></p></li></ul></li></ul><h3><strong>&#128640; New data cleanup tools</strong></h3><ul><li><p>The new data cleanup tools are officially out this week! They will help reduce time spent on integration-related issues where user data is misaligned.</p></li><li><p><strong>Additional Links:</strong></p><ul><li><p>Here&#8217;s a guide on how it works: {{link}}</p></li><li><p>Here&#8217;s a demo of it working: {{link}}</p></li></ul><p></p></li></ul><h3>&#127780;&#65039;<strong>&nbsp;Updated user signup flow.</strong></h3><ul><li><p>I'm currently running inventory on the release to determine if there&#8217;s anything left, but, likely, all of the work is officially out of dev and in team testing/ready for QA.</p></li><li><p>We&#8217;ve found some areas of code around the account creation endpoint that need to be refactored, but we&#8217;ve agreed to do this post-release.</p></li><li><p><strong>Additional Links:</strong> </p><ul><li><p>Relevant tickets:<em> </em>{{link}}</p></li><li><p>See designs here: {{link}}</p></li><li><p>Try it out in this demo environment: {{link}}</p><p></p></li></ul></li></ul><h3>&#127774;<strong> New and improved favoriting feature</strong></h3><ul><li><p>This week, we had a design review on the latest progress and discussed the technical approach. The review specifically covered cleaning up the existing system, adding the ability to bulk favorite, and permeating favorites across other features. </p></li><li><p>From here, we&#8217;ll start ironing out the first release and continue to iterate across the different areas of the project. As the initial release definition starts coming together, we should have a much more robust update next week.</p></li><li><p><strong>Additional Links:</strong> </p><ul><li><p>Check out the technical approach doc: {{link}}</p></li><li><p>Check out the current designs here: {{link}}</p></li></ul></li></ul><div><hr></div><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Score up the value of a project]]></title><description><![CDATA[Reach, Conversion Rate, and Confidence level is all you need.]]></description><link>https://www.rewritetheroadmap.com/p/how-to-score-up-the-value-of-a-project</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-score-up-the-value-of-a-project</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sun, 22 Sep 2024 23:56:20 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/aaa3e385-fe9a-4efe-8338-899388ce142e_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When determining what projects should be on the roadmap, it&#8217;s valuable to have some way to quantify your beliefs and, through that quantification, determine whether projects are actually worthy (or just appear to be at face value). </p><p>Scoring is the process through which you can add some rigor to your roadmap building and, as a result, improve your odds of success. Overall, scoring brings a lot of positive dynamics to your team:</p><ol><li><p><strong>It forces you to define value:&nbsp;</strong>Scoring pushes you to describe value in tangible terms. This adds rigor to the way projects are selected and how they extract value. Over time, this habit will improve your ability to identify high-upside projects, and it might also spread to others across your organization.</p></li><li><p><strong>It helps shape projects in ways that improve their odds of success: </strong>By scoring projects, you&#8217;ll more consistently identify why some projects are unlikely to succeed and mitigate those risks ahead of time.</p></li><li><p><strong>It creates clarity and alignment about the value you&#8217;re trying to unlock:</strong> Scoring allows everyone to understand the value to be extracted by doing the project by clearly describing that value. By understanding this, others can better form opinions about how to best tackle the project and debate on whether that value is worth the investment in the first place.</p></li></ol><p>With all this in mind, I want to propose a simple exercise you can use to score projects quickly.</p><div><hr></div><h2>Value = Reach  * Conversion Rate * Confidence Level</h2><p>A project&#8217;s value is almost always defined by the combination of two factors:</p><ol><li><p>The addressable market of your solution (Reach)</p></li><li><p>The conversion rate of that market to your solution (Conversion Rate)</p></li></ol><p>Ultimately, these two variables are kind of all you need to score any project. There are many frameworks out there (most of which work), however, I&#8217;ve found that with these two variable I have 90% of what I need to pick projects with a relative degree of accuracy. </p><p>The final variable is Confidence Level, which, simply put, mitigates opinion and ego. By putting reach, conversion rate, and confidence level together you have something that easily describes the value of any given project (as long as that project is defined by a metric of some kind).</p><p>Let&#8217;s dig into these variables a bit more closely:</p><p></p><h2>Reach</h2><ul><li><p>Reach describes the market/opportunity size of your solution. Reach is not tied to impact, instead, it looks to quantify the number of addressable entities for your project.</p></li><li><p>Reach is measured in the number of people, events or time, impacted for a specific period. The time period is variable and generally depends on your company&#8217;s payback period for initiatives.</p></li></ul><ul><li><p>Sometimes projects have reach outside of a single dimension, for example some projects reduce retention while also improving organizational efficiency. I often score projects against different dimensions to put forth the clearest argument for its value.</p></li></ul><p></p><h2>Conversion rate:</h2><ul><li><p>Conversion rate measures the % of your reach that the solution will actually solve for. Conversion rate (or win rate) is essentially your gut feeling of how effective your solution will be.</p></li></ul><ul><li><p>Discussions about conversion rates often lead to realizations about your ability, or lack thereof, to extract value.</p><ul><li><p>This ends up being the heart of value discussions as reach tends to be more easily agreed upon. If everyone agrees on the reach, conversion rate is then a tangible way to talk through impact for a project and align expectations.<strong>  </strong></p></li></ul></li><li><p>Conversion rate is also a tangible metric which can be figured out (through experimentation, research, etc) which will often allow you to further advance the likelihood of a project even if it&#8217;s rejected upfront.  </p></li></ul><p></p><h2>Confidence level:</h2><ul><li><p>Confidence level is an easy way to ensure that opinions are weighted against good practices. </p><ul><li><p>The confidence level is set based on if you&#8217;ve undergone specific activities that derisk the project and, as a result, are more confident that the project&#8217;s value is what you say it is. </p></li></ul></li><li><p>These are the options available for Confidence level:</p><ul><li><p>Supported by Experiment / Existing Feature = 90%</p></li><li><p>Supported by Feedback / Product Data = 67%</p></li><li><p>Supported by Market Analysis = 45%</p></li><li><p>Supported by Vision Alignment = 23%</p></li></ul></li></ul><ul><li><p>As you&#8217;ve probably surmised from the above, the project&#8217;s value expectation suffers the more fuzzy its foundation of insight is. This creates an incentive to derisk projects by pursuing activities that naturally increase confidence in your problem and its execution. </p></li><li><p>Confidence level can be an especially good tool for getting past situations where highest paid person&#8216;s opinion wins. Projects that are only backed by opinion will naturally score worse and be harder to defend whereas projects that have been derisked in some way will score best.</p></li></ul><p></p><div><hr></div><h2>Example Scoring:</h2><p>Let&#8217;s put these variables into practice. Below are 3 different projects scored using the above variables:</p><h4>1. Signup Optimization Project:</h4><p>Let&#8217;s say you&#8217;re thinking about a project related to optimizing a key step within your signup process.</p><ul><li><p><strong>Reach: </strong></p><ul><li><p>In this situation, reach is the total number of people who attempt to sign up but don&#8217;t complete it within a time period. </p></li><li><p>More specifically, let&#8217;s say it&#8217;s the number of people who bounce during a specific step that you&#8217;ve identified is problematic. Let&#8217;s say per year it&#8217;s 20,000 people.</p><ul><li><p><strong>Reach Used: </strong>20,000 over 12 months.</p></li></ul></li></ul></li><li><p><strong>Conversion Rate: </strong></p><ul><li><p>In this situation, the conversion rate is the expected percentage of people who attempt to sign up but don&#8217;t complete it, but who would now complete it because that step&#8217;s friction is removed.</p><ul><li><p>Let&#8217;s say that you believe that you can convert about 50% of the people you&#8217;ve identified in reach by addressing the issues with that step.</p></li><li><p><strong>Conversion Rate Used:</strong>  50%.</p></li></ul></li></ul></li><li><p><strong>Confidence Level: </strong></p><ul><li><p>Let&#8217;s say you&#8217;ve done a funnel analysis of signup and drop-off points, you&#8217;ve localized the problem with data, and you have a sense of why it happens but haven&#8217;t really tested it or talked to customers).</p></li><li><p><strong>Confidence Level Used: </strong>In this situation, you&#8217;re using the 67% confidence level as your score is supported by product data</p></li></ul></li><li><p><strong>Final Score:</strong></p><ul><li><p>20,000 people * 50% of those converting via signup * 67% confidence level = 6,700</p></li><li><p>If you pursued this project you&#8217;d generate 6,700 incremental leads signing up every year.</p></li></ul></li></ul><p></p><h4>2. Value Expanding Feature Project:</h4><p>Let&#8217;s say you&#8217;re building a feature that expands the value of your business by introducing a way to solve a problem you haven&#8217;t addressed. </p><ul><li><p><strong>Reach: </strong>In this situation, reach is the total number of existing or new customers that could potentially become customers of this new solution over the course of 1 year.</p><ul><li><p>Let&#8217;s say it&#8217;s 500 new customers at an average ARR value of $10,000. 500*10,000 = 5M.</p></li><li><p><strong>Reach Used: </strong>$5 Million in ARR signed up within 1 year.</p></li></ul></li><li><p><strong>Conversion Rate: </strong>In this situation,<strong> </strong>conversion<strong> </strong>rate is % of existing or new customers that will become customers of this new solution.</p><ul><li><p>Let&#8217;s say you believe that within a year, you can convert 10% of those customers to this new product.</p></li><li><p><strong>Conversion Rate Used: </strong>10%</p></li></ul></li><li><p><strong>Confidence Level: </strong></p><ul><li><p>Let&#8217;s say you&#8217;ve looked at the market and have a sense of competitors doing this and that it is desired but you&#8217;ve never connected the dots in terms of if your customers would actually care to have this).</p></li><li><p><strong>Confidence Level Used: </strong>In this situation, you&#8217;re using the 45% confidence<strong> </strong>level as your score is only supported by market analysis</p></li></ul></li><li><p><strong>Final Score:</strong></p><ul><li><p>5,000,000 * 10% of those converting via signup * 45% confidence level = 225,000</p></li><li><p>If you built this new solution you&#8217;d be able to generate $225k in additional ARR in the first year of launching it.</p></li></ul></li></ul><p></p><h4>3. Internal Efficiency Feature Project:</h4><p>Let&#8217;s say you&#8217;re building a feature that reduces the duration of a certain internal operations process. </p><ul><li><p><strong>Reach: </strong></p><ul><li><p>In this situation, reach is the total number of hours that that internal process currently takes to perform per quarter. Let&#8217;s say it&#8217;s 300 hours quarterly.</p></li><li><p><strong>Reach Used: </strong>300 hours per quarter.</p></li></ul></li><li><p><strong>Conversion Rate: </strong></p><ul><li><p>In this situation, conversion<strong> </strong>rate is the expected % of hours that are removed as a result of the solution. </p></li><li><p>Let&#8217;s say you believe you can get 70% of those hours removed with this solution.</p></li><li><p><strong>Conversion Rate Used: </strong>70%</p></li></ul></li><li><p><strong>Confidence Level: </strong></p><ul><li><p>Let&#8217;s say you&#8217;ve done the internal process yourself and hacked at a new way to do it that the solution will be based on).</p></li><li><p><strong>Confidence Level Used:</strong> In this situation, you&#8217;re using the 90% confidence<strong> </strong>level as your score is only supported by market analysis.</p></li></ul></li><li><p><strong>Final Score:</strong></p><ul><li><p>300 * 70% of those converting via signup * 90% confidence level = 189 </p></li><li><p>If you pursued this project you&#8217;d remove 189 hours of operations work per quarter.</p></li></ul></li></ul><p></p><div><hr></div><h2><br>Scoring Related Notes:</h2><p>There are a lot of important ideas that pertain to scoring (too many to share in one post). The following are some of the most important learnings I felt were worth including in this article:</p><ul><li><p><strong>Scoring doesn&#8217;t replace doing PM work.</strong></p><ul><li><p>Scoring can easily become a crutch for poor PM' ing, as it can give a false sense of rigor and sophistication. Ultimately, someone who has a clear vision, business sense, a deep understanding of customers, and a strong team can achieve an effective roadmap without ever scoring a project. </p></li><li><p>At different times I&#8217;ve encouraged my team to drop the rigor and trust their beliefs to mitigate for this risk. In many cases, if there&#8217;s a deep belief in the project built on a strong knowledge base (that was primarily developed through customer contact), then you should probably just do the project regardless of how it scores.</p></li></ul></li><li><p><strong>Scoring isn&#8217;t the what determines which projects get done and which ones don&#8217;t.</strong></p><ul><li><p>I originally believed that the purpose of scoring was to decide which projects should get done and which ones shouldn&#8217;t. </p></li><li><p>I soon realized that when you pursue a highly regimented scoring approach all you&#8217;re left with is measurable work and unfortunately, a lot of really important work (brand sentiment, experience quality, system health, debt creation/mitigation to name a few) is immeasurable or hard to measure. </p></li><li><p>A project&#8217;s score within a framework is not the deciding element for if it should be done, it is a data point in your consideration for doing it and an entry point for everyone else to help validate your project&#8217;s value. </p></li></ul></li><li><p><strong>Cost isn&#8217;t a part of this model because it does not impact value, it instead impacts return on investment.</strong> </p><ul><li><p>The model does not go into timing or cost as those are properties that define return on investment. Conversation around those variables is important but when evaluating a project I prefer to start with a pure value assessment since doing investment management analysis for a project that has no upside, and as a result that you&#8217;ll never do, is a waste of time.</p></li><li><p>Once a project is scored  you should be able to have thorough conversations about the worth and timing of projects and the necessary adjustments to extract value.</p></li></ul></li><li><p><strong>Standardization can be counterproductive:</strong></p><ul><li><p>A lot of frameworks try to compare projects with totally different goals against one another by consolidating value into an abstract score (if no one knows what the number means then it&#8217;s not a useful number). </p></li><li><p>I&#8217;ve found this to be detrimental to my ability to identify value and pick the right projects. It&#8217;s really hard to compare apples to oranges; if you do, all you&#8217;re doing is following whatever natural bias you have towards either apples or oranges.</p></li><li><p>I prefer just talking about value and making decisions on what the business values more over trying to translate that value into some opaque metric and building a dependency on that metric to make decisions.</p></li></ul></li><li><p><strong>Don&#8217;t score for the sake of scoring.</strong></p><ul><li><p>Not every project needs scoring and in fact scoring can be highly detrimental to your ability to get certain projects through even when they&#8217;re overwhelmingly the right decision. </p></li><li><p>Scoring is effective for large projects with clear tangible value. When you have strategic partnership, key customer pain points, small iterations, or other hard to measure opportunities you should likely revisit your need to score as the scoring framework might unfairly punish these fuzzier items.</p></li><li><p>For projects that don&#8217;t score well but that I&#8217;m still going to do, I won&#8217;t have an actual metric value for doing it because that wasn&#8217;t why the project made it to the roadmap. It made it to the roadmap because of my belief in its value and I&#8217;d rather express that value the best I can rather than make up a value number I don&#8217;t believe in.</p></li></ul></li><li><p><strong>Re-Scoring is as valuable as scoring.</strong></p><ul><li><p>Once projects are done do your best to re-score them with the actual results. This can be an incredibly insightful process for you and your team as you&#8217;ll learn more about the nature of your product and roadmap and why things work/don&#8217;t work. </p></li><li><p>This type of review will help you the types of projects you go for and sharpen your instincts for which projects work and which ones don&#8217;t.</p></li></ul><p></p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Deal with product fires]]></title><description><![CDATA[Stop. Drop. Roll. Then fix the bug.]]></description><link>https://www.rewritetheroadmap.com/p/how-to-deal-with-fires</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-deal-with-fires</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Mon, 02 Sep 2024 10:58:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8029bf5c-cc44-49f2-940c-03d88699951d_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Fires are situations where an issue or bug has escalated to the point of major impact on the business. In startups, fires occur often and can have crucial ramifications for the business. As a result, firefighting is an easy way for PMs to quickly build equity with the business and become a trusted team member.</p><p>Firefighting comes down to 4 key skills:</p><ol><li><p>Diagnosing</p></li><li><p>Mitigating</p></li><li><p>Communicating</p></li><li><p>Resolving</p></li></ol><p>This article goes into firefighting and how to consistently diagnose, mitigate, communicate, and resolve issues.</p><div><hr></div><h2>Diagnosing</h2><p>Diagnosing comes down to 2 key ideas: having the right people and environment in place to diagnose and following a structured path that has a high likelihood of resulting in a correct diagnosis.</p><p></p><h3><strong>Environment and People:</strong></h3><p>Diagnosis can be the most challenging step as the anxiety and stress from the issue can get in the way of critical thinking. As much as you can create an environment for thoughtful problem-solving this will significantly improve the odds of success and the speed at which you find a solution. Set yourself and others up for success by using calm tones when discussing the issue, remind and reassure everyone that you&#8217;ll arrive at a solution, and, when applicable, use humor to defuse tension and contextualize the issue.</p><p>Most importantly, ensure that the right stakeholders are involved with the right level of involvement. Ensure that you have people actively working on the issue who have the ability to solve the issue and that you&#8217;re in contact with teams that are dealing with the backlash of the issue. If you don&#8217;t know who should be working on the issue escalate the question to someone who can make that determination.</p><p>I&#8217;ve found that the single most valuable thing a PM brings to diagnosing is focus. It is hard to stay focused on doing the right things due to stress, incomplete information, desire to know more, and uncertainty. As a PM your job is to constantly align the team on identifying the best next step, pursuing that step, and reassessing the path. If the team is consistently doing this you&#8217;ll effectively diagnose and resolve any issues you face. </p><p></p><h3><strong>Path to Diagnosis:</strong></h3><p>I&#8217;ve found that the most reliable path to a diagnosis is to create a timeline of events that can effectively explain why the issue occurred and why we&#8217;re seeing the current impact. Timeline building is a great way to ensure that you don&#8217;t miss any key variables in the diagnosis and once you have that narrative defined it&#8217;ll be much easier to identify the solution. The narrative should answer a few questions:</p><ol><li><p>What is happening and where is it happening?</p></li><li><p>How does it happen?</p></li><li><p>Who is affected now and who might be affected later?</p></li><li><p>Why did it start happening and why didn&#8217;t we notice it until now?<br></p></li></ol><p>When going through this process you should be capable of proving your narrative is right by checking your assumptions against existing cases and finding consistency across every case you see. If every impacted user or instance is consistent with your assumptions then you likely have proven yourself right. Finally, you need to ask yourself if you were wrong why would it be? This extra step of challenging your assumption can be key to expanding your understanding of your problem and identifying missing pieces to your problem.</p><p>With this information, you can work with technical and operations teams to correctly define the solution for the issue and the follow-through.</p><div><hr></div><h2>Mitigating  </h2><p>Mitigating is about reducing the breadth and depth of impact of the fire, in some cases, you can skip this step altogether and just focus on solving the issue. Mitigation is valuable when fully solving the issue will take too long and where stopping the bleeding can help buy you time and focus. If you can&#8217;t get affected users back to normal ask yourself if you can prevent further users from being exposed as this often can make a huge difference in the total damage of the issue. </p><p>Mitigation should not be done by default and should instead be a step decided upon after analysis. I&#8217;ve found the following rules as a good way to decide if it&#8217;s worth investing:</p><ul><li><p><strong>Do</strong> <strong>invest</strong> if it aligns with the ultimate solution.</p></li><li><p><strong>Do</strong> <strong>invest</strong> if it significantly improves the quality of life for those impacted or reduces further damage.</p></li><li><p><strong>Don&#8217;t</strong> <strong>invest</strong> if it significantly gets in the way of solving the issue. </p></li><li><p><strong>Don&#8217;t invest </strong>if there&#8217;s any risk the mitigation compounds the issue.</p></li></ul><p>If you are pursuing mitigation ensure that your stakeholders understand the nature of this action and that they don&#8217;t interpret this step as actually solving the problem. </p><div><hr></div><h2>Communicating</h2><p>Throughout the issue, you need to keep people up to speed on what&#8217;s happening so that they can tackle their end of the problem-solving. Great communication can often be a game-changer during fires. There are three key ideas to great communication:</p><ol><li><p>Communicate what you know</p></li><li><p>Communicate what you&#8217;re doing</p></li><li><p>Communicate regularly</p></li></ol><p><strong>Note: </strong>When communicating you want to avoid misleading teams, especially if they&#8217;re going to bring information to customers, as misinformation can often make the problem worse. You should not make assumptions or guesses at the cause or impact of the issue if you do not understand it as it may create further confusion and stress, but you can provide not fully validated information if you properly caveat it and the likelihood that you&#8217;re right is &gt;=80% as this can help teams navigate the issue.  </p><p>Deliver your updates in structured messages calling out what&#8217;s happening, who&#8217;s impacted, what&#8217;s been done to address it, what&#8217;s next, and key things that need to be done (if applicable). </p><p>You want to communicate what you reliably know as well as what the team&#8217;s next steps are. Specifically, you want to make sure the following things are understood:</p><ul><li><p>The reach and depth of the issue.</p></li><li><p>Who else may be impacted in the future.</p></li><li><p>How you&#8217;re mitigating that impact.</p></li><li><p>What the team is doing next to mitigate or resolve the issue fully (and ideally what the perceived timeline of that is). </p></li></ul><p>Ultimately you&#8217;re trying to ensure that everyone is up to speed as close to the real-time as possible but you should always make resolving the issue your first priority. When it comes to how often to share updates shoot for every 30-60 minutes during the early stages of an issue as this will: 1. build trust across the group that you will keep everyone up-to-speed as needed (and as a result minimize distractions) and 2. empower people to properly follow-up with customers/stakeholders.</p><div><hr></div><h2>Resolving</h2><p>When it comes to resolving the issue, the key thing a PM can do is provide the business and customer context so that the right solution is pursued. In any resolution decision-making process options will be presented and these options will vary in their user impact and costs. By understanding your business and customers you can help teams make the right decision for how to solve the issue.</p><p>When looking for the ideal solution remember that:</p><ul><li><p>The ideal solution to any issue will remove the problem entirely for the affected users. </p><ul><li><p>It will often do this while not creating a secondary more significant problem although sometimes it may create a less significant one.</p></li></ul></li><li><p>Oftentimes, to fully resolve the issue a more significant solution, system refactoring, or revamping will be needed.</p><ul><li><p>It&#8217;s important to delineate what the today part of the solution is and what the tomorrow part of the solution is so that everyone is aligned on how to proceed.</p></li></ul></li><li><p>When pursuing a solution, do NOT attempt to release an untested, or minimally tested, solution, especially under heavy fatigue. </p><ul><li><p>The single worst thing you can do in a fire situation is attempt a solution you haven&#8217;t fully vetted as this can, and will, often make things worse. The wrong solution can easily turn a moderate problem into a huge problem and it can easily reset your problem solving furthering the damage and frustration.</p></li></ul></li></ul><p>As the technical solution is being implemented, work with ops teams on the customer follow-through for the solution. Oftentimes, the lengths you and the team go to to put back what was broken and make amends with customers can significantly improve the standing of your business and the relationship with your customers. Nail down the customer communication and any operational follow-through whether it&#8217;s cleanup, reimbursement, or committing to preventative work.</p><div><hr></div><h2>Post Fire</h2><p>Once an issue is fully resolved invest in doing a retrospective about what things went wrong and how can you prevent the same thing from happening again. Identify in these sessions the work that needs to be done to minimize future issues as well as the protocol to be followed. </p><p>Whenever possible put monitoring or tracking in place for the issue area as this will help you identify and resolve issues faster in the future. Strong monitoring is the best way to prevent issues from escalating and every fire is an opportunity to improve how you&#8217;re tackling prevention.</p><p>As mentioned before, follow-up work will likely be necessary to prevent the issue from happening again. Use the momentum of the fire to tackle any system issues that may lead to further fires. Fires create momentum to improve the overall health and safety of your system, use that momentum to get the necessary projects in the roadmap. </p><p>Overall, firefighting is a stressful part of PM&#8217;ing but it can also be an immensely powerful tool for building reliability as a team and in your system. Take each fire as an opportunity to learn more about your system, your team, and yourself and you&#8217;ll find that these learnings will help you become a better PM.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Thoughts on: Operations based solutions vs software based solutions]]></title><description><![CDATA[Software is cool, but duct tape is cheaper.]]></description><link>https://www.rewritetheroadmap.com/p/ops-vs-software</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/ops-vs-software</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sat, 23 Mar 2024 15:18:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6b1c6c18-78a4-4a8e-a404-4a4f54ac7ec4_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There is a natural tendency to <strong>overestimate software-driven solutions </strong>and <strong>underestimate operations-driven solutions.</strong> This bias leads to an over-investment in software and an under-investment in communication, process, experimentation, and deal-making.</p><p><strong>Operations-based solutions tend to maximize outcomes within the constraints of today</strong>. A few examples:</p><ul><li><p>The fastest way to improved internal efficiency is not to build a new internal tool but instead to cut scope and do less.</p></li><li><p>The fastest way to higher adoption is not to build a new onboarding feature but to individually connect with customers, understand their pain points, and onboard them. </p></li><li><p>The fastest way to drive more conversions is not to build an extra paid feature but to nail down communications, sales processes, pricing, or even run promotions.</p><p></p></li></ul><p><strong>Software, on the other hand, tends to break current constraints and expands the business&#8217; problem-solving range and depth</strong>:</p><ul><li><p>In the former scenario, a new internal efficiency tool could wipe an efficiency problem away completely by automating that workflow. </p></li><li><p>A new onboarding feature could establish a new floor for adoption.</p></li><li><p>A new paid feature could raise the conversion rates across the user base. </p><p></p></li></ul><p>When businesses put all their eggs in the software basket they inevitably leave a lot of value on the table because every problem is weeks or months away from being solved and you can only really solve a few problems at a time. When businesses put all their eggs in the ops basket they can easily develop a short-sighted mindset present and create debt that&#8217;ll be hard to shake in the future.</p><p>It is true that operations solutions can break constraints and software solutions can optimize within constraints, signing a massive partnership, for example, is a transformative, ceiling-raising operations solution. However, this general mental model can help evaluate the right short-term and long-term approach to a problem.  </p><p></p><div><hr></div><p></p><h2>Properties of each solution type: <br></h2><h4><strong>Software-based solutions</strong> </h4><ul><li><p>Require a high degree of investment in terms of development but also in terms of integrating the new feature successfully into the business (building workflows, communication, support, sales, etc). </p></li><li><p>Carry long-term maintenance/support/optimization costs that are often overlooked.</p></li><li><p>Because of this overall cost, they have to deliver substantial amounts of value to achieve positive ROI. The investment cost is also almost always paid upfront. </p></li><li><p>Given the investment scope, businesses can also only afford to pursue a few solutions like this at a time.</p></li><li><p>Iteration cycles are slower as each cycle requires additional development which makes it harder to obtain results and validate the solution.</p></li><li><p>The total possible value for a solution of this kind is often very high since software can make <strong>almost </strong>anything possible.<br></p></li></ul><h4><strong>Operations-based solutions</strong></h4><ul><li><p>Generally have much lower upfront costs. </p></li><li><p>Are easier to manage, support, and discontinue long-term. Businesses can quickly abandon ops processes if these aren&#8217;t working. </p></li><li><p>This low-cost, low-commitment nature makes them easier to sign off on and leverage to extract value quickly. </p></li><li><p>Businesses can generally pursue a great number of ops solutions at any given time (although this capability creates coordination and cohesion challenges). </p></li><li><p>Iteration cycles are faster since each cycle has a low investment cost which makes it easier to obtain results and validate the solution.</p></li><li><p>The total possible value for a solution of this kind is limited by the constraints of the people, time, capabilities, and capital available to the business.  <br></p></li></ul><p>Given the difference in mechanics, each solution type is well-suited for certain problems but ill-suited for others. A few questions can be helpful to determine if ops or software is the right path for a problem: </p><ul><li><p>Could we reasonably solve this without software? Would that solution be faster? Could we unlock 60-80% of the value with that approach? Is this something that we&#8217;re comfortable not building into the product?</p><ul><li><p>If all are yes then an ops solution could reasonably solve the problem in a comparable way to software making it a good option. </p></li><li><p>If only some are yes then ops might still be well suited but ops is likely an intermediary solution or a precursor to a long-term software solution. </p></li><li><p>If all are no then this is very likely a software-shaped problem needing a software-based solution.</p></li></ul></li></ul><p></p><div><hr></div><p></p><h3>Software-based ops (hybrid model)</h3><p>A new hybrid approach has become more prevalent with the rise of low-code platforms and LLMs. In many cases, hybrid solutions are fundamentally shaped like ops solutions in terms of cost and cycle speed but have payoffs that are more software-shaped. This makes them very well suited for a great deal of problems. </p><p>In particular, I&#8217;ve found these to be well-shaped for areas such as:</p><ul><li><p>Internal efficiency (replacing pure software internal tooling)</p></li><li><p>Experimentation (expanding the possible experiments you can run), </p></li><li><p>Replacing high manual labor-based features (replacing things like customizable reports and assets, personalized email campaigns, and marketing page enablement). </p></li></ul><p></p><div><hr></div><h2>How this concept affects PMs:</h2><p>In many orgs, teams will become software-centric or ops-centric. I&#8217;ve found that it can often be the PM&#8217;s job to help teams delineate when one or the other is best suited. In creating this open-mindedness PMs and teams can become more effective and nimble in their problem-solving and collaboration.</p><p>Effective product management at a small company is often about maximizing the volume and quality of ops bets and software bets the business is making. Indexing too much on one or the other will often lead to disappointment. Software bets take too long to pay out and operations solutions can&#8217;t deliver the long-term upside the business needs to continue growing. By effectively leveraging both approaches businesses can effectively maximize opportunities short term while setting themselves up for high upside opportunities long term. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to: Do a design review as a product manager ]]></title><description><![CDATA[Win over your designers and avoid becoming that PM...]]></description><link>https://www.rewritetheroadmap.com/p/how-to-do-a-design-review</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-do-a-design-review</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sun, 21 Jan 2024 18:18:43 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/58a8a5d0-c255-456d-880e-b2808fa3901c_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Where PMs fail when reviewing designs</h2><p>Many PMs will approach reviews with a creator mindset and quickly spit out many potential UX or UI updates, in some cases even feature ideas, without acknowledging the thinking and approach laid out by the designer. In contrast, some PMs do not go deep enough during design reviews leading to designs that can often lack the alignment and depth to be effective when built.</p><p>It is key to acknowledge that the designer is the driver of the design and it is through their understanding of the business case, customer, and context that the team will succeed in finding the right experience. Collaboration is necessary, of course, but the designer is ultimately running point on the design.  The PM, however, is running point on ensuring the design drives value and delivers on the intended use case by ensuring a use case and value-focused mindset during the process. Through this combination of roles, the discovery and design process finds the necessary pressure, trust, and collaboration to deliver effective design solutions. </p><div><hr></div><h2>The PM&#8217;s role in a design review:</h2><p>The PM&#8217;s job is to ensure that the conditions are in place for the designer to succeed. <strong>Simply put, the goal is to sharpen design thinking by challenging, clarifying, and contextualizing where applicable</strong>. The PM should bring a clear and in-depth understanding of the customer, use case, market, and business implications as well as an open mind and understanding of basic design concepts. As a reviewer, the PM is seeking, throughout the review, to:</p><ol><li><p>Clearly understand the laid-out approach</p></li><li><p>Assess the design approach in its ability to solve the use case</p></li><li><p>Correct any misconceptions about the business context</p></li><li><p>Challenge the approach and identify risk areas/complexity that need to be mitigated</p></li><li><p>Ensure that the design approach is aligned with the business case, the philosophical direction of the product, and the end user.</p></li><li><p>Help refine the key follow-up questions/areas to be investigated by the designer.</p><p></p></li></ol><p>By doing the above things you&#8217;ll become a more effective partner for the designer because you&#8217;ll be helping the designer find the path and enriching their mental model for the solution. The designer is the driver, only they can get you to the outcome the team needs, but that doesn&#8217;t mean they won&#8217;t get lost - this is where the PM can provide the most value, ensuring that we can find the way. </p><p></p><h4><strong>When do things enter the PM domain?</strong></h4><p>I think there are a few places where one can draw the line for where the PM can, and should, jump in:</p><ul><li><p><strong>The investment cost line:</strong> </p><ul><li><p>When a path&#8217;s cost exceeds the acceptable investment level for the problem it&#8217;s partly the PM&#8217;s job to help the team understand a threshold has been reached and the solution needs to be downscaled.</p></li><li><p>It&#8217;s important with this type of intervention, to try to understand why the designer ended up where they are. Are they seeing something you&#8217;re not, or is there a misalignment in how certain variables are weighed? </p></li><li><p>This is most effectively done with team input and understanding, which requires nailing the exact reasoning for why it would be over-investing. </p></li></ul></li><li><p><strong>The business requirement line: </strong></p><ul><li><p>When a path is solving alternative business requirements or business requirements that don&#8217;t exist, or not solving the core business requirements at all, then a PM can step in and pull back. </p></li><li><p>The PM is responsible for the outcome being driven and (ideally, always) has the most use case knowledge. This makes them the most capable actor to re-direct focus. </p></li><li><p>Again, here it&#8217;s important to try to understand why the designer ended up where they are. </p></li></ul></li><li><p><strong>The decision needs to be made line:</strong></p><ul><li><p>When an area of the project becomes blocked there&#8217;s no agreed-upon solution, and one can&#8217;t be arrived at through research/testing/validation, it&#8217;s the PM&#8217;s job to make the decision and take accountability for its outcomes.</p></li><li><p>In many projects, this won&#8217;t be required, but when it is required, it can be one of the more important things the PM can do as it ensures the project continues moving forward. </p></li></ul></li><li><p><strong>The work is not good enough line:</strong></p><ul><li><p>If you&#8217;ve spent the time to fully understand the laid-out approach and at the end feel like it&#8217;s not good enough and is missing elements that have to be there for launch, it is in your domain to course correct and push the designer to dig deeper. </p></li><li><p>This can be the hardest intervention to make, but when it happens it&#8217;s the most important intervention because it ensures that you don&#8217;t put something out there that fails to deliver value.</p></li></ul></li></ul><div><hr></div><h2>How to approach a design review:</h2><ul><li><p><strong>Ask yourself if you&#8217;re at the beginning or finishing phase.</strong></p><ul><li><p>Discovery tends to be in either the beginning or the finishing phase. During the beginning phase, you&#8217;re trying to nail down the philosophy and requirements for your solution. The beginning is open and benefits from gathering information, talking to customers, experimenting, and having long philosophical conversations. It&#8217;s the going broad stage of the work. </p></li><li><p>During the finishing phase, you&#8217;re trying to move the work towards a buildable place. The finishing stage is focused on closing out open questions, focusing on the core value to be delivered, editing down to the essentials, and ensuring what we end up building nails the problem/philosophy/value we set out to achieve. </p></li><li><p>It&#8217;s important to always have a strong sense of if you&#8217;re beginning or finishing and adjust your approach accordingly. If you don&#8217;t know where you are in the design process, you might end up investing in processes that aren&#8217;t necessary and the process will lose momentum and focus. </p></li></ul></li><li><p><strong>During the review, start by understanding the state of the design.</strong>  </p><ul><li><p>It&#8217;s very easy to get lost in a misunderstanding about the level of definition something has. Understand the progress level of the work so that you don&#8217;t spend time on the things that they haven&#8217;t thought much about. Focus on the things they spent the most time on in the review, or if they&#8217;re going the wrong way, focus on clarifying the focus/understanding the misunderstanding.</p></li><li><p>Ask the designer things like </p><ul><li><p>&#8220;What things should I pay attention to and which ones aren&#8217;t ready?&#8221;, </p></li><li><p>&#8220;Is the work still very conceptual, or more evolved&#8221;, </p></li><li><p>&#8220;Should I pay any attention to the UI or just the experience?&#8221;</p></li></ul></li></ul></li><li><p><strong>Then dive deep into understanding the design. Hold your thoughts until the review is finished.</strong></p><ul><li><p>Spend as much time understanding the approach as possible before making any pronouncements. Try to understand why the designer made each choice and how each choice works together to create a mental model for the solution. <strong>This should be 1/2 of the review with the second half being a discussion. </strong></p></li><li><p>During the review emphasize that you&#8217;re trying to understand the approach fully and that you&#8217;ll be asking many questions, that these will not be judgmental, just information gathering. For understanding, oftentimes intense questioning is necessary, this can make designers feel challenged. You want to mitigate this going in.</p></li><li><p>In terms of questioning, focus on understanding the designer&#8217;s decision-making process at great depth, look at each frame and each experience step, and ask the designer things like:</p><ul><li><p>&#8220;Walk me through how you got to this implementation, what considerations did you have, and why this felt like the best decision?&#8221;</p></li><li><p>&#8220;What led you to place this step here specifically? How does this placement affect the rest of the experience?&#8221;</p></li><li><p>&#8220;What do you think are the most important factors to mitigate? Why those factors, and how do you think the approach mitigates those?&#8221; </p></li><li><p>&#8220;How do you see the intended user using the feature? Walk me through the flow for x use case ?&#8221;</p></li><li><p>&#8220;What do you think will work and what are you worried will not? What did you deliberate the most on with this approach?&#8221;</p></li><li><p>&#8220;What led you to this specific UI (dig into specific affordances, placement, components, and how they affect the user&#8217;s experience and interpretation of the experience?&#8221;</p></li><li><p>&#8220;How do you expect the user to feel throughout this experience?&#8221;</p></li><li><p>&#8220;Do you feel that the user will clearly understand at each stage why this is being asked of them or what their options are?&#8221;</p></li><li><p>&#8220;What are the principles that you feel are really important in this experience?&#8221;</p></li></ul></li></ul></li><li><p><strong>Once the design approach is fully understood, hone in on the key questions worth discussing.</strong></p><ul><li><p>Once you fully understand what&#8217;s being presented, focus on highlighting the areas that work and the areas you&#8217;re worried about. Discuss each of these to the depth applicable. Mostly focus on what&#8217;s missing or what&#8217;s going the wrong way over what everyone likes. <strong>Discussions around these areas are where the majority of the progress will be made.</strong></p></li></ul></li><li><p>Look for problems at a few different levels:</p><ul><li><p><strong>Path:</strong> will users effectively get from point A to point B or is there danger they&#8217;ll get lost or stuck?</p></li><li><p><strong>Comprehension:</strong> will users understand the affordances in place, is there enough context building, is there an overwhelming amount of information, do users understand the possible actions and their outcomes?</p></li><li><p><strong>Requirements:</strong> are we solving for the requirements or are we off track, are there new requirements that may have surfaced, will this thing deliver the value we expect? </p></li><li><p><strong>Complexity: </strong>is the solution being presented overly complex from any lens, is it overly complex to build, hard to maintain, hard to support, has redundant parts, has unnecessary parts, is it overkill, or should it be simpler?</p></li><li><p><strong>Placement: </strong>Why is this feature in one part of the product and not the other? Why does this path take us here? Why does it exist at the 1st layer of interaction and not the second? </p></li><li><p><strong>Sentiment: </strong>is there a specific feeling or notion that is key to the user&#8217;s success with this experience? Does the laid-out experience keep that in mind/achieve it? </p></li><li><p><strong>Principles: </strong>are there essential principles to this experience that must permeate the experience (for example, the experience should always feel non-intimidating, or users should feel rewarded at each step)? Does the design factor these principles in? </p></li></ul></li><li><p><strong>Finally, focus on the clear next steps to move the project forward. </strong></p><ul><li><p>Always end with clear next steps or areas of questioning, and align with the designer on what these are and why they&#8217;re valuable. Be very conscious about what the process needs, in particular:</p><ul><li><p>Do we just need to keep exploring the use case further and go more in-depth into certain areas?</p></li><li><p>Do we need to validate certain notions so that we can safely move forward?</p></li><li><p>Do we need to be more focused on a very specific use case/customer journey and run the design through that use case/customer journey?</p></li><li><p>Do we have a lot of unknown unknowns, should we go talk to customers to find these?</p></li><li><p>Do we need technical input to ensure that what we&#8217;re coming up with is buildable?</p></li></ul></li><li><p>Design reviews will very often end with new questions. It&#8217;s important to highlight the progress made in closing older questions and pinpoint the new areas of investigation. If both the PM and designer are aligned on these, you&#8217;ll often make quick progress on the problem to be solved.<br></p></li></ul></li></ul><p><strong>Note: </strong>Over-invest in philosophical conversations when there isn&#8217;t a strong internal sense of what you&#8217;re trying to do, don&#8217;t over-invest in philosophical conversations when the truth can easily be arrived at with exposure to the customer. Remember that sometimes, you might need to ship the thing and see what happens with the scale. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to: Coexist with your predecessor’s work]]></title><description><![CDATA[When the ghost of products past won&#8217;t stop haunting you.]]></description><link>https://www.rewritetheroadmap.com/p/how-to-coexist-with-your-predecessors-work</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-coexist-with-your-predecessors-work</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Wed, 13 Dec 2023 11:29:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3782b96b-5fff-442f-b836-1190bb66c890_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>You&#8217;ve just joined an organization and while at first, you were excited about the cool products and features you were going to get to work on, you&#8217;ve now realized there&#8217;s a lot swept under the rug that needs to be dealt with. You&#8217;re frustrated about the state of things and overwhelmed by how much is now in your hands to wrangle.</p><p>It is tempting when facing this situation to enter a victim mindset and blame any problems that come up on the poor decisions of your predecessor or the organization at large. This guide has the goal of helping you find a more productive, healthier relationship with the work that was done before you so that you can ultimately be more effective in your role. </p><div><hr></div><h2><strong>Empathize, empathize, empathize </strong></h2><p>Understanding your predecessor&#8217;s work can help you significantly fast-track your insight into the business, products, and organization. Your predecessor was not a moron. The chances are that your predecessor was quite bright and impactful but worked in a limiting environment. Respect their work and their capabilities.</p><p>Try to understand, In detail, the decision-making processes of your predecessor. Specifically, try to understand the following:</p><ul><li><p>How did they approach their work?</p></li><li><p>Why did they approach their work in that way?</p></li><li><p>What were they prone to doing/missing as a result?</p></li><li><p>What did they excel at? </p></li></ul><p>The moment you become dismissive of someone else&#8217;s work you&#8217;ll become myopic and fail to see the value and patterns they have created which can slow down your rate of progress.</p><div><hr></div><h2><strong>Understand the products you&#8217;re inheriting, and fight the urge to burn it all down</strong></h2><p>Some PMs immediately move to &#8220;rebuild&#8221; mode. Rebuilding is often a massive mistake as it is resource-consuming, requires a lot of organizational equity, and, if done too early, will likely kill value inadvertently. </p><p>If you approach your inheritance as if you were starting at -100%, you&#8217;ll end up spending all your time getting to 0%. Think about your inheritance in terms of starting with an advantage that can be built upon.  In most cases, fine-tuning, optimization, tweaks, and simplification can help you supercharge what you already have. Strive to be fair but opinionated as you develop your understanding of the work and ask yourself: </p><ul><li><p>What works well?</p></li><li><p>What doesn&#8217;t work well? </p></li><li><p>What is close to working well but needs a few changes?</p></li><li><p>What do customers love?</p></li><li><p>What features do customers use the most?</p></li><li><p>What is the team most passionate about? </p></li><li><p>What strikes you as the most compelling? </p></li></ul><p>Give yourself the time to understand your products. Don&#8217;t knock down a structural pillar out of spite, and don&#8217;t take shots in the dark out of frustration.</p><div><hr></div><h2><strong>Understand the decision-making environment at the time</strong></h2><p>It is essential to understand the decision-making environment to effectively understand why a feature exists in the way it does. If you don&#8217;t understand why a decision was made at the time, it&#8217;ll be much harder to make a new decision on the feature now.</p><p> To better understand the decision-making environment at the time, ask yourself questions like:</p><ul><li><p>What was the process like when these features were built?</p></li><li><p>At the time, was there enough talent or domain expertise on the team to solve this problem?</p></li><li><p>Were there internal or external pressures when these decisions were made?</p></li><li><p>Has the business significantly changed around this feature?</p></li><li><p>What would have happened with my process if deployed at that time in the organization?</p></li></ul><p>If you don&#8217;t understand the company&#8217;s culture and its progression, you&#8217;ll eventually find it working against you and your goals. Most of the time, you&#8217;ll find that bad historical outcomes were a result of poor processes, low skill, or pressures from leadership, investors, or high-profile customers. Outside of those factors, you&#8217;ll find that some features might have just overstayed their welcome and that the business has changed since they were originally launched. </p><div><hr></div><h2><strong>Work to define what your version of the role will be like</strong></h2><p>When inheriting a product and roadmap, it is key to establish your version of the job and build alignment with stakeholders. </p><p>In many cases, there will be give and take, it&#8217;s hard for people to dramatically shift expectations on a role, but you can work on this expactation over time. While this type of expectation can initially be set during interviewing, in many cases, expectation setting is an ongoing part of working within an organization. </p><p>Take the time as you build a working relationship with your manager and team to talk through what you believe is most essential in the role, what you expect of yourself and of others, and what will happen if you&#8217;re capable of pursuing the role in this way. Having these kinds of conversations can be essential for you to carve out the version of the role you believe in while understanding what areas you&#8217;ll need to address and support so that your team continues to operate effectively. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Thoughts on: The 16 things I wish I knew starting out as an APM]]></title><description><![CDATA[From imposter syndrome to developing opinions.]]></description><link>https://www.rewritetheroadmap.com/p/16-things-i-wish-i-knew-as-an-apm</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/16-things-i-wish-i-knew-as-an-apm</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sat, 25 Nov 2023 01:09:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/f2fc6766-8ec1-450b-929a-f6ba55636ff4_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is a collection of the top 16 things I wish I had known when I started. I update this every once in a while, as my perspective changes.</p><div><hr></div><h4><strong>1) Yes, imposter syndrome is normal</strong>.</h4><p>Yes, it will get better with time. You&#8217;ll be surprised how time will make even the most intimidating tasks feel second nature. That being said, use that feeling of fear, shame, and helplessness as fuel to get you to the next level.</p><p>Most of all, start getting comfortable with not knowing. Every year there will be many, many challenges that will be totally new and unknown to you. When encountering these, focus on the feelings of curiosity and fascination to build a better relationship with not knowing.</p><div><hr></div><h4><strong>2) It&#8217;ll never be enough&#8230; and it&#8217;ll never end.</strong></h4><p>Fall in love with the journey or you&#8217;ll end up hating the work. It will kind of always be a grind and it will always challenge you. </p><p>In some environments, others will indirectly chip away at your love for the work. Protect your passion as much as you can. Find ways to make the work move you. Find the magic in the creative process and find the magic in the impact the work has on people to keep yourself motivated and invested.</p><div><hr></div><h4><strong>3) Stop trying to control everything.</strong></h4><p>Everything will get 10x easier once you realize control is pointless. You can&#8217;t control others, even when you have authority, you&#8217;ll barely be able to control people.</p><p>Focus instead on being a source of clarity, focus, energy, empathy, and passion. Creating and harnessing momentum is much easier and effective than pushing and dragging others.</p><div><hr></div><h4><strong>4) You&#8217;re not paid by the word.</strong></h4><p>Precision and clarity are critical to getting ideas across. The less you say, the more each word matters. How you say it can also make all the difference.</p><div><hr></div><h4><strong>5) Disagreements are a natural part of the creative process.</strong></h4><p>Never lose sight that the creative process requires conflict. At the same time, ensure that conflict can happen in a safe way. Never make it personal, always take the time to understand the other person&#8217;s thought process, and challenge ideas while acknowledging that you don&#8217;t know what the best outcome is (only that you&#8217;re seeking it). </p><p>Don&#8217;t be afraid to get into a really passionate discussion about something. If things end poorly, take the time to reach out and acknowledge your appreciation for the other person&#8217;s work, passion, and desire to get it right. Always mend fences when fences need mending. </p><div><hr></div><h4><strong>6) Always take the time to form an opinion.</strong></h4><p>At the end of the day, your perspective is one of the key things you can add to a team. Remember though, your opinion is only valuable if you&#8217;ve really put in the work. You also don&#8217;t have to be a contrarian or fit some idealized template, it&#8217;s perfectly fine to look at a situation and have nothing additional to say if the team has said all that needed to be said.</p><p>In most cases, if the team knows you&#8217;re reliable, and that you&#8217;re going the extra mile, they&#8217;ll respect what you have to say. Even in environments where you feel like your opinion is not wanted, you&#8217;ll find that if you continue to develop it, you may eventually find a way to break through. A good opinion in the right situation makes all the difference.</p><div><hr></div><h4><strong>7) Strong opinions, loosely held.</strong></h4><p>You&#8217;re seeking truth in projects, not self-validation. You want to be incredibly resolute about what you believe in but, at the same time, be ready to throw that belief out the window if you&#8217;re presented with highly compelling evidence or arguments against it. Being stubborn when you&#8217;re wrong is useless. Being stubborn when you&#8217;re right can be essential to the success of a project.</p><p>It takes a lot of willpower to make a meaningful impact, but it also requires an open mind and a strong ability to reset. </p><div><hr></div><h4><strong>8) Relationships matter a lot.</strong></h4><p>Building and maintaining trust with stakeholders is essential to any success you&#8217;ll have. Relationship management will be key throughout your journey whether it is to foster a strong collaborative process, persuade others to change their mind, get help when necessary, or battle tough times. Without others trusting you, you&#8217;ll find the job to be impossible. </p><div><hr></div><h4><strong>9) You deal in the What and Why, only sometimes in the How.</strong></h4><p>If you&#8217;ve come from an individual contributor background this will be really hard to shake off. Once you do shake it off, you&#8217;ll find that you can become an amazing partner to your teammates. The What and the Why is the stuff you have time and skill set for. In most cases, you won&#8217;t have the time, or oftentimes, the skill and knowledge to deliver an effective How. </p><p>Over time you&#8217;ll find a need for the How and the joy it can bring. Find safe spaces for the How whether it is in your personal projects or side projects in your work. If you abandon the How you&#8217;ll find it harder and harder to relate to your team over time.</p><div><hr></div><h4><strong>10) Never compromise on common sense.</strong></h4><p>You&#8217;ll be blown away by how quickly common sense goes out the window. You&#8217;ll also be amazed how quickly you&#8217;ll lose it yourself. Compromising gets things done but push to never compromise on common sense. Without common sense, a lot will break with your products.</p><div><hr></div><h4><strong>11) Nurture the creative environment</strong></h4><p>Create an environment for creative decision-making by communicating clear use cases, goals, and priorities, and having an open mind. Spend a lot of time learning how to make yourself a better partner to designers, engineers, and other creators. I highly recommend reading about music producers such as Brian Eno and Rick Rubin, and how they partner with artists in the creative process. </p><div><hr></div><h4><strong>12) Cherish the craft</strong></h4><p>Software is a craft and most people on your team are artisans. Think about what motivates every team member and why they&#8217;re in this line of work.</p><p>When thinking of software, a famous line from Moneyball comes to mind:<em> &#8220;How can you not be romantic about baseball?&#8221;</em></p><p>It is an unbelievable thing we get to do every day. In the work, in the details, there is magic. Creating is an incredible privilege, creating for others (especially creating things you love) is even more so. Don&#8217;t lose sight of that privilege. </p><div><hr></div><h4><strong>13) You literally can&#8217;t solve every problem</strong></h4><p>There are too many problems. New problems are made every day (if not every hour). The sooner you stop playing wack-a-mole the more effective you&#8217;ll be. </p><p>Constantly ask yourself &#8220;What is the highest value problem?&#8221;, and do your best to constantly pursue that problem. Additionally, seek to build systems and documentation that enable others to tackle problems that you care about solving.  </p><div><hr></div><h4><strong>14) If you believe in an idea, pursue and validate it</strong> </h4><p>Change is a constant. If you think something is worth pursuing, validate it by yourself and with the team. It might end up getting prioritized in the roadmap. Your belief that something is worth doing is one of the strongest intrinsic guides you&#8217;ll have through this messy world. Nurture it, develop it, and lean into it. </p><div><hr></div><h4><strong>15) Simplicity is a key</strong></h4><p>Simplicity without quality loss is the key to many of your problems. In any problem you&#8217;re seeking to solve, start by finding the essential. When you&#8217;ve arrived at the essential you&#8217;ll find, most times, that you&#8217;re much closer to the answer than you previously thought (if not there already). </p><p>The areas in which I&#8217;ve had the most impact were the areas where I was able to reduce lots of complexity to an easy to understand concept that accurately described the phenomenon or mechanic at hand. Most progress is built on a very well understood concept or sets of connected concepts. The more effectively you can understand and effectively describe complex ideas the more impactful and wide reaching your work will become.</p><div><hr></div><h4><strong>16) Be Yourself </strong></h4><p>The best PM version of yourself will be founded on who you are as a person.&nbsp;You&#8217;ll never be your best by being someone else. If you focus on finding the best PM version of yourself you&#8217;ll be something no one else can be. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to: Find low-hanging fruit projects]]></title><description><![CDATA[8 key ways to find big wins in small packages]]></description><link>https://www.rewritetheroadmap.com/p/how-to-find-low-hanging-fruit</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-find-low-hanging-fruit</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sun, 19 Nov 2023 18:49:38 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3c751d63-35e7-4dab-af3c-fbd0b81fc479_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>When looking at potential projects you can find a few common shapes:</p><ul><li><p>Low value | high cost &#8594; AKA mistakes </p></li><li><p>Low value | low cost &#8594; AKA why bother</p></li><li><p>High value | high cost &#8594; AKA core roadmap items</p></li><li><p>High value | low cost &#8594; AKA low-hanging fruit </p></li></ul><p>As a result of being high value and low cost, low-hanging fruit is some of the most desired work for an organization. It is, however, hard to find low-hanging fruit if you aren&#8217;t regularly and deliberately looking for it.</p><p>This guide seeks to provide paths that can yield low-hanging fruit opportunities.</p><div><hr></div><h2>8 ways to find low-hanging fruit:</h2><p></p><h4><strong>1) Find the friction</strong> <br></h4><h5>Why it works:</h5><ul><li><p>Most problems in your product will not be reported by end users unless they are leading to the total failure (or significant breakdown) of a feature. </p></li><li><p>As a result, there&#8217;s a very good chance that your product has friction in some key places that are invisible to you but highly disruptive to your users. </p></li><li><p>If you find and unblock those pathways and you&#8217;ll see a higher fraction of users performing actions successfully. Do this for an important enough pathway and you&#8217;ve found a big win. <br></p></li></ul><h5>How to look for it:</h5><p>Start by thinking of the most important actions that happen in your product, then try to identify any points of friction in the process of performing those actions. </p><p>Ask yourself &#8220;Would people reasonably understand the path to take?&#8221;, &#8220;Is the feature slow, is load time an issue?&#8221;, &#8220;Is it a big pain to get from point A to B with this feature?&#8221; and &#8220;Does this experience suck?&#8221; to find potential improvement areas.</p><p>Additionally, look at recordings, do a few feature walkthroughs yourself, look at feature data, talk to customers about it, run usability testing or feature walkthrough sessions, and/or map a funnel for the feature if possible as these can help uncover those points of friction. </p><div><hr></div><h4><strong>2) Find a small win for the biggest possible audience<br></strong></h4><h5>Why it works:</h5><ul><li><p>Big wins are generally much harder to find than small wins. </p></li><li><p>There are more small wins out there than big wins. </p></li><li><p>It is, often times, easier to find a small win for a big population than a big win on a small population (or a big win on a big population). A small win on a big population will, often times, have the same effect as a big win on a small population.</p></li><li><p>It is therefore an effective strategy to try to identify small wins on the broadest population possible as these will be easier to find and likely to yield big effects on your product and users. <br></p></li></ul><h5>How to look for it:</h5><p>Look for areas of your product that experience a great deal of traffic or that substantial portions of your population rely on. Focus on identifying small potential ways to significantly improve those areas. </p><p>Ask yourself &#8220;What are some ways we can improve efficiency by 1-10%?&#8221;, &#8220;What&#8217;s the easiest and cheapest way to make a massive improvement without significantly changing the product we have today?&#8221;, or &#8220;What&#8217;s the clearest no-brainer improvement we can make?&#8221; to start identifying these areas. </p><div><hr></div><h4><strong>3) Find the hidden bug in a core system<br></strong></h4><h5>Why it works:</h5><ul><li><p>The longer a system has existed and the more other systems leverage it, the more likely there are hidden issues with it or with the systems that rely on it. </p></li><li><p>Similarly to looking for friction, finding and removing the hidden disruption in a high-leverage system will often create many positive outcomes. </p></li><li><p>In many cases, teams will inadvertently build solutions around broken features. By solving critical bugs you can remove that unnecessary overhead and make your system more reliable in the process. <br></p></li></ul><h5>How to look for it:</h5><p>Think about your critical systems and assume there&#8217;s something wrong with them. Subsequently, spend time trying to &#8220;break&#8221; the feature to see if you can identify issues. </p><p>Ask yourself &#8220;If this feature is broken, why would we not know?&#8221;, &#8220;What are the worst ways this feature could be broken?&#8221;, and &#8220;What part of this feature does no one really know how it works&#8221; to find potential leads. </p><p>Additionally, spend time with your team and other internal stakeholders of the feature and try to dig for any known issues or concern areas that might exist. Stakeholders can often guide you in fruitful directions as they&#8217;ll have a strong understanding of the feature as well as fears about how it could fail. </p><div><hr></div><h4><strong>4) Find the biggest expenses<br></strong></h4><h5>Why it works:</h5><ul><li><p>The longer a 3rd party exists in your system the likelier it is that you&#8217;re overusing it without realizing it.</p></li><li><p> Cost problems tend to compound over time as teams can grow to leverage the third party in inefficient ways or turbocharge expenditures accidentally. </p></li><li><p>If your business doesn&#8217;t do a good job cleaning 3rd party bills or have effective expenditure monitoring, then you&#8217;re likely to find many wins around cost-cutting and making your business more profitable.<br></p></li></ul><h5>How to look for it:</h5><ol><li><p>Take all the expenses your business has, especially the expenses related to software and 3rd party tools (think about any feature that leverages a third-party system that charges at a fixed (or better yet variable) rate). </p></li><li><p>Map all those expenses in a spreadsheet along with the cost, billing cycles, any variables that you control, the usage data for that feature, and the qualitative importance of the feature.</p></li><li><p>Once you have that list in place, start with the highest-order expenses and work your way down trying to find which ones you can cut, simplify or downscale. Chances are that you&#8217;ll find expenditure bloat somewhere. </p></li></ol><div><hr></div><h4><strong>5) Find the most painful internal processes<br></strong></h4><h5>Why it works:</h5><ul><li><p>Non-technical teams will often get stuck in terrible processes because they lack the development support to fix the issue or because the company is stuck in an outdated mental model for a problem. </p></li><li><p>These processes can create significant wasted effort as well as affect morale. Worst of all, they can lead to the business incorrectly assuming it can&#8217;t scale because it assumes the operational model is unchangeable (without significant investment). </p></li><li><p>You can often find simple yet powerful solutions to these issues. These solutions often unlock significant growth potential for the business and free up teams to more effectively use their time.<br></p></li></ul><h5>How to look for it:</h5><p>If possible start with a concrete measurement of the most time-sucking processes for internal teams. If these are not available, ask team leads and individuals about what they think these processes are and where they rank one against the other. </p><p>Ask yourself, &#8220;What&#8217;s a process that needs to get done but currently gets done very inefficiently?&#8221;, &#8220;Could that process be done with a different tool and as a result be made significantly more efficient?&#8221;, &#8220;What&#8217;s the 20% version of this process that would yield 80% of the result?&#8221; and &#8220;Could we scrap that process entirely, if not what would be the second best path?&#8221; to identify leads.</p><p>Dig into internal processes and look for the ones that would be easiest to solve. In these processes, you can often find many easy internal efficiency wins. For more information on how to tackle internal efficiency projects <a href="https://www.rewritetheroadmap.com/p/how-to-approach-internal-tool-projects">check out this article</a>.</p><div><hr></div><h4><strong>6) Find the leak<br></strong></h4><h5>Why it works:</h5><ul><li><p>A bad enough data problem will consistently disrupt the user experience, your team&#8217;s focus, and introduce risk to your system. </p></li><li><p>In many cases addressing data inconsistency/utility issues can result in an efficiency boost to your product or can enable the team to drive a better experience down the line. </p></li><li><p>In most cases, teams already know where the bad data issues lie (if not where they start) making these kinds of problems some of the easiest to identify and understand. <br></p></li></ul><h5>How to look for it:</h5><p>Is there a data set in your system that&#8217;s totally broken? For example, do you have data that should be standardized but is not, data that should be in one format but is in a less useful format, or data that is only captured half the time? </p><p>Ask yourself &#8220;Where is the bad data becoming bad?&#8221;, &#8220;What are we doing today to address it?&#8221;, &#8220;What is the impact of this bad data both in terms of issues but also missing upside?&#8221;, and &#8220;What is the ideal solution to this bad data creation problem?&#8221; to better understand the leak and the potential value of fixing it.</p><div><hr></div><h4><strong>7) Find a way to connect the dots <br></strong></h4><h5>Why it works:</h5><ul><li><p>It is often the case that two experiences should have been connected but your team never had the time to complete the connection. </p></li><li><p>Solving these cases can sometimes unlock a 1+1 = 3 scenario as the synergies between experiences will amplify the outcome for the user or supercharge the overall experience.<br></p></li></ul><h5>How to look for it:</h5><p>There are a couple of good ways to find these kinds of opportunities:</p><p>Look to better leverage raw materials.</p><ul><li><p>Think about your product&#8217;s raw material. What are the outputs that are being created by your systems? These can be things like videos, favorites, likes, comments, posts, documents, etc. Now think about how that raw material could be leveraged elsewhere in your product to achieve a positive effect. </p></li></ul><p>Look to unite two separate experiences.</p><ul><li><p>Think about two key actions/workflows/experiences your users are performing that have overlaps in the use case, data, or mindset. Then think about what the benefits would be from connecting those two experiences in a more seamless way.</p></li></ul><div><hr></div><h4><strong>8) Find the feature that needs retiring<br></strong></h4><h5>Why it works:</h5><ul><li><p>Sunsetting features can create efficiency and morale boosts in ways that adding features cannot. </p></li><li><p>Features long past their usage will slow the system down, detract from the team&#8217;s core focus, and require deep knowledge that is often isolated to certain individuals. </p></li><li><p>As most organizations and teams have a hard time sunsetting past work, there is often a lot of opportunity to pursue this angle. <br></p></li></ul><h5>How to look for it:</h5><p>Look for things in your system that create complexity and cost for your team but that have questionable or unknown value. </p><p>Ask yourself &#8220;What are our least used features?&#8221;, &#8220;What are the features we hear the least about?&#8221;, &#8220;What are the features that the team hates interacting with or supporting the most?&#8221;,  &#8220;What do we keep getting bugs about that we hate working on?&#8221;, and &#8220;What happens if we do away with the feature altogether?&#8221; to find potential areas of your product that need to be cut out or simplified.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to: Report your experiment’s findings]]></title><description><![CDATA[Part 4 of a four part series on how to identify, sell, run and report on experiments]]></description><link>https://www.rewritetheroadmap.com/p/how-to-report-your-experiments-findings</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-report-your-experiments-findings</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sun, 12 Nov 2023 23:25:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/494522db-0d98-446f-88df-ba92e7b904d5_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<ol><li><p><a href="https://www.rewritetheroadmap.com/p/identify-experiment-opportunities">How to: Identify a good experiment opportunity</a></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-get-buy-in-for-your-experiment">How to: Get buy-in for your experiment</a></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-run-an-experiment">How to: Run an experiment</a></p></li><li><p><strong><a href="https://www.rewritetheroadmap.com/p/how-to-report-your-experiments-findings">How to: Report your experiment&#8217;s findings</a> &#8592; You&#8217;re Here</strong></p></li></ol><div><hr></div><h2>The experiment is complete&#8230; What now?</h2><p>Once the experiment is finished, assuming everything about the design of the experiment and the tracking of the results was set up correctly, there are three key actions to take:</p><ol><li><p>Analyze your results</p></li><li><p>Report your results</p></li><li><p>Identify next steps</p></li></ol><div><hr></div><h2>Analyze your results:</h2><p>Analysis is often a surprisingly challenging part of the experimentation process. While you&#8217;re looking to extract truth from data, you&#8217;ll find that it is often possible to use data to create many seemingly believable narratives. This makes it challenging to land on solid truth. </p><p>During analysis, you&#8217;re looking to create a clear separation between the core hypothesis validation that the experiment was designed around, from the other interesting factors that might have surfaced during the experiment. Focusing on your core hypothesis is a good way to stay grounded in your search for truth.</p><p>If the experiment was set up effectively this should be simple as you&#8217;ll have clear data indicators of if the hypothesis was proven or disproven. If your experiment was more open-ended, then you may not have a great deal of clarity but the experiment might have helped point you in a more promising direction.</p><p>The following are a few questions that can help navigate the analysis process:</p><ol><li><p><strong>Core Hypothesis Questions:</strong></p><ol><li><p>Was the experiment conclusive? Can I make a pronouncement on my initial hypothesis?</p><ol><li><p>Always go back to your initial hypothesis and analyze the data you obtained in context of that hypothesis (specifically if it proves/disproves it).</p></li><li><p>If applicable, check for statistical significance to determine if there&#8217;s a clear result for A/B or multivariate tests.</p></li></ol></li><li><p>What conclusions can I make about the experiment&#8217;s results that I would bet my life on? </p><ol><li><p>Push yourself to an extreme standard as this will help clarify the things you full-heartedly believe in, from the things that you sort of or mostly believe in. </p></li><li><p>If you report assumption-riddled insights, you may lead your team down the dangerous path of developing solutions that won&#8217;t work as the underlying premise for those solutions is untrue. Ensure that you&#8217;re clearly separating the proven insights from the ones that haven&#8217;t been validated. </p></li></ol></li><li><p>What things would perform exactly the same if a 3rd party were to re-run the experiment?</p><ol><li><p>This is another way to evaluate the results you&#8217;re seeing, this time through the lens of repeatability. </p></li><li><p>Results that you strongly believe could be replicated, or that others could replicate in the future, are results that you can count on. If you have something that will perform consistently you might have the foundation for a new feature, product, or process.</p></li></ol></li><li><p>How did the key metrics I was tracking perform? What could their performance mean in relation to the core hypothesis?</p><ol><li><p>Did the metrics perform according to your initial expectations or were you surprised? Is there a discernible pattern or are some metrics moving in contradictory ways? </p></li><li><p>Look at the different possibilities your metrics are illustrating but remember the bar of truth you&#8217;re looking for. It&#8217;s ok for an experiment to end with new questions and some ambiguity as long as you&#8217;re acknowledging that that&#8217;s the case. Most experiments result in follow-up research being required.</p></li></ol></li></ol></li><li><p><strong>Other questions worth asking:</strong></p><ol><li><p>Are there any qualitative aspects worth highlighting?</p></li><li><p>What variables were most surprising/unsurprising?</p></li><li><p>What would be the most interesting things to follow up on?</p></li><li><p>Were there clear outliers in the experiment?</p></li></ol></li></ol><p>Once you&#8217;ve answered the above questions, you should be capable of putting together a report on the experiment and determining what comes next. </p><p>Before covering reporting, there are a few things to keep in mind during analysis:</p><p></p><h4>Inconclusive results</h4><p>If your experiment was inconclusive, the first place to start is with the experiment design<em>:</em></p><ol><li><p>Did you build the right sample? </p></li><li><p>Did you set up an experience that effectively tested your hypothesis? </p></li><li><p>Were there variables in the experiment that you might not have been aware of initially that would have significantly affected the results? </p></li><li><p>Have you run the experiment long enough? </p></li></ol><p>In some cases, you might have solid experiment design and still get no results. Generally, this means that there&#8217;s no significant change in behavior. If this is the case, that is a valuable finding in itself and should be reported as such. </p><div><hr></div><h3>Report your results:</h3><ul><li><p>Being capable of effectively translating your findings to a broader audience is one of the single most impactful things you can do as a product manager. </p></li><li><p>If you&#8217;re experimenting effectively, the information and insights that you&#8217;re uncovering are some of the most well-proven and reliable insights accessible to the business. As a result, you can have a strong effect on how the business operates and thinks. </p></li><li><p>To do this, however, you need to be very effective at reporting information. In many cases, you also need to spend a great degree of care and time putting together compelling resources that others can leverage to understand your work.</p></li></ul><p></p><h4>The key information items of a report:</h4><p>At the end of the day, in most organizations, the stakeholders interacting with your experiment only care about the results and that the results are reliable (understanding process and methods is secondary for most stakeholders). </p><p>You&#8217;ll also find that getting your results understood is something that only happens through proactive action. The more friction there is to learn about your findings the less likely anyone is to learn them. The more ambiguity, the more likely each stakeholder is to come up with their own interpretation.</p><p>In terms of reporting, the key aspects to cover are:</p><ol><li><p><strong>The hypothesis</strong></p></li><li><p><strong>The experiment conditions</strong></p></li><li><p><strong>The results</strong></p></li><li><p><strong>The insights from the results</strong></p><p></p></li></ol><p>In practice, a report will sound something like this:</p><blockquote><ol><li><p>&#8220;We had the following hypothesis&#8221;</p></li><li><p>&#8220;To test that hypothesis we set up this experiment&#8221;</p></li><li><p>&#8220;The following things happened during the course of the experiment&#8221;</p></li><li><p>&#8220;Based on these results we have strong proof that the following things are true&#8230; We also believe that the hypothesis has been proven/disproven...&#8220;</p></li></ol></blockquote><p></p><p>To ensure your results are not lost, focus on clean, crisp, impossible-to-miss reporting. Ask yourself &#8220;What are the key things that everyone should echo after reviewing?&#8221; Focus on delivering those key things and minimize possible distractions.</p><p></p><h4><strong>Format and delivery:</strong></h4><p>Without proper distribution and packaging your findings will get lost in the shuffle. This is an essential aspect of reporting, the following are two (of many) ways to do this:</p><ul><li><p><strong>A written doc or email</strong> covering the information and data from the experiment is a low-effort way to formalize the findings.</p><ul><li><p>I&#8217;ll generally use this medium if the results are worth reporting on but the underlying thesis is not fully consolidated.  </p></li><li><p>I try to send these to any relevant stakeholders for the experiment (i.e. anyone who would care to/should read the report). </p></li><li><p>To make this readable, look to highlight key portions such as the hypothesis, results and insights. Try to add a &#8220;too long, didn&#8217;t read&#8221; section at the top of your email covering the most essential insights to ensure at least something is retained.</p></li></ul></li><li><p><strong>A slide deck </strong>is a strong medium if the results are high leverage or if you&#8217;re at the culmination of a lot of research and work. Yes, slide decks have a bad rep but if you&#8217;re not in a reading-friendly organization (in most cases you won&#8217;t be) they can be a very effective tool to get people to quickly understand and absorb your work&#8217;s findings. </p><ul><li><p>The deck should cover (as mentioned above) the hypothesis, the experiment, the results, and the insights. </p></li><li><p>Each section is 1-2 slides, contains the highest importance information, and is visually tailored (writing, assets used in the experiment, videos of user behavior, data visualization, etc) to best convey the idea behind the slide. </p></li><li><p>You can always add an annex at the end with more detailed information on the experiment if necessary.</p></li><li><p>Worth noting, the goal of a slide deck is not to run through a presentation, but instead to create an easily consumable report that can also serve as the basis for further discussion. </p></li></ul><p></p></li></ul><p>Regardless of the medium or distribution method for your insights, always ensure two things happen:</p><ol><li><p><strong>You have made your findings as digestible as humanly possible</strong></p><ol><li><p>The bar to strive for is: If everyone in your organization read your report and were asked to state the findings every person would say the exact same thing.</p></li></ol></li><li><p><strong>You have allowed people to digest findings async but have also established time to review them synchronously with key stakeholders</strong></p><ol><li><p>The async review lets stakeholders digest findings, the sync review often becomes a rich collaboration and discussion space around your insights. </p></li><li><p>Both are extremely valuable and lead to a successful internalization of the findings you&#8217;ve presented.</p></li></ol></li></ol><div><hr></div><h3>Identifying next steps</h3><p>The next step for an experiment should revolve around one question: <strong>is it worth continuing down this line of thinking?</strong> </p><p>The answer to that question depends on the perceived value to be gained, the level of insight you have on the matter, and the context of the surrounding business. With that in mind, there are 3 possible paths after an experiment is analyzed and reported:</p><ul><li><p>Experiment or research further</p></li><li><p>Pivot </p></li><li><p>Productize</p><p></p></li></ul><h4>Experiment or research further</h4><p>This is a valuable path to pursue if you have:</p><ol><li><p>Clear areas of further investigation that are promising and likely to lead to insights that will bring value to the business. </p></li><li><p>The buy-in from the organization to continue your research path. </p></li><li><p>No clear productization path for what you&#8217;ve learned or limited value in productizing at this time.</p></li></ol><p>If you are to continue experimenting or researching then you should consider the following risks to your research path:</p><ul><li><p><strong>Loss of organizational buy-in</strong></p><ul><li><p>Keep in mind when evaluating the option to do further research that there is often internal pressure. It&#8217;s easy to lose momentum and buy-in in some cultures the longer the research takes. </p></li><li><p>Consider if the organization may be losing patience or feeling the pressure to move towards action. Selling is key to preventing this but sometimes so is acting (by productizing or taking other steps to unlock some of the value you&#8217;ve been experimenting around).</p></li></ul></li><li><p><strong>Limited future value</strong></p><ul><li><p>While it isn&#8217;t necessary to have a complete understanding of how the research you&#8217;re doing will unlock value, it should at least be in an area of the business that is rich in opportunity. </p></li><li><p>This means that there should be a reasonable expectation that better understanding the problems you&#8217;re looking at will unlock value in the future.</p></li><li><p>To an extent, you should also consider the future viability of productizing around your research path. Specifically, there should be an aligned expectation (between you and the organization) that productizing around your research area is not only a likely outcome, but a shared priority.</p></li></ul></li><li><p><strong>Narrow focus</strong></p><ul><li><p>Another thing to keep in mind when deciding to do more research is if you can start taking your findings and piecing them together into a larger theory. </p></li><li><p>Shifting your mindset from a single hypothesis to a larger theory can help you find a broader, more impactful area of focus for subsequent work.</p></li><li><p>Ask yourself questions like &#8220;What is the broader pattern that may be happening here?&#8221;, &#8220;If I distilled what I&#8217;m learning down to its simplest element, what would that be?&#8221; or &#8220;What would be the bigger truth here that could unlock a lot of value?&#8221; to get yourself into this mindset. </p></li></ul><p></p></li></ul><h4>Pivot </h4><p>This is will happen if:</p><ol><li><p>The areas of further investigation are not promising or likely to lead to insights that will bring value to the business. </p></li><li><p>There is limited/no buy-in from the organization to continue your research path. </p></li><li><p>There is no clear productization path for what you&#8217;ve learned or limited value in productizing.</p></li></ol><p>There are some key things to keep in mind when pivoting away from a research path:</p><ul><li><p><strong>Document and catalogue your findings</strong></p><ul><li><p>In the future you may come back to your research as the context of the business changes.</p></li><li><p>Throughly document your work and make it easy to find/search for. Future you will appreciate this.  </p></li></ul></li><li><p><strong>Make sure stakeholders are aligned</strong></p><ul><li><p>Many times stakeholders will have been invested in your research and will be surprised by a pivot.</p></li><li><p>Ensure that you&#8217;ve built the shared understanding and buy-in to pivot effectively. Explain why the pivot is happening and clarify where the work will go from here.</p></li></ul></li></ul><p>If you consistently report and sell the value of your research you might be able to prevent pivots, block pivots or gain an extra bit of room to finish out your research. This is why formalizing findings and building alignment is key to longevity in your research focus.</p><p>If you are forced into a pivot, remember that this door may be open for you again and that your work has not been wasted.</p><p></p><h4>Productize </h4><p>Productization is the best outcome as this means your findings are valuable enough to be made part of the product or processes of the company. Productizing simply means taking the mechanics of the experiment and doing the necessary work to enable those mechanics to function or be used at scale.</p><p>This is a valuable path to pursue if you have:</p><ol><li><p>A clear productization path for what you&#8217;ve learned and the right timing to execute on these insights</p><p></p></li></ol><p><strong>Things to keep in mind when shifting to productization:</strong></p><ul><li><p>If you are thinking of productizing, first ask yourself: &#8220;Is there more I need to understand about this before driving value/productizing?&#8221;</p><ul><li><p>Ensure that you are getting enough validation to derisk productization, but, at the same time, don&#8217;t be afraid to push for productizing mechanics that are getting strong results. </p></li><li><p>The investment to be made for productization is significant, make sure that your data and process is solid and that the insights are reliable enough to spend money on.</p></li><li><p>If you are being asked to productize based on an experiment and you&#8217;re not confident in that path, bring your reasoning and your data to the stakeholder. </p><ul><li><p>When there is a disconnect in productizing where one party wants to do it and the other does not, it tends to be because each party has a different interpretation of what the experiment showed. </p></li><li><p>You might not end up getting the stakeholder to buy-in to your position but the conversation will reveal the reason for the misalignment and you might be able to problem solve from there. </p></li></ul></li><li><p>Productizing and continuing the research path is not mutually exclusive. </p><ul><li><p>The more you&#8217;re capable of parallel pathing these two actions the more positive your organization will see your experimentation. </p></li><li><p>You&#8217;ll also start benefitting from productization as the data from the products you&#8217;re building will help feed your research path by uncovering new insights.</p></li></ul></li></ul></li></ul><div><hr></div><h3>Parting thoughts:</h3><ul><li><p>The more experimentation and research you do the more comprehensive your understanding of the business will be. </p></li><li><p>The deeper your understanding of the business the better able you&#8217;ll be able to create cohesive theories that describe its underlying mechanics and outcomes. </p></li><li><p>Big, validated, cohesive theories are incredibly value for organizations as they allow for new strategies to be developed and investments to be made in productizing in accordance with those theories. </p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading this post! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to: Write user stories]]></title><description><![CDATA[Crafting effective user stories for efficient software development]]></description><link>https://www.rewritetheroadmap.com/p/how-to-write-user-stories</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-write-user-stories</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sat, 04 Nov 2023 16:08:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3d210efb-00f0-43da-b094-a5c8c07234fe_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Why good user stories matter:</h2><p>The cost of detecting and fixing issues increases exponentially the longer they remain undetected in the workflow. Issues introduced at the beginning of a project are often, as a result, the most expensive issues to detect and fix.</p><div class="pullquote"><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b4Dx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b4Dx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 424w, https://substackcdn.com/image/fetch/$s_!b4Dx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 848w, https://substackcdn.com/image/fetch/$s_!b4Dx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 1272w, https://substackcdn.com/image/fetch/$s_!b4Dx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b4Dx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png" width="475" height="461.95054945054943" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1062,&quot;width&quot;:1092,&quot;resizeWidth&quot;:475,&quot;bytes&quot;:373627,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b4Dx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 424w, https://substackcdn.com/image/fetch/$s_!b4Dx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 848w, https://substackcdn.com/image/fetch/$s_!b4Dx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 1272w, https://substackcdn.com/image/fetch/$s_!b4Dx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2408c10-d919-44f6-8233-67507a001aa8_1092x1062.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> Researchers at Hewlett-Packard, IBM, Hughes Aircraft, TRW, and other organizations have found that purging an error at the beginning of construction allows for rework to be done 10 to 100 times less expensively than when it's done in the latter parts of a project.</p></div><p>For many teams, the user story is the consolidation point of an idea into an achievable piece of functionality. This makes user story creation a key point in the development cycle as it can significantly influence the project&#8217;s success and cost down the line. As a result, by delivering high-quality stories, you will be capable of effectively minimizing the downstream likelihood (and cost) of issues giving your projects a better chance to succeed. </p><div><hr></div><h2>User story disclaimer:</h2><p>Not everyone needs to use user stories or benefit from them. I&#8217;ve found user stories to be one of the lowest-effort ways to have documentation, alignment, and tracking for the development of a solution. </p><p>I&#8217;ve also tried many other ways, all having worked to different degrees of success/failure. It is not the system that makes your work succeed, it is how the team works together that truly matters. User stories are just another way to help teams work together more effectively.</p><p>That being said, many organizations find user stories to be the way they prefer to work so understanding how user stories can be made into effective collaboration, documentation, and tracking tools is a valuable skill for a PM to have.</p><div><hr></div><h2>User story principles:</h2><p>I&#8217;ve found that to maximize the utility of user stories there are a few core principles to keep in mind: </p><ol><li><p><strong>User stories are owned by the entire team:</strong></p><ol><li><p>User stories represent agreements between stakeholders. They reflect the functionality the team has agreed to build along with the context behind that choice and its execution.</p></li><li><p>While there may be one main writer for the story, each team member can, and should, provide input on the story itself and feel ownership over it.</p></li></ol></li><li><p><strong>User stories should be written in such a way that anyone, even those who weren't involved in the writing/review process, can quickly understand them.</strong></p><ol><li><p>This ensures that anyone can easily pick up a story, regardless of their prior level of involvement on a project or how long it's been since the story was groomed. The following are the key items people should be able to quickly grasp when reading a user story:</p><ol><li><p>The affected user</p></li><li><p>The functionality to be built</p></li><li><p>The intended outcome</p></li><li><p>The requirements to achieve that outcome</p></li><li><p>How you will validate that the implementation achieves the intended outcome</p></li><li><p>The effort required to tackle the story.</p></li></ol></li></ol></li><li><p><strong>It takes teamwork to move stories through the workflow: </strong></p><ol><li><p>For a user story to be moved across the workflow there needs to be understanding and buy-in for the work to be done. Without team alignment, the story itself is worthless as it represents a truth that only a fraction of the team understands and believes. </p></li><li><p>Making this acceptance of a user story&#8217;s readiness into an explicit, workflow-based, action that the team performs together can help ensure each story is viable and maximized.</p></li></ol></li><li><p><strong>User stories represent an end-user outcome that can be enabled by executing the story:</strong></p><ol><li><p>Stories are expressed from the perspective of the end user (which in some cases may be a system) and explain a feature and its outcome.  </p></li><li><p>Without an end-user value mindset, user stories can easily become ineffective and hard to grasp. Always ensure that the story actively describes an outcome to be achieved for the end user and the means to achieve it.</p></li></ol></li></ol><div><hr></div><h2>How to structure user stories:</h2><p>I generally construct user stories using 4 key sections. These sections are:</p><ol><li><p>User story</p></li><li><p>Additional context</p></li><li><p>Acceptance criteria</p></li><li><p>Testing plan</p><p></p></li></ol><h4><strong>User story:</strong></h4><pre><code><strong>Template: 

</strong>As a <strong>{{user type}}</strong> I want to be capable of <strong>{{action to be performed}}</strong> so that I can <strong>{{value to be experienced}} 

Example: 

</strong>As a <strong>paid user</strong> I want to be capable of <strong>favoriting a publication from the news feed</strong> so that I can <strong>track publications I&#8217;m interested in hearing more from.</strong></code></pre><ul><li><p><strong>What it is: </strong>A description, from the standpoint of the user, of the outcome to be delivered and the means to achieve it.  </p></li><li><p><strong>Why it matters:</strong> This is the easiest, fastest way to create context for the entire team on the goal of the work item. By having this context anyone can understand and make effective decisions when executing on the story.</p></li><li><p><strong>Questions for quality control:</strong></p><ul><li><p>Can someone who didn&#8217;t write the story understand its goal just by reading this section? </p></li><li><p>Does the whole team understand the user, the action, and the value? Do different people have different interpretations? </p></li><li><p>Is the user type clear and identifiable? </p></li><li><p>Is the story specific enough? Is the story too specific?</p></li><li><p>Is the end goal clear and is it easy to understand why it would be valuable?</p></li></ul></li></ul><p></p><h4><strong>Additional Context</strong></h4><pre><code><strong>Template:

</strong>There&#8217;s no template for additional context but generally this section is needed if there is supporting context that can be helpful to make a story likelier to be effectively tackled. 

Things that fit this section:
&#8226; Links to relevant materials such as historical documentation on the feature the story touches.
&#8226; Explanations of the business context or motivation that lead to tackling the story.
&#8226; Related issues/stories that may be worth consulting. </code></pre><ul><li><p><strong>What it is: </strong>The supporting context for the user story. </p></li><li><p><strong>Why it matters: </strong>Because sometimes user stories benefit from having additional context. This can help ensure that stories are fully understood.</p></li><li><p><strong>Questions for quality control:</strong></p><ul><li><p>Is the team likely to understand the context of the story? </p></li><li><p>Is there a context for the story that if understood reduces risk?</p></li><li><p>Is there additional information about the underlying feature or system that would be important to understand when tackling the story?</p></li><li><p>Is there anything that could easily be missed that could lead to the story being incorrectly addressed?</p></li><li><p>Have we tackled something like this in the past? Should that previous work be consulted when tackling the story?</p></li></ul></li></ul><p></p><h4>Acceptance Criteria (ACs)</h4><pre><code><strong> Template:</strong>

There&#8217;s no template for ACs, different stories require different ACs. It&#8217;s up to the team to ensure these work for the story. A good starting point is:
Given {{the case in the story}}:
1. {{This is the state}}.
2. {{This action}} {{results in this outcome}}.
3. {{This action}} {{results in this outcome}}.
(&#8230;)


<strong>Example:</strong> 

As a paid user on the feed, I should be capable of:
1.Seeing the favoriting button in the news card. The button should accurately reflect if the media publication is favorited/not favorited.
2. If the button was inactive (not favorited), clicking it should lead to the publication being favorited and the button should update to active. 
3. If the button was enabled, clicking it should lead to the publication being unfavorited and the button should update to inactive.</code></pre><ul><li><p><strong>What it is: </strong>The high-level requirements that, if met, validate that the feature is working and delivering the user story outcome. </p></li><li><p><strong>Why it matters: </strong>Because it creates a shared understanding of what requirements must be met for the story to be deemed complete. This makes it simple for anyone to determine if a story is complete or not.</p></li><li><p><strong>Questions for quality control:</strong></p><ul><li><p>If this user story gives us trouble, what will be the reason?</p></li><li><p>Is the team capable of testing the ACs in order to validate that the feature works?</p></li><li><p>Did the team go through these ACs, review them, and agree they&#8217;re ready? </p></li><li><p>If someone wasn&#8217;t involved in the AC writing, would that person be capable of understanding them and testing the user story using them?</p></li><li><p>If just these ACs are met, does this story successfully pass testing? If not what&#8217;s missing?</p></li><li><p>Do we need a testing plan? Can we use the ACs to test it in this case or would we benefit from more info here?</p></li><li><p>Is there anything about testing these ACs that is messy/easy to miss? </p></li></ul><p></p></li></ul><h4>Testing Plan</h4><pre><code><strong>Template:
</strong>
There&#8217;s no template for the testing plan, different stories require different testing plans or may not require them at all (if the ACs are clear and guiding enough). It&#8217;s up to the team to ensure these work for the story.

<strong>Example:
</strong>
Checking that favoriting publications works

1.Login as a paid user
2. Go to the feed
3. Click the favorite button on the news card. This should flip the button to active. 
4. Check the DB for a favorite record from the media publication for that user. 
5. Click the favorite (unfavorite in this case) button for the same news card. This should flip the button back to inactive.
6. Check the DB for the same record. This record should now be soft deleted.

1. Login as a free user
2. Go to the feed
3. There should not be a favorite button on the news card. </code></pre><ul><li><p><strong>What it is: </strong>A series of steps the team can undergo to validate that the feature is working as intended. </p></li><li><p><strong>Why it matters: </strong>For most tickets, the ACs are enough for the team to test. But for some tickets, you need more definition on how to test. This section allows the team to align on the testing plan when needed and document it for future use.</p></li><li><p><strong>Questions for quality control:</strong></p><ul><li><p>Do we actually need a testing plan for this? </p></li><li><p>How do you successfully test this feature step by step? </p></li><li><p>Are there cross-user considerations? </p></li><li><p>Is testing of this dependent on something else? </p></li><li><p>Do we need a manual action (from a dev or someone else) to test this? Like running a command.</p></li><li><p>Can this be tested on our regular testing environments? </p></li><li><p>What user accounts do we need, and what data do we need? Do we have a means of generating that testing data?</p></li></ul></li></ul><div><hr></div><h2>How to scope user stories</h2><p>The hardest part of defining a user story is determining its scope. While this is more art than science, there are questions you can ask yourself to help find the right scope:</p><ol><li><p><strong>Is the user story the right size? </strong></p><ol><li><p>User stories that are too large can slow down progress and hurt team morale as they make teams feel like the work is just not moving forward. Most times stories can be broken down into smaller stories which allow the team to more effectively, build, test, and release leading to faster value delivery. </p></li><li><p>It&#8217;s important to stress test stories for size and to keep in mind that most stories tend to be over-scoped. If you&#8217;re already breaking stories down, keep in mind that the reverse can happen as well. Sometimes teams can get overzealous with breakdowns leading to stories that don&#8217;t deliver value, feel incomplete, or are generally overvaluing size over value. Think about story size in terms of the end-user and the context of that story in the project, ask yourself &#8220;Can it be smaller&#8221; but also &#8220;Should it be larger&#8221;. </p></li><li><p>During the course of a project, it&#8217;s easy to bundle together items that don&#8217;t need to be bundled. It&#8217;s important to also ask yourself if there are pieces of the work that could be realistically done independently, or if there are parts that could hold back the story but shouldn't. This can help clarify story size and decouple the core parts of a story from the less essential parts.  </p></li></ol></li><li><p><strong>Does the story make sense from the standpoint of the end user?</strong></p><ol><li><p>It&#8217;s easy to lose sight of the end-user when optimizing user stories and this can easily lead to stories that are perfectly shaped for moving through a workflow but deliver poor or incomplete value on delivery. </p></li><li><p>Ensure that the stories you&#8217;re writing deliver meaningful change to the end user be that end user an actual user or a system. Ask yourself if the story delivers actual value if achieved or if there is something missing.</p></li></ol></li><li><p><strong>Is the story testable?</strong></p><ol><li><p>If a story cannot be tested you cannot ensure that it&#8217;ll deliver on requirements and as a result you cannot ensure that it is completed. For a story to be effectively deemed complete it needs to be tested which implies you have a clear way to verify requirements. </p></li><li><p>If this is not possible, the team will need to determine why it&#8217;s not possible and problem-solve from there. In many cases, if a story is not testable it is because there are dependencies that need to be solved first which makes it so that the story needs to be tackled later on once those dependencies are met.</p></li><li><p>Don&#8217;t move stories into the workflow without knowing exactly how to verify that they were completed or you run the risk of realizing the story isn&#8217;t achievable in its current state and needs to be rethought.</p></li></ol></li><li><p><strong>Is the story in the right place in the sequence?</strong></p><ol><li><p>During projects, one of the main reasons why stories can fail is because they are being worked on in the wrong sequence which can make it so that the story is not testable or that if completed it does not meaningfully move the project forward. </p></li><li><p>If tackling stories within the context of a project it&#8217;s key to understand the order in which stories are to be tackled as this will ensure that each story can be effectively tested and completed. Ensure that the team has thought about how the sequential execution of stories will lead to a finished deliverable. </p></li></ol></li><li><p><strong>Is there a clear delineation between what is to be covered in the user story vs elsewhere?</strong></p><ol><li><p>At all times it should be clear what information is to be covered in user stories and what information is covered elsewhere. For example, it should be well understood by the team if user stories cover design specifications or if these are covered in design specks or other materials.</p></li><li><p>Ensure that everyone understands what information the user story is responsible for and what materials cover the remaining information needed to tackle the story. It is very easy to miss requirements as a result of having ambiguity in terms of what information lives where.</p></li></ol><p></p></li></ol><div><hr></div><h2>Parting thoughts:</h2><p>It is important to remember that processes exist to ensure that certain things happen in the workflow. As a result, processes should be evaluated against their ability to ensure those things happen. It is also important to evaluate if those things happening are actually making the workflow better. </p><p>User stories exist, mostly, to help teams better define the work to be done, document it, and create alignment on it. If user stories are getting in the way of the team effectively collaborating and defining work then consider if 1. you and your team are using stories correctly or if 2. user stories are a fit for your team overall. </p><p>Ultimately, team workflows work best when they are iterative and allowed to evolve based on the needs and structure of the team but they also need time to improve and find the right shape. Don&#8217;t be afraid to question if the stories you&#8217;re writing are adding value and remember why you&#8217;re writing them in the first place.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Source: <a href="https://www.amazon.com/Code-Complete-Practical-Handbook-Construction-dp-0735619670/dp/0735619670/ref=dp_ob_title_bk">Code Complete: A Practical Handbook of Software Construction, Second Edition - by Steve McConnell</a> which itself references: <em>"Design and Code Inspections to Reduce Errors in Program Development" (Fagan 1976), Software Defect Removal&nbsp;(Dunn 1984), "Software Process Improvement at Hughes Aircraft" (Humphrey, Snyder, and Willis 1991), "Calculating the Return on Investment from More Effective Requirements Management" (Leffingwell 1997), "Hughes Aircraft's Widespread Deployment of a Continuously Improving Software Process" (Willis et al. 1998), "An Economic Release Decision Model: Insights into Software Project Management" (Grady 1999), "What We Have Learned About Fighting Defects" (Shull et al. 2002), Balancing Agility and Discipline: A Guide for the Perplexed&nbsp;(Boehm and Turner 2004).</em></p></div></div>]]></content:encoded></item><item><title><![CDATA[How to: Approach internal tool projects]]></title><description><![CDATA[Turning backstage tools into show-stopping successes]]></description><link>https://www.rewritetheroadmap.com/p/how-to-approach-internal-tool-projects</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-approach-internal-tool-projects</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Sun, 29 Oct 2023 18:16:58 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6ea1e6a5-77cf-4f84-9aa0-3ba093291308_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>Internal tools are first-class citizens:</h3><p>Internal tools get a bad rap. Many PMs find them to be &#8220;entry-level work&#8221; not worthy of their time and devoid of the magic and flash of consumer-facing products. The reality is, however, that high internal efficiency is a core aspect of a healthy company and a hallmark of a strong culture. </p><p>Growth is only possible when the team and products can support the scale of the user base. Without efficiency, you will create poor experiences for customers, drag down your product&#8217;s performance, and kill team morale (especially on customer-facing teams). With efficiency, doors open. Teams can then focus on key initiatives and you&#8217;ll get pulled less and less to deal with fires.</p><p>This guide covers how to maximize your time working with these tools.</p><div><hr></div><h2>A practical approach to building internal tools</h2><p>I follow a basic framework for tackling internal tool projects composed of five steps:</p><ol><li><p>Understand the problem and feel the pain yourself</p></li><li><p>Get your team to learn what you&#8217;ve learned</p></li><li><p>Find the right path for the solution</p></li><li><p>Execute with quality </p></li><li><p>Celebrate the improvement </p></li></ol><div><hr></div><h3>1. Understand the problem and feel the pain yourself:</h3><h4><strong>Quantitative Understanding:</strong></h4><p>The key question to understand with internal tooling is &#8220;What is the tangible value of solving this problem?&#8221;. If you understand the answer to this question very well, the solution will be easy to find. To answer this question you need to find the metric-based representation of the problem&#8217;s impact.</p><p><strong>For Example:</strong></p><p>Let&#8217;s say your company requires each new customer to be manually set up, and this setup is complex and time-consuming which results in customers not being set up within a reasonable time frame.</p><ul><li><p>Here you can measure the impact of this problem in a few different ways:</p><ul><li><p>The total support time spent getting customers set up.</p></li><li><p>The number of days that it takes the average customer to get set up.</p></li><li><p>The percentage of customers being set up outside a reasonable time frame. </p></li><li><p>Ticket and call volumes from customers due to set-up delays.</p></li><li><p>Sentiment scores from customers who experienced this issue. </p></li></ul><p> </p></li></ul><p>By creating a problem value statement that is metric-based all stakeholders can more easily align on the problem to be solved. This can create focus in your discussions as well as allow you to better understand how a proposed solution could impact the metric.</p><p>Not every problem can be reduced to a metric, many problems also have multiple value adds if solved, and some problems have immeasurable value. However, being capable of measuring a problem&#8217;s value in a clear way can make a huge difference in your approach and ability to focus.</p><p></p><h4><strong>Qualitative Understanding:</strong></h4><p>Sometimes PMs get overfocused on metrics as these can provide comfort to analytical-type individuals. However, a quantitative understanding of the problem is only a partial understanding. To truly understand the issue you need to experience it first-hand.</p><p>Here I look to do a few things:</p><p><strong>Understand how the team feels and thinks about the problem.</strong></p><ul><li><p>Get the perspective on the issue from stakeholders. In particular, talk to the individual team members facing the issue every day, their manager, and your manager. Find the misunderstandings and the overlaps, and understand how different people feel about the issue and why they feel that way. This will give you a clear understanding of your organization&#8217;s perspectives on the issue which can help you understand expectations and better navigate those expectations. </p></li><li><p>If each stakeholder has a totally different understanding and evaluation of the problem to be solved, you&#8217;ll end up experiencing a lot of friction in the process as stakeholders may not understand why you&#8217;re going down a specific solution path.<br> </p></li></ul><p><strong>Understand how the internal team is hoping the issue gets solved:</strong></p><ul><li><p>9 times out of 10 if you&#8217;re getting an ask on internal efficiency it&#8217;ll come prescribed as a specific feature that an internal team needs. </p></li><li><p>While user-driven feature requests are many times not useful, internal team-driven requests can sometimes break this rule. In organizations that don&#8217;t invest much in internal efficiency you&#8217;ll often find that team leads will have created their own solutions to these problems in lots of clever, and sometimes not so clever, ways. This can help you fastrack your understanding of the problem and its solution.</p></li><li><p>Don&#8217;t take the solution as gospel but don&#8217;t dismiss it outright. I&#8217;ve found with internal tooling that the average prescribed solution has tended to be a pretty good possible solution (although mileage may vary for you).</p></li></ul><p><strong><br>Understand the problem yourself by experiencing it first-hand.</strong></p><ul><li><p>Upon taking all the feedback, you have to do the work yourself. There&#8217;s no better way to get your own opinion on the issue than living a day in the shoes of those dealing with it. </p></li><li><p>Ask to take on the responsibility of one of the team members for a specific process or join a ride-along. </p></li><li><p>See the problem for yourself and more importantly, feel the pain for yourself. This very hands-on experience will make the problem more tangible in your mind.</p></li></ul><div><hr></div><h3>2. Get your team to learn what you&#8217;ve learned:</h3><p>Depending on how your development process runs, you may have more or less of an opening to bring in other stakeholders. In the right environment, you can have most, if not the whole, development teams participate in the discovery process and gradually iterate on the problem through cycles of learning, building, testing/validating. However, I&#8217;ve found this to be exceptionally hard at most startups and in most cultures due to lack of time, process, skill, or all three together. </p><ul><li><p>If you&#8217;re in a more waterfall-type process, remember that you can bring your findings to the team as you&#8217;re consolidating them, while first-hand experience is best, a steady series of updates and documentation can help someone be more integrated in discovery even if they can&#8217;t participate directly. </p></li><li><p>Another thing I like to do for these kinds of projects is to have the development team go through a simulated exercise that has them doing the work that the internal team is suffering through. Giving your development team first-hand experience of the issue and layering context to that experience makes the problem to be solved really tangible and gives them a direct relationship with that problem. This sets everyone up for a more effective discovery process and allows them to return to the exercise experience for guidance. </p><p></p></li></ul><p>The key thing to achieve is to ensure the team understands the problem they&#8217;re solving to a similar degree as you do. If you&#8217;ve achieved this you&#8217;ll find that the outcome you and the team will deliver will in many cases be 10x what you&#8217;d get if only you understood the problem deeply.</p><div><hr></div><h3>3. Find the right path for the solution:</h3><p>To find the right path you need to ensure that you&#8217;re properly weighing the variables of the problem and evaluating possible solutions accordingly. There are a few things you can do here:</p><p></p><h4><strong>Challenge assumptions, deconstruct and reconstruct your understanding of the problem:</strong></h4><p>Try to identify and deconstruct any assumptions that are being made about the problem and about a possible solution. Ask yourself </p><blockquote><p>&#8220;What are we relying on to be true? Is that actually true&#8221;</p><p>&#8220;How do we make this problem completely disappear&#8221; </p><p>&#8220;What&#8217;s the 20% solution that delivers the most value&#8221;</p><p> &#8220;What&#8217;s the ideal solution to this problem&#8221;</p></blockquote><p>I&#8217;ve found that pushing my thinking to extremes and challenging assumptions has yielded better solutions not because you end up building the extreme, you almost never do so, but because this type of thinking often helps cut through group thoughts, biases, and assumptions resulting in a clearer focus on the parts of the problem that add the most value.</p><p>Don&#8217;t worry about straying too far off course, it&#8217;s easy to get back on track if you understand the problem to be solved. But note that most teams tend to avoid challenging lines of thinking and intense discussions out of fear of conflict. Better solutions come from productive creative conflict, pushing the team&#8217;s thinking to challenge assumptions can help you do this in a positive way. </p><p></p><h4><strong>Test your solution hypothesis with the internal team, simulate it, and try to break it:</strong></h4><p>As you&#8217;re concepting the solution, bring in the internal team for feedback. Internal teams are the best kinds of users because most of the time you can bring them in as often as you&#8217;d like and they&#8217;ll spend as much time testing and talking to you as you need. Keep them in the loop at a minimum but consider taking advantage of their time and having them challenge the solution path you&#8217;re working through.</p><p><strong>Once you have a clear path,</strong> try to simulate what the workflow would look like with it in place. If you can&#8217;t simulate it, run through the solution step by step with the internal team instead and evaluate how it performs or how the internal team would expect it to perform. </p><p>As much as you can, try to challenge the path you&#8217;re thinking of taking and try to make it fail. Look to find ways it could fail, and try to identify flaws by questioning it under different scenarios. If your path makes its way through, you probably have something that will work in the real world.</p><p><strong>Go back to your problem&#8217;s key metric and evaluate how your solution would impact it:</strong> </p><p>For every internal efficiency project, I try to also put together a projection of what the world will look like with the solution in it. Go back to your initial quantitative metric for the problem and ask yourself &#8220;What is my expectation of the impact our proposed solution would have on the problem?&#8221; Putting together an estimate forces you to think through the value you&#8217;re solving for and the value you think you can capture. This type of calculation can help you refine your strategy and thinking around the problem.</p><p></p><h4><strong>Consider build, buy, change process, sunset, customer-facing options: </strong></h4><p>It&#8217;s very easy to fall into the trap of always solving problems by building, but remember building is one of many options to solve a challenge:</p><ul><li><p><strong>Building:</strong> Ask yourself also if what you&#8217;re trying to solve is core to what you do or secondary (but still a requirement). Building isn&#8217;t a one-time cost, it presents a continual need to support and maintain what you&#8217;ve built as well as added complexity to your system. If you're building, ensure you understand the commitment involved and have determined it&#8217;s worth it over all other options.</p></li><li><p><strong>Buy:</strong> If something is not core to what you do and requires a high degree of expertise and maintenance, then buying may be a better path. Third-party solutions can be the right choice in situations where there are clear market solutions and where the problem to be solved is not core to your business. </p></li><li><p><strong>Change/sunset process:</strong>  There are cases where you can change the process in such a way that much of the problem disappears. This isn&#8217;t very common but can be the case if the original process had a core assumption that you&#8217;ve proven to be untrue. If this is the case, sometimes you can work with the internal team to shift the process in a way where the problem you&#8217;re trying to solve essentially disappears. </p></li><li><p><strong>Make the solution customer-facing instead:</strong> While this can be a tougher sell, in some cases you can delegate part of the problem to the customer or position the solution in a way where the problem is removed via customer behavior. Consider if the problem you&#8217;re solving is actually just a symptom of a problem with how your product works for users and if a better end-user solution would mitigate the issue.</p></li></ul><p></p><h4><strong>Ensure everyone is rowing in the same direction:</strong></h4><ul><li><p>Ensure that everyone who is a stakeholder in the project is bought in on your solution path and understands the reasoning behind it. </p></li><li><p>Building buy-in will reduce the amount of friction you will feel during the project&#8217;s execution. If stakeholders don&#8217;t understand or believe in your proposed solution they&#8217;re likely to try to course-correct it which can lead to disruptions and tension. </p></li><li><p>With internal tooling, internal buy-in is very important because internal teams have more of a say in product decisions than regular customers do (or at least are capable of communicating their thoughts and feelings directly). Before anything gets built, make sure the internal team is bought in as this will make the partnership smoother and more productive.</p></li></ul><div><hr></div><h3>4. Execute with Quality</h3><p>To ensure that your solution is the best it can be there are a few things to keep in mind when executing that make a big difference:</p><ol><li><p><strong>Test for real scenarios.</strong> </p><ol><li><p>Don&#8217;t just test and audit like a robot, think about how the tool will be used by your team and try to simulate it. </p></li><li><p>If possible bring in the team for stress testing. Realistic testing will often find issues that pure auditing will not.</p></li></ol></li><li><p><strong>Get the internal team to be hands-on.</strong> </p><ol><li><p>Getting stakeholders to try out your solution in a testing environment early can uncover issues with your implementation. As development occurs, keep the internal team in the loop and provide them with visibility into the project as this will help catch issues but also increase morale.</p></li></ol></li><li><p><strong>Use realistic data and volume (stress test as much as you can).</strong> </p><ol><li><p>Don&#8217;t just test for an optimal scenario, make it as realistic as possible and test for extremes. </p></li><li><p>It&#8217;s easy to miss something because you&#8217;re testing with data that was convenient. Ask yourself, what does the worst real-world scenario look like, now imagine it&#8217;s twice as bad. Test with the date from that imagined scenario. </p></li><li><p>Also, ensure that your team members are testing with different data sets, it&#8217;s easy to miss issues when everyone is working with the same data.</p></li></ol></li><li><p><strong>Don&#8217;t throw your solution into the wild at the wrong time.</strong> </p><ol><li><p>Whatever you do, don&#8217;t try to rush a tool at the last minute to support a workflow. Tools need iteration and are generally never perfect on the first release. </p></li><li><p>Give teams time to ease up to the tool. Don&#8217;t go live on the worst week of the year. Shipping at the wrong time can create a strain on teams and decrease morale.</p></li></ol></li><li><p><strong>Build some padding in your roadmap for follow-ups.</strong> </p><ol><li><p>It&#8217;s tough for some internal tools to be fully tested without having real-world usage. Give yourself some space after release for iteration and for addressing any issues found in the wild.</p></li></ol></li><li><p><strong>Be realistic and honest about what you can and can&#8217;t get done.</strong> </p><ol><li><p>It&#8217;s normal in any project to be constrained in terms of the investment you can make. Acknowledge that you won&#8217;t be able to solve every problem and be honest with internal teams about this. </p></li><li><p>That being said, don&#8217;t lose perspective on the value you can create and how it&#8217;ll affect that team. Focus on that value and keep that focus during the development process.</p></li></ol></li><li><p><strong>Build support materials and documentation to ensure the solution is understood. </strong></p><ol><li><p>This applies to any kind of work but it is especially important for internal tooling. You should ensure that you have clear documentation explaining what the internal tool was built for as well as what it does, and how it does it. </p></li><li><p>Try to be as detailed and technical as possible in your documentation. Without documentation, the likelihood of forgetting key details about the tool is high which can lead to issues down the line. </p></li></ol></li><li><p><strong>Have a clear tracking plan around your metric to learn about how your solution is performing:</strong></p><ol><li><p>Before going live ensure that you have a way to understand if the solution you&#8217;re putting in place is effectively addressing the problem. </p></li><li><p>Understanding the solution&#8217;s real-world impact on the problem&#8217;s metric can help show you how to improve the solution and validate if you made the right decisions and assumptions in your process.</p></li></ol></li></ol><div><hr></div><h3>5. Celebrate the improvement:</h3><p>Building internal tools is a highly collaborative, highly personal endeavor as the tools you&#8217;re building affect the internal team&#8217;s daily lives. Ensure that wins with internal tooling are recognized and come back to the value they&#8217;re delivering. Praise those who participated in making the tool better both on your team and on the internal team who will be using it. </p><p>Over time you&#8217;ll find that the internal team will be energized from the tooling and your working relationship will grow stronger. Remember that while projects like this can sometimes be draining, they often mean a lot to those who benefit from them. </p><p>Ultimately you&#8217;re trying to create valuable products for people, internal tooling may have a small audience in many cases, but the work you&#8217;re doing optimizes those people&#8217;s work lives and helps your business grow.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Run an experiment]]></title><description><![CDATA[Part 3 of a four part series on how to identify, sell, run and report on experiments]]></description><link>https://www.rewritetheroadmap.com/p/how-to-run-an-experiment</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-run-an-experiment</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Thu, 19 Oct 2023 09:11:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/59dbf5fc-197f-40a6-8eee-2ea2fc64f9cd_420x300.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<ol><li><p><a href="https://www.rewritetheroadmap.com/p/identify-experiment-opportunities">How to: Identify a good experiment opportunity</a></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-get-buy-in-for-your-experiment">How to: Get buy-in for your experiment</a></p></li><li><p><strong><a href="https://www.rewritetheroadmap.com/p/how-to-run-an-experiment">How to: Run an experiment</a> &#8592; You&#8217;re Here</strong></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-report-your-experiments-findings">How to: Report your experiment&#8217;s findings</a> </p></li></ol><div><hr></div><h2>How to Experiment:</h2><p>The easiest mistake to make with experimentation is to set the experiment up in a way where nothing is learned. The vast majority of the time, this happens due to improper setup. </p><p>In particular, there are three key areas to consider when it comes to designing an experiment.</p><ul><li><p>The people in the experiment.</p></li><li><p>The structure of the experiment.</p></li><li><p>The tracking and measurement of the experiment.</p></li></ul><p>This guide will dive into each area and what to consider at each level.</p><div><hr></div><h2>The people in the experiment:</h2><p>When you set out to conduct an experiment you&#8217;re generally trying to prove that something is true for a large group of people.</p><p>The experiment is then testing your assumption about/on that population. In most cases, you can&#8217;t/shouldn&#8217;t test on the full population as it will have adverse effects on the business. If you send a new test email to all your users, for example, you might be exposing everyone to a poorer experience and you will also have to wait in order to be able to test a new version of the same email. To avoid this problem, you can build sample groups from the larger population and test on those groups. </p><p></p><h4><strong>Sampling:</strong></h4><p>By taking a sample group from the larger population, you can ensure that the group you&#8217;re experimenting with will yield results that will also be true for the full population. At the same time, by breaking up your population into smaller groups you&#8217;ll be able to run more iterations of the experiment than if you just tried on the full population from the start. </p><p>To create a sample you have to define two things:</p><ul><li><p><strong>The characteristics of the people in the sample.</strong></p><ul><li><p>Practically speaking what are the properties the population shares that make them distinct from other users? </p></li><li><p>Ideally, these characteristics are things you can easily use to segment your users in order to build the sample.</p></li></ul></li><li><p><strong>The amount of people you need in the sample to get accurate results.</strong></p><ul><li><p>The sample size calculation for the majority of experiments can be done using a simple sample size calculator (here&#8217;s one from <a href="https://www.surveymonkey.com/mp/sample-size-calculator/">SurveyMonkey</a> that I&#8217;ve used in the past).</p></li><li><p>These calculators can help you ensure that your sample is big enough to generate results that are very likely to be representative of the results you&#8217;d get if you were exposing the full population to the experiment. </p></li><li><p>They also can help you better understand how statistical analysis is done and how to interpret your results. <br></p></li></ul></li></ul><h4><strong>What if my population is made up of lots of different groups?</strong></h4><p>In some cases, you might be dealing with a population made up of different sub-groups each with their own unique characteristics.</p><p>In this case, you have to ask yourself if your hypothesis could be influenced by the particular properties of those groups. If it could be, the next question is if it warrants a different experiment or if you can just have that differentiation be something you include in your analysis. In most cases, this differentiation can just be a variable in your analysis as it won&#8217;t materially affect the experiment design. If your experiment is directly designed for a specific sub-group then you probably should segregate the user types.</p><p></p><h4><strong>What if there are very few people in my population?</strong></h4><p>In some cases, the population is too small to build a proper sample. In these cases, you&#8217;re generally better off just testing with the full population as the sample size requirement will essentially be very close to the full population size.</p><p>Sometimes, you can shift your experiment to focus on frequency to build confidence. For example, exposing the same user group to the experience several times over the course of a long window of time. That being said, I&#8217;ve found that going with the full population instead has tended to be the right path in these scenarios.</p><p></p><h4><strong>Focusing on a very small number of users:</strong></h4><p>There are cases (as I&#8217;ll cover below) where you&#8217;re trying to understand how to build something new. In these cases, sometimes your best path is to focus on how to manually support your intended experience for a small number of people and learn from their experience.</p><p>The thing to keep in mind in this context is that any learnings you&#8217;re making with a small number of people have higher risk since there is a chance the people you&#8217;re working with are outliers and the group is too small to cover the full variations of behavior you&#8217;d get from the larger population.</p><p>This shouldn&#8217;t discourage you from attempting this approach as you can always bring in a bigger sample to try your new experience later on. </p><div><hr></div><h2>The structure of the experiment:</h2><p>The structure of the experiment is the mechanism through which the experiment is conducted. There are some common templates that people gravitate towards for conducting experiments:</p><ol><li><p><strong>White Glove Service / Concierge Treatment:</strong></p><ol><li><p>Here you&#8217;re manually supporting an experience in order to understand what the same experience would look like and how it would perform if it were built.</p></li><li><p>Imagine you&#8217;re trying to understand if it&#8217;s worth building a recommendations feature for your product. In this design, you&#8217;d manually curate recommendations for a small number of users yourself. This would teach you both how to best make recommendations as well as the user impact of having recommendations in their experience.</p></li><li><p><strong>Best suited for: </strong>Validating if something is worth building.</p></li><li><p><strong>Downsides: </strong>You&#8217;re generally limited in your sample size since you&#8217;re manually supporting this experiment. </p></li></ol></li><li><p><strong>A/B/Multivariate Testing</strong></p><ol><li><p>These are scenarios where you&#8217;re running one or more versions of the same thing against one another to determine which one performs best. Within this format there are two main kinds of design:</p><ol><li><p><strong>Big swing experimentation</strong></p><ol><li><p>Here you&#8217;re testing out a completely different concept from the existing implementation. </p></li><li><p>Imagine you have an onboarding experience that has mediocre performance. In this design, you&#8217;re coming up with a completely different approach for your onboarding experience and testing it in parallel with your current onboarding flow. This allows you to try to establish a higher floor of performance than what you previously had. You won&#8217;t know exactly why it&#8217;s better (because you&#8217;re essentially testing a totally new version and can&#8217;t measure what about it made it better) but better is better so you&#8217;ll take it. </p></li><li><p><strong>Best suited for: </strong>Cases where you feel like a big pivot is needed but aren&#8217;t ready to replace what you had.</p></li><li><p><strong>Downsides: </strong>Your new version has a tendency to underperform if it&#8217;s going up against something that has been deeply optimized. You won&#8217;t learn why you got better/worse.</p></li></ol></li><li><p><strong>Variable based optimization</strong></p><ol><li><p>Here you&#8217;re focusing on a key variable, changing it, and testing the changed version against your existing version. </p></li><li><p>Imagine you have something that works pretty well but you feel like it can be better. This approach allows you to try to improve but still know exactly what led to the improvement by focusing changes on only a single variable.</p></li><li><p><strong>Best suited for: </strong>Optimizing something that&#8217;s already working.</p></li><li><p><strong>Downsides: </strong>If you&#8217;re trying to quickly improve something this generally won&#8217;t be an effective approach as it&#8217;s highly focused on a single variable and it can take many changes to improve significantly.</p></li></ol></li></ol></li></ol></li><li><p><strong>Buffet Style Experiment</strong></p><ol><li><p>Here you&#8217;re presenting the customer with lots of options in a cohesive way and seeing what options they pick.</p></li><li><p>Imagine you have multiple filters you&#8217;re thinking of building but you&#8217;re trying to figure out which one should be built first. In this experiment design, you are finding a way to expose the customer to all the options you&#8217;re thinking about and see which ones they gravitate to the most. You&#8217;ll know based on the engagement with the selection which ones are most immediately compelling to the user.</p></li><li><p><strong>Best suited for: </strong>Figuring out what things users gravitate towards. </p></li><li><p><strong>Downsides: </strong>It&#8217;s very easy to structure the experiment in a way where there&#8217;s a bias towards one of the options (leading customers to gravitate towards it0 which can mislead you. </p></li></ol></li></ol><p></p><h4>Experiment Duration:</h4><p>When it comes to experiment duration the key questions to consider are:</p><ul><li><p>What is the length of time that it would require for all users in my experiment to complete the experience/experiment path? </p></li><li><p>Given that the necessary time has elapsed, have I achieved meaningful results?</p><ul><li><p>For A/B types of experiments, you can use <a href="https://www.surveymonkey.com/mp/ab-testing-significance-calculator/">SurveyMonkey&#8217;s</a> statistical significance calculator to determine if your results are significant. </p></li><li><p>Statistical significance will tell you that the results you&#8217;ve experienced with your sample group are likely to be experienced by the full population if that population were exposed to the same conditions.</p></li><li><p>For other kinds of experiments if you&#8217;ve built the sample correctly then you should be able to assume that the results you&#8217;ve seen with your experiment are representative of the results that you&#8217;d see with the full population.<br></p></li></ul></li></ul><p>If you don&#8217;t have clear results then you can either add more people to the experiment or run the experiment for longer. In cases where you have results that don&#8217;t seem better or worse than your baseline, you may be able to conclude that the new version is not better and focus on coming up with new experiments.</p><p></p><h4>Experiment Medium and Fidelity:</h4><p>For experiments around features, it can sometimes feel like you&#8217;re blocked from experimenting because you&#8217;re unable to deploy experiments in the same medium or with the same level of fidelity as your ultimate feature idea.</p><p>In many cases, you might be able to successfully experiment using other mediums (email being one of my favorites) you just have to ensure that it&#8217;s a reasonable assumption that the behavior would translate across mediums (ex: people would engage with this email in a similar way as they would engage with the in-product version of this feature). </p><p>For in-product experimentation, consider investing in feature flagging or other similar techniques to be capable of experimenting with select groups of users. In my experience, these capabilities have significantly increased the range of experiments available and have yielded the most accurate results since they allow you to get in-product engagement. </p><p>In terms of development try to have higher fidelity on the customer-facing portion of the experiment and hack together anything else that&#8217;s not visible (might as well because it&#8217;ll be cheaper and if the experiment doesn&#8217;t pan out you&#8217;ll likely have to throw it away).</p><p>Ultimately, it&#8217;s important to note that the experiment is a good enough proxy for a potential feature or process. It won&#8217;t ever perform exactly like the real thing, but it will ideally give you insight into how the real thing may behave if built.</p><div><hr></div><h3><strong>The tracking and measurement of the experiment: </strong></h3><p>Sometimes in the excitement to try things we forget that without the right data at the end of the experiment, we won&#8217;t be able to take anything from it. </p><p>Before the experiment starts you should have a clear understanding of:</p><ol><li><p>Key metrics and how to track them</p></li><li><p>What success and failure look like</p></li></ol><p></p><h4>Key metrics and how to track them</h4><p>The experiment is essentially an exercise to extract data that is reliable for analysis and drawing conclusions. Without the data, the experiment is essentially throwaway.&nbsp;</p><p>Generally, in terms of metrics, I like to have:</p><ul><li><p>One or two key metrics that are the clear guide for the experiment&#8217;s success or failure.</p></li><li><p>A few additional metrics that are interesting and could affect my understanding of the results. </p></li></ul><p>These metrics should all be retrievable by you at the point where the experiment is finished.&nbsp;This means that before starting you should know exactly how you will retrieve these metrics. </p><p>The key is to set up your data analysis process (how you will acquire those metrics and analyze them) before the experiment starts and not during or after (can&#8217;t stress this enough). Every time I&#8217;ve neglected this rule I&#8217;ve found myself missing key data points in my analysis and had to do further experimentation. <br></p><p><strong>Quantitative + Qualitative</strong></p><p>Many times I like to couple my quantitative analysis with qualitative analysis for a more well-rounded picture. Here you can often include surveys as part of your experiment (at key points that don&#8217;t disrupt the experiment) and schedule interviews/sessions with customers to further discuss their experience with the experiment. If you have session tracking tools, those can be very valuable for understanding user behavior in your experiment (although nothing beats getting feedback from customers directly).<br></p><h4>What success and failure look like </h4><p>It&#8217;s always valuable to determine what success and failure look like for your experiment as well as determining if you&#8217;d be happy with the success state. This success/failure should be based on the key metric(s) you&#8217;ve selected and be something tangible like hitting a certain threshold on that metric(s). </p><p>Once you actually run the experiment, coming back to this success and failure state can be enlightening. Evaluating your results against your initial expectations can help clarify what you&#8217;ve learned and how your expectations have changed since running the experiment. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading this post! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[How to: Get buy-in for your experiment]]></title><description><![CDATA[Part 2 of a four part series on how to identify, sell, run and report on experiments]]></description><link>https://www.rewritetheroadmap.com/p/how-to-get-buy-in-for-your-experiment</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/how-to-get-buy-in-for-your-experiment</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Tue, 10 Oct 2023 10:14:36 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/907459f1-b9ce-4c18-882d-b510b4983d2c_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<ol><li><p><a href="https://www.rewritetheroadmap.com/p/identify-experiment-opportunities">How to: Identify a good experiment opportunity</a></p></li><li><p><strong><a href="https://www.rewritetheroadmap.com/p/how-to-get-buy-in-for-your-experiment">How to: Get buy-in for your experiment</a> &#8592; You&#8217;re Here</strong></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-run-an-experiment">How to: Run an experiment</a> </p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-report-your-experiments-findings">How to: Report your experiment&#8217;s findings</a></p></li></ol><div><hr></div><h2>Why some organizations are averse to experimentation:</h2><p>You might have had the luck to work at an organization that doesn&#8217;t see experimentation as a valuable activity (you might be working there now). This write-up intends to provide some tools for breaking through that aversion and getting your experiments off the ground.</p><p>If you work at an experimentation-averse organization, the reasoning against experimentation sounds a bit like this:</p><blockquote><p><em>&#8220;It&#8217;s too hard and too time consuming.&#8221;</em></p><p><em>&#8220;We already know the answer, why do we need to validate something we already know?&#8221;</em></p><p><em>&#8220;Experimenting is going to take away time from more important things that we need to focus on.&#8221;</em></p><p><em>&#8220;We need to move quickly and experimenting is going to slow us down.&#8221;</em></p><p><em>&#8220;This is going to create a poor user experience for our users.&#8221;<br><br>&#8221;We don&#8217;t have time for this.&#8221;</em></p></blockquote><p></p><p>While the reasoning may sound different from situation to situation, it often translates to the same idea: </p><ul><li><p>The organization believes that its idea-to-value flow is already effectively yielding results. As a result, the organization doesn&#8217;t see a reason to introduce more mechanics to this flow.</p></li></ul><p>While in some cases product managers and other individuals can &#8220;boil the ocean&#8221; with their research and experimentation, in most cases experimentation aversion stems from a lack of exposure to effective experimentation and a sense of security towards the company&#8217;s past way of working. </p><p>To get past this aversion, experimentation needs to be demystified and better understood. </p><div><hr></div><h2>Opening the door to experimentation:</h2><p>In my experience, there&#8217;s always a gatekeeper for experimentation. They can be someone who&#8217;s directly managing you and is accountable for your output and time or they can be someone who is the owner of a key tool you need to leverage in your experiment. </p><p>In order to get experimentation on the map you need to convince this person that your effort is worthwhile while simultaneously aligning on a reasonable cost investment for your experiment. The following are a few things to keep in mind when having these conversations:<br></p><ol><li><p><strong>Make sure your reasoning for experimentation is rock solid:</strong></p><ol><li><p>The key thing going into any conversation is to have clear-cut reasoning for the experiment. This means a few things:</p><ol><li><p>You have a deep understanding of the issue you&#8217;re looking to experiment on but you&#8217;re blocked from fully understanding the issue due to missing key insights. </p></li><li><p>You&#8217;ve attempted to gain those insights via other suitable means and hit a wall and the experiment has a high chance of unlocking those key insights.</p></li><li><p>Having those insights will improve your chances of future success or mitigate significant risk. </p></li><li><p>You have a clear idea of how the experiment would work that is reasonable in terms of cost to setup and duration.</p></li></ol></li><li><p>If the above items are covered then you&#8217;re starting your conversation from a position of strength. </p></li></ol></li><li><p><strong>Make sure to explain your reasoning for conducting an experiment:</strong></p><ol><li><p>In many cases, there&#8217;s an asymmetry of information between you and the gatekeeper when it comes to your work. While you&#8217;re spending considerable time in the weeds they might be highly focused on the &#8220;bigger picture&#8221; or overall not invested in your area of work. </p></li><li><p>It&#8217;s then important to establish a shared foundation of knowledge so that the gatekeeper understands why you ended up deciding an experiment was necessary. </p></li><li><p>Once that shared knowledge exists, it&#8217;s then time to explain why an experiment would be worthwhile (what risk you&#8217;re mitigating for, what upside you might verify, or overall what value you&#8217;re expecting from an experiment). </p></li><li><p>Ideally, by doing this, you&#8217;ll have built enough buy-in to experiment, but some situations require further handling and convincing. In those cases, you need to be more thoughtful about the gatekeeper's concerns and philosophical disagreements with your reasoning.</p></li></ol></li><li><p><strong>Minimize cost/loss-of-focus concerns by addressing them head-on:</strong></p><ol><li><p>Many organizations opt out of doing experimentation because they believe that the cost of setup for an experiment is too high. This leads to people opting to execute the project version of the experiment instead. </p></li><li><p>There is an obvious paradox here, the project version of an experiment is, almost always, significantly more expensive to execute. What organizations really mean here is that running an experiment has costs outside of the costs they generally pay and requires effort outside of the effort they generally exercise. </p></li><li><p>Here you have to align with the gatekeeper on what an acceptable investment looks like (in addition to the potential value and reasoning) since this will be an investment outside of the ones they&#8217;re used to making. If the gatekeeper believes you won&#8217;t over-invest they will often be open to the experiment (even if it&#8217;s just to appease you). </p></li><li><p>Think about this as a resource allocation problem for the gatekeeper, and work with them to identify the acceptable time and capacity allotment for the experiment. If the gatekeeper understands you&#8217;re being mindful it&#8217;ll make them much more open to trying things out.</p></li></ol></li><li><p><strong>Don&#8217;t make it a battle of beliefs. Instead, focus on the upside/downside of the experiment:</strong></p><ol><li><p>Many times the reasoning for not running an experiment ends up being about the gatekeeper&#8217;s belief around a key variable or idea the experiment is looking to validate. If the belief is not based on fact, and you can&#8217;t get the gatekeeper to realize that, then you might need to pivot away from beliefs and focus instead on the upside/downside of the experiment. </p></li><li><p>Here are a few pathways for discussion that can be more productive:</p><ol><li><p>In cases where you&#8217;re experimenting to validate a feature. Focus on how much can be learned via the experiment and how expensive and risky it would be to only learn as a result of a release. The experiment may help you minimize scope, hone in on the most valuable aspects of the feature, and prevent you from having to do follow-up work. </p></li></ol></li><li><p>In cases where the experiment is only seeking to increase your knowledge base, identify the possibilities that could come out of the insight and use those possibilities to help the gatekeeper open their mind to the experiment. Most people love discovering new things especially if those new things can help create more value, key in on that value as fuel for your experiment.</p></li><li><p>In cases where there&#8217;s risk, focus on helping the gatekeeper create a clear idea of the bad things that can happen if you move forward without experimenting. Make this risk very tangible to them so that they can imagine the world where the risk is realized. From here you can use experimentation as the risk mitigation tool that will help them prevent that risk from realizing.</p></li></ol></li><li><p><strong>Get other members of the team to lobby for the experiment:</strong></p><ol><li><p>If other stakeholders of the team also believe in the experiment ask them to voice this to the gatekeeper. Having multiple voices echoing the same ideas can have a compelling effect on the gatekeeper&#8217;s thinking about the experiment and its value.</p></li><li><p>Keep in mind that you&#8217;re not the sole owner of the experiment, the gatekeeper and the team working on the experiment are also owners of that work. If everyone rows together with conviction you&#8217;ll find that a lot of the friction in this process will start to dissipate over time.<br></p></li></ol></li></ol><p><strong>Pick your battles:</strong></p><p>In some cases, it might not be worth spending your equity on pushing experimentation. This is especially the case when the learning is low-upside and you&#8217;re still on shaky ground with experimentation as a concept within your org.</p><p>On the flip side, if you have something you really believe in, something that&#8217;s very polarizing in the organization, or something that feels very risky to the team then those tend to be good areas to invest in since you&#8217;re likelier to uncover insights and as a result validate the value of experimentation.</p><p>Being thoughtful about when to use experimentation can sometimes be the most valuable thing to counter aversion. Experiments with pay-offs are ultimately the best path towards combating aversion as they can be used as proof that experimentation is valuable.</p><div><hr></div><h2>Asking for forgiveness instead of permission: </h2><p>It needs to be mentioned that there are pathways to experimentation where permission isn&#8217;t necessarily granted. Sometimes organizations work themselves into a dysfunctional workflow that prevents them from unlocking further value and become unable to improve on that workflow. In a few rare cases, learning the truth about something that is holding the company back is worth crossing a line. That being said experimentation with buy-in is always the best version of this work and the goal should be to make experimentation an aligned approach.  </p><p>There are some key things here to keep in mind / avoid when swimming in these murky waters:</p><ul><li><p>Don&#8217;t run an experiment for something you&#8217;ve been explicitly denied permission for (changing a small variable on your original plan does not constitute a totally different experiment either). </p><ul><li><p>You&#8217;re just going to piss people off and it&#8217;s kind of an asshole move. The upside is seldom justified and the downside is massive.</p></li></ul></li><li><p>Don&#8217;t do it as a practice since you&#8217;ll just get permanently shut down.</p><ul><li><p>The undercover approach should be a rare occurrence, a card you can play when you really need it. If you keep doing it, again, you&#8217;re going to piss people off and come off as an asshole.</p></li></ul></li><li><p>Don&#8217;t do it if there&#8217;s a real negative downside.</p><ul><li><p>If you&#8217;re affecting customers negatively, if you&#8217;re putting another team in a bad place, if there are real negative consequences to your experiment then it&#8217;s probably not the right path.</p></li></ul></li><li><p>Don&#8217;t talk about it constantly, don&#8217;t formalize it.</p><ul><li><p>If you make a big fuss over the thing&#8230; You&#8217;re probably going to get the thing shut down. </p></li></ul></li><li><p>Don&#8217;t overinvest.</p><ul><li><p>What&#8217;s worse than doing something you shouldn&#8217;t be doing? Doing it a lot. Think about this experiment as low-stakes research that you&#8217;re doing on the side, not as your big project.</p></li></ul></li><li><p>Know what you&#8217;re doing:</p><ul><li><p>If you don&#8217;t understand your business, customers, teams, and strategy then don&#8217;t do it, you&#8217;re very likely to mess things up for everyone. Do the work before crossing the experimentation line.</p></li></ul><p></p></li></ul><p>Again, a bought-in group is always better than a group where everyone is going their own way. While experimenting without consent can yield results, it&#8217;ll almost never yield buy-in. Make people stakeholders and share the wins, this is most often the path toward an experimentation-focused culture.</p><div><hr></div><h2>Lexicon for convincing people to experiment:</h2><p>I often find that a concept is only useful if I have the words to express it. With that in mind, I wanted to share some of my favorite lines when talking about experimentation and what they intend to achieve. Hopefully, this helps you build your own language and script for these kinds of conversations:</p><ul><li><p><em>&#8220;While we have a good idea that this is true, an experiment will help us learn more about how to build it effectively saving us from having to do more iterations after launch.&#8221;</em></p><ul><li><p><strong>Why it works: </strong>You&#8217;re moving the conversation away from the assumption and instead focusing on adding value (by better understanding how to solve the problem and doing it in a cheaper way than releasing).</p></li></ul></li><li><p><em>&#8220;I know that you&#8217;re not as high on this as I am and that to you this feels like a bit of a waste of time. Is there an amount of time I could invest into this that you feel would still be ok?&#8221;</em></p><ul><li><p><strong>Why it works: </strong>You&#8217;re disarming the gatekeeper by addressing a big concern head-on (by doing this giving them confidence you&#8217;re going to be thoughtful about this initiative). You&#8217;re then collaborating with them to find a suitable middle ground and asking them to problem-solve with you vs approve/deny you.</p></li></ul></li><li><p><em>&#8220;If we&#8217;re wrong about this, why do you think that would be the case? What&#8217;s the worst scenario for this? What could we have missed? Would we be open to testing that to be sure?&#8221;</em></p><ul><li><p><strong>Why it works: </strong>You&#8217;re pivoting the conversation to focus on the downside instead of assumptions and using the feelings/thoughts tied to the downside to help fuel the desire to experiment.</p></li></ul></li><li><p><em>&#8220;I really believe that I can find some valuable learnings here and minimize some potential risks&#8230; I&#8217;d like to spend some time on it unless you&#8217;re opposed to it?</em></p><ul><li><p><strong>Why it works: </strong>You&#8217;re making your belief and passion explicit and essentially asking the gatekeeper for an opportunity to pursue this passion. They may be inclined to allow you because they know you&#8217;ll be more bought in as a result of having been allowed to do this.</p></li></ul></li><li><p><em>&#8220;The thing that makes me want to do this isn&#8217;t necessarily just getting to this answer, but it&#8217;s instead what this answer unlocks. Think about what we could do if we validate X&#8230; We could do this thing, that thing, maybe even this other thing&#8230;&#8221;</em></p><ul><li><p><strong>Why it works: </strong>In cases where the key insight isn&#8217;t as interesting to the gatekeeper but it is to you, focusing on what it unlocks helps the gatekeeper understand why you&#8217;re excited. Making the possibilities more tangible can help the gatekeeper see your experiment in a totally new way and as a result, be more bought in. </p></li></ul></li></ul><div><hr></div><h2>Parting thoughts on getting buy-in:</h2><ul><li><p>I&#8217;ve found most people are excited about uncovering new insights, being proven right/wrong, and preventing mistakes from happening. The experimentation process provides that potential for excitement and it&#8217;s partly your job to foster enthusiasm around that process.</p></li><li><p>Cultivate excitement with those that are experiment averse and you&#8217;ll sometimes find that they&#8217;ll become advocates of experimentation over time. With enough successful experimentation, the culture of the organization might start leaning towards experimentation.</p></li><li><p>While the road to get buy-in might be hard, it wouldn&#8217;t have been possible without the gatekeeper&#8217;s approval. Remember to always validate the investment the gatekeeper has made in enabling you to experiment as this will help them feel ownership of experimentation and can help turn them into a long-term believer. </p></li><li><p>Finally, remember that most people are trying to do their best for the customer/product/company/team. Think about gatekeepers as partners and you&#8217;ll find ways to grow that partnership. Think about gatekeepers as blockers and you&#8217;ll often find yourself running into their walls.</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to: Identify a good experiment opportunity]]></title><description><![CDATA[Part 1 of a four part series on how to identify, sell, run and report on experiments]]></description><link>https://www.rewritetheroadmap.com/p/identify-experiment-opportunities</link><guid isPermaLink="false">https://www.rewritetheroadmap.com/p/identify-experiment-opportunities</guid><dc:creator><![CDATA[Luis Camara]]></dc:creator><pubDate>Wed, 04 Oct 2023 11:06:04 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0bbfc363-c5ce-4a4d-80b9-34b724b871c1_1456x1048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Experiments can establish the foundation of new processes, features, and ways of thinking as well as save companies from making bad decisions. This series intends to provide a detailed guide on experimentation. It will likely be most helpful to people who are just starting to introduce experiments into their workflow.</p><ol><li><p><strong><a href="https://www.rewritetheroadmap.com/p/identify-experiment-opportunities">How to: Identify a good experiment opportunity</a>  &#8592; You&#8217;re Here</strong></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-get-buy-in-for-your-experiment">How to: Get buy-in for your experiment</a></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-run-an-experiment">How to: Run an experiment</a></p></li><li><p><a href="https://www.rewritetheroadmap.com/p/how-to-report-your-experiments-findings">How to: Report your experiment&#8217;s findings</a></p></li></ol><div><hr></div><h2>Why experimentation matters</h2><p>It&#8217;s easy to confuse facts with beliefs because both of them feel the same and sound the same. However, while facts are by definition always true, beliefs are only sometimes true.</p><p>The ability to discern fact from belief is then a key skill to uncovering the world&#8217;s truth and as a result, ensure your product matches that truth.</p><p>Experimentation is one of the key tools that allows you to explore and map that distinction. It is, in essence, a tool for fact-finding. It is also, in the world of product development, generally a lower cost, lower risk, and lower commitment option than its alternatives.  </p><div><hr></div><h2>A general framework for what experiments are good/bad for:</h2><p>It&#8217;s important to establish that experiments aren&#8217;t the perfect tool for every job. Experiments are very well suited for some things but poorly suited for others. </p><p>The following is a list of the different things experiments are well and poorly suited for: </p><ol><li><p><strong>Well Suited: </strong></p><ol><li><p>Simulating how a feature may perform if built.</p></li><li><p>Iterating on a pre-existing solution.</p></li><li><p>Mapping customer behavior for a given scenario.</p></li><li><p>Exploring the productization of an insight (along with testing that insight&#8217;s validity).</p></li><li><p>Validating if a certain product or process can be removed.</p></li></ol></li><li><p><strong>Poorly Suited:</strong></p><ol><li><p>Solving a business problem long term. </p><ol><li><p>Ultimately you&#8217;ll need a product/process to be created to fully solve most problems. </p></li></ol></li><li><p>Understanding users, markets, behaviors, and journeys (in some cases).</p><ol><li><p>Generally, you&#8217;re better off conducting user research than using experimentation to validate certain kinds of insights. </p></li><li><p>Additionally, if you haven&#8217;t looked at data and/or talked to customers then you might be skipping a step by running the experiment.</p></li></ol></li><li><p>Simulating scenarios that are very long-term, intangible, or high risk to the company:</p><ol><li><p>Certain scenarios are not well suited for experimentation, if there&#8217;s considerable risk to your company, if the insight requires a lot of time to validate or if you&#8217;re trying to test big intangible decisions like partnering/acquiring/merging with another company then experimentation is likely a bad tool.</p></li></ol></li></ol></li></ol><p>If your experiment idea falls under the &#8220;bad&#8221; list then ask yourself: 1. What is the goal for the experiment 2. Is an experiment the best way to get there? </p><p>Additionally, I want to stress that it&#8217;s worth experimenting when the insight is worth knowing in the first place. If the insight is trivial or can otherwise be learned in easier ways (like data analysis, customer interviews, usability testing, etc) then experimentation may not be worth doing in the first place. This seems obvious but it can be hard in practice to make this deliberation</p><div><hr></div><h3><strong>A strong hypothesis</strong></h3><p>This is arguably the murkiest part of experimentation because different people can have different opinions on what constitutes a strong hypothesis. I like to think that a strong hypothesis is a hypothesis that:</p><ol><li><p><strong>Is logical</strong> and at least a few others would think your premise has validity.</p></li><li><p>Is based on <strong>first-hand experience or second-hand reporting from customers </strong>of the conditions, customers, and environments at play. </p></li><li><p>There&#8217;s a reason <strong>confirming/denying the hypothesis is valuable</strong>, knowing this thing will give you the necessary information to make an important decision.<br></p></li></ol><h4>Types of Hypothesis:</h4><p>This is not an extensive list of the types of hypotheses but some of the more common patterns that come to mind:</p><ol><li><p>X happens because of Y</p></li><li><p>When exposed to multiple options, customers will pick X</p></li><li><p>The outcome is better if x is changed</p></li><li><p>What if X existed/no longer existed?</p></li></ol><p></p><h5><strong>X happens because of Y</strong></h5><p><em>Example: People buy the Caesar salad because it comes with croutons.</em> </p><ul><li><p>In this context, the experiment is the method by which you validate that x does happen because of y and to what degree y can impact x. </p></li><li><p>Notice that I used a simple relation of cause and effect, this is because the majority of the time multiple relationships of cause and effect are very hard to test in an experiment. </p><ul><li><p>If you&#8217;re trying to prove multiple relationships most of the time you&#8217;re better off designing multiple experiments and performing those independently as this will give you the best odds of getting valuable data from your efforts. </p></li></ul></li></ul><p></p><h5><strong>When exposed to multiple options, customers pick x:</strong></h5><p><em>Example: If provided with multiple kinds of salads to pick from people would pick Caesar salad.</em></p><ul><li><p>In this scenario we&#8217;re essentially seeing what happens when we expose customers to multiple competing variables and we&#8217;re measuring which variables customers gravitate towards. </p></li><li><p>This is a tricky type of hypothesis because the results are very experiment design dependent. It&#8217;s easy to tip the scales towards one variable or another. However, in my experience, it&#8217;s a very helpful question to ask that if validated allows the business to properly weigh the importance of different variables in their product.</p></li><li><p>It&#8217;s also important to note that most times the experiment won&#8217;t give you the answers on why customers pick x, it&#8217;ll just tell you that they will. You&#8217;ll have to conduct research to truly understand the why.</p></li></ul><p></p><h5><strong>The outcome is better if x is changed:</strong></h5><p><em>Example: We&#8217;d sell more Caesar salads if they came with smoked salmon.</em> </p><ul><li><p>This is a classic optimization hypothesis here you&#8217;re just testing that a change to the product would lead to improved results. </p></li><li><p>Again, valuable question to ask that is very dependent on the execution but when approached correctly can lead to real improvements in the product experience.</p></li><li><p>The size of X is also variable. Many times people think of optimization as only being able to change a single variable to avoid inconclusive results. In some cases, if your performance is poor you&#8217;re better off experimenting with a completely different version of your experience to try to establish a new floor of performance that&#8217;s higher than your previous experience. While you won&#8217;t know exactly why the new version is better, in some cases, you&#8217;re better off optimizing something that performs well than something that performs poorly.</p></li></ul><p></p><h5><strong>What if x existed / no longer existed:</strong></h5><p><em>Examples: What if we sold Cobb salad? / What if we stopped selling Caesar salad? </em></p><ul><li><p>If you&#8217;re thinking about building a new feature then an experiment can be a very valuable way to determine if you should follow through with your plan. </p><ul><li><p>In this scenario, the experiment is then attempting to measure/understand how customers would interact with that feature if it was built by creating a simulated version of that feature.  </p></li></ul></li><li><p>On the other hand, experiments are also really useful to prove that you can sunset features that you believe are no longer viable.</p><ul><li><p>In this scenario, you&#8217;re in most cases hiding or removing the feature from some/all customers and measuring the results (seeing if feedback comes in, seeing if there&#8217;s a change in engagement, etc). This de-risks sunsetting features significantly.</p></li></ul><p></p></li></ul><div><hr></div><h2>Following your gut (or someone else&#8217;s)</h2><p>I tend to find that good experiments are driven by very strong feelings or very negative feelings.</p><p>If I have very strong feelings about something it tends to be worth digging further. This isn&#8217;t because I&#8217;m some sort of savant but because strong feelings tend to be indicators of somethingness. What that somethingness is or if it&#8217;s actually there is unknown, the experiment will bring that out. </p><p>That somethingness many times is a set of patterns, behaviors, trends, or mechanics that we&#8217;re unable to quantify or explain at the time but that can eventually form the basis of something in the business. </p><p>Strong feeling driven experimentation can often yield positive outcomes in one way or another:</p><ul><li><p>When you&#8217;re validated in your feelings you can add more ammo to whatever case you&#8217;re trying to make.</p></li><li><p>When the results are negative you&#8217;re then faced with confronting the reality of the outcome and learning from it. Many times these poor outcomes lead to important changes in perspective and direction.<br></p></li></ul><p>While positive strong feelings about something are valuable, negative strong feelings are just as valuable. Annoyance, frustration, impatience, insecurity, fear are all great indicators of somethingness. People can be very right but they can also be very wrong about things, flipping your mindset to take advantage of both your beliefs and anti-beliefs is an easy lever to increase the amount of opportunities you&#8217;re identifying.</p><p></p><h4>Leveraging other people&#8217;s gut feelings</h4><p>All of this goes for people you trust, I find that in any team there are a few people who tend to have a general prescience about things. Those people&#8217;s feelings can also act as a compass for the direction to focus on. It&#8217;s important also to factor in the different exposure areas for different people on your team, a designer, a developer, a customer success specialist, a sales person with the prescience trait will often have feelings about very different areas of the business or the experience. Actively seeking to hear their opinions and thoughts about the product and business can often yield valuable opportunities. </p><p>It&#8217;s also important to highlight that things that many people consistently believe have a fun tendency to be wrong (often very wrong). This is because shared beliefs are often built on a tradition of knowledge and that tradition of knowledge is often based on opinion. These are very good experiment areas because they are often located at core areas of knowledge of the business and identifying misalignments in those areas can help the business find the things they should actually believe. </p><p>I have found that central ideas to mine and other people&#8217;s decision making were in some cases inaccurate or generally incomplete. They made sense conceptually and felt completely rational and beyond reproach and as a result we assumed they were self-evident. When you get a big enough idea like that wrong, you can spend months or even years going down less valuable paths. To find these I ask myself questions like: &#8220;What would really suck if we were totally wrong on&#8221;. This often yields the highest value areas to validate.<br></p><blockquote><p><strong>Story time: </strong></p><p>For a period of about 2 years I was convinced that there was a pricing problem with a product I was managing and I had a strong belief that our pricing was a key blocker in our growth. </p><p>I debated and lobbied for an experiment to test this assumption and eventually was granted the opportunity. We found a clear way to test multiple different pricing options (all lower in different ways) in a cohesive experience to the customer. This was a several month experiment measuring short term conversion rate and long term retention.</p><p>When the results finally materialized the original pricing I had been so strongly against was by far the best option far outperforming everything else both in conversion rate and retention. </p><p>After a good amount of analysis trying to disprove this reality I settled on a new reality - pricing wasn&#8217;t the problem. This insight was fundamental for me and others to move forward in new directions which did prove fruitful.</p><p>Had we changed pricing from the start we would have likely taken significant losses and learnt that lesson the hard way. Had I not done the experiment I&#8217;d probably still be arguing for it today convinced pricing was the blocker for the growth I was seeking. </p></blockquote><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.rewritetheroadmap.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Rewrite the Roadmap! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>