The ICE framework: Scoring experiment backlog

Mon Jun 23 2025

You know that sinking feeling when you look at your experiment backlog and realize you have 47 ideas but only bandwidth for three? Yeah, we've all been there. The worst part isn't having too many ideas - it's picking the wrong ones and watching three months of work barely move the needle.

That's where the ICE framework comes in handy. Created by Sean Ellis (the growth hacking guy), it's basically a scoring system that helps you quickly figure out which experiments are worth your time. No complex spreadsheets, no three-hour meetings - just a simple way to separate the winners from the time-wasters.

Introducing the ICE framework for experiment prioritization

The ICE framework boils prioritization down to three questions: how much will this move the needle (Impact), how sure are we it'll work (Confidence), and how painful will it be to build (Ease). You score each factor from 1-10, multiply them together, and boom - you've got your priority list.

What makes ICE so popular with product and growth teams is that it's ridiculously fast. You can score a dozen ideas in 20 minutes instead of debating them for hours. It's particularly great for spotting those quick wins - the experiments that'll give you results next week instead of next quarter.

But here's the thing: simplicity cuts both ways. The same speed that makes ICE useful can also make it shallow. You might end up chasing easy wins while bigger, more strategic opportunities sit on the shelf. As teams on Reddit's product management forums point out, ICE works best for relative prioritization - choosing between similar options rather than setting your entire product strategy.

Still, when you need to move fast and pick from a pile of decent ideas, ICE gets the job done. It's especially powerful when you combine it with other methods or use it as a first-pass filter before deeper analysis. Think of it as your quick-and-dirty sorting hat for backlog prioritization.

Applying the ICE framework: Evaluating impact, confidence, and ease

Let's get practical about scoring. For Impact, you're basically asking: if this works perfectly, how many people care? Look at the percentage of users affected, potential revenue lift, or improvement in your north star metric. Itamar Gilad's Confidence Meter suggests backing up your scores with actual evidence - user research beats gut feelings every time.

Confidence is where things get interesting. This isn't about being optimistic; it's about having proof. Pull data from:

  • Previous A/B tests that worked (or failed) in similar ways

  • Customer interviews that validate the problem

  • Competitor results you can benchmark against

  • Input from engineers who've built similar features

The key is being brutally honest. If your confidence is based on "the CEO thinks it's cool," that's probably a 3, not an 8.

Ease is deceptively tricky to score. Sure, changing button text is easier than rebuilding your checkout flow, but you need to think beyond just dev hours. Factor in design time, QA complexity, dependencies on other teams, and whether you'll need to convince legal or compliance. One trick from product managers who actually use ICE: break the project into tasks and count how many people need to say yes.

Tools like Foxly can help standardize your scoring process, but the real secret is getting everyone in the room. Your engineer's "ease" score will be way more accurate than your guess, and your data scientist can reality-check those impact estimates.

Limitations of the ICE framework and when to consider alternatives

Here's where I'll be blunt: ICE can turn your team into feature factories pumping out easy wins while your product strategy goes nowhere. The framework's biggest weakness is that it treats all impact equally - a 10% improvement to a tiny feature scores the same as a 10% improvement to your core experience.

The subjectivity problem is real too. Without clear scoring guidelines, you'll get wildly different numbers from different people. Your optimistic PM gives everything 8s and 9s while your cynical engineer hands out 2s and 3s. Teams on Reddit regularly complain about ICE discussions devolving into opinion battles.

When you need more nuance, frameworks like RICE or DRICE add important dimensions:

  • Reach: How many users actually encounter this?

  • Return on Investment: What's the payoff relative to the cost?

DRICE particularly shines for finding those high-ROI projects that ICE might bury. You know, the ones that take a bit more work but deliver 10x the value.

The dirty secret is that ICE often rewards laziness. Why tackle that complex but game-changing feature when you can ship five button color tests instead? This is fine for optimization teams focused on incremental gains, but if you're trying to build the next big thing, you'll need to look beyond ICE scores.

Enhancing prioritization by integrating ICE with other methods

Smart teams don't use ICE in isolation - they use it as one tool in their prioritization toolkit. Here's what actually works:

Start with ICE for the initial sort, then layer on strategic considerations. Maybe you score everything with ICE first, then ask: which of these top-scoring ideas actually ladder up to our quarterly goals? The RICE and DRICE frameworks mentioned earlier can help add this strategic lens.

Building a data-driven culture makes every framework better, including ICE. When your impact scores come from actual user research instead of boardroom assumptions, the whole system works better. Teams that regularly analyze their experiment results get progressively better at predicting impact and confidence.

Want to level up your ICE game? Try these approaches:

The teams that succeed with ICE treat it as a living system, not a rigid formula. They adapt the scoring based on what they learn, combine it with other methods when needed, and never forget that the framework serves the strategy, not the other way around. When you implement quality checks and experiment scores alongside ICE, you create a more robust prioritization process that balances speed with thoughtfulness.

Closing thoughts

The ICE framework isn't perfect, but it doesn't need to be. It's a starting point for prioritization conversations, not the final word. Use it when you need to move fast, score a bunch of similar ideas, or get everyone on the same page about what matters.

Just remember: the best prioritization happens when you combine ICE's simplicity with strategic thinking, real data, and a healthy dose of skepticism about your own assumptions. Don't let the ease score seduce you into building only the simple stuff.

Want to dig deeper? Check out Statsig's experimentation guides for more on building a culture of testing, or explore how teams use DRICE when ICE isn't cutting it.

Hope you find this useful!



Please select at least one blog to continue.

Recent Posts

We use cookies to ensure you get the best experience on our website.
Privacy Policy