The problem: Jira priorities are opinions

Jira gives you a Priority field out of the box — Highest, High, Medium, Low, Lowest. In most teams, 70% of the backlog is "High" within a year. It's not that everything is genuinely high priority. It's that the field is subjective, so people default upward.

The consequence: sprint planning becomes a negotiation. The loudest stakeholder wins. Engineers don't know which ticket to pull next when two "High" issues are waiting. And product managers can't explain — to their team or their leadership — why one feature ships before another.

RICE scoring fixes this by making the trade-offs explicit. Two features can both matter and both be urgent. RICE forces you to say which one matters more, and to show your work.

Quick answer: Jira does not include RICE scoring natively. You add it via a Marketplace app that creates custom fields — sliders for each dimension, automatic score calculation, and a ranking table across your backlog. Setup takes under 10 minutes.

The RICE formula explained

RICE was developed by Sean McBride at Intercom as a way to standardize feature prioritization across a growing product team. The formula has four factors:

RICE Score Formula
RICE Score  =  Reach  ×  Impact  ×  Confidence Effort
R
Reach
Users affected per quarter
I
Impact
Effect per user (0.25–3)
C
Confidence
Certainty (10–100%)
E
Effort
Person-months of work

Reach — how many people benefit?

Reach is the number of users or customers who will be affected by this feature within a defined time period — typically one quarter. Use real numbers, not percentages. "500 users per quarter" is better than "medium reach."

The key is consistency: all issues in your backlog must use the same time window. Mixing quarterly and annual estimates will skew every score in your ranking.

Impact — how much does it move the needle?

Impact is a multiplier that represents the effect on each individual user. Intercom defined a standard scale to keep estimates comparable across features:

Impact label Multiplier When to use
MassiveCore workflow change — users will notice it daily
HighSignificant improvement — noticeably faster or easier
MediumUseful addition — nice to have, some will use it
Low0.5×Minor improvement — affects edge cases or power users
Minimal0.25×Negligible effect on most users

Impact is the most subjective dimension. The multiplier scale helps — it forces you to commit to "this is 2× better than average" rather than vague labels like "important."

Confidence — how sure are you?

Confidence is expressed as a percentage and reflects how well-supported your estimates are. Intercom's original guidance suggested three levels: 100% (high confidence — you have data), 80% (medium — some data, some assumption), 50% (low — mostly assumption). Values below 50% are valid but signal that the issue needs more discovery before scoring reliably.

Confidence acts as a risk discount. A feature with a fantastic RICE score at 50% confidence will rank lower than a more modest feature at 100% confidence — which is exactly what you want. Don't inflate confidence to make a pet feature look good.

Effort — how much does it cost?

Effort is the total time all team members will spend on this issue, expressed in person-months. A frontend engineer and a backend engineer working for two weeks each = 1 person-month. Include QA, design, and PM time — not just engineering.

Effort is the denominator: a high-effort issue is penalized even if it has excellent reach and impact. This is RICE's key insight — it's not just about value delivered, but value per unit of work.

A worked example

Let's score two features competing for the same sprint slot:

Feature A — Bulk email export
Reach400 users/quarter
Impact1 (Medium)
Confidence80%
Effort2 person-months
RICE Score = (400 × 1 × 0.8) ÷ 2 160
Feature B — Advanced filter in settings
Reach80 users/quarter
Impact2 (High)
Confidence100%
Effort0.5 person-months
RICE Score = (80 × 2 × 1.0) ÷ 0.5 320

Feature B ranks higher — despite being wanted by far fewer users — because it delivers twice the impact per user, costs a quarter of the effort, and you're certain it will work. RICE surfaces this trade-off cleanly where a pure "Priority: High" label on both would hide it entirely.

Common mistake: Mixing time windows for Reach. If one issue uses "annual users" and another uses "monthly active users," the scores are not comparable. Decide on a time window (typically one quarter) and use it for every issue in the backlog.

When RICE is the right fit — and when it isn't

RICE works best when your team can estimate reach with real data — user analytics, support ticket volume, interview counts. It rewards features that affect many people, which aligns well with product teams working on consumer-facing or broadly-used software.

RICE is less suited for teams where:

For teams unsure which framework fits, see the comparison guide: RICE vs. WSJF vs. ICE — which Jira framework is right for you?

How to set up RICE scoring in Jira

The steps below use Priority Scoring for Jira — a Forge-native app that adds RICE, WSJF, and ICE as native custom fields. It runs entirely on Atlassian's infrastructure ("Runs on Atlassian" badge), so no data leaves your Jira site.

  1. 1
    Install Priority Scoring from the Atlassian Marketplace. Search for "Priority Scoring" and install it to your Jira Cloud site. A free trial is available with all features included.
  2. 2
    Add the RICE Score field to your project screen. Go to Jira Settings → Work Items → Screens, find the screen for your project, and add the "Priority Scoring — RICE" field. Without this step, the sliders won't appear on issues.
  3. 3
    Open a Jira issue and set your first RICE score. You'll see four sliders — Reach, Impact, Confidence, and Effort. Drag each slider to the appropriate value. The RICE score is calculated live and displayed above the sliders.
  4. 4
    View your ranked backlog in the Priority Scoring dashboard. Open the Priority Scoring global page. Switch to the RICE framework, select your project, and see all scored issues ranked from highest to lowest score — with sprint, due date, and assignee columns.
  5. 5
    (Optional) Configure score thresholds. In Settings, set Low and High thresholds for the RICE framework. Issues below the low threshold display a 🔴 badge; above the high threshold a 🟢. Teams at a glance see which issues are priority without reading individual scores.
Jira issue view showing the Priority Scoring RICE, WSJF, and ICE custom fields with sliders for each dimension and live score calculation

RICE, WSJF, and ICE score fields on a Jira issue — sliders for each dimension, score calculated live

Seeing your ranked backlog

Individual issue scores are useful, but the real value is the ranked view across your entire project. The Priority Scoring dashboard shows a sortable table of all scored issues — rank, RICE score (with color badge), issue key, summary, sprint, assignee, and due date.

Sort descending to work from the top. Issues sitting at the bottom of the ranking — especially those with a 🔴 badge — are the ones to question: should they be in the backlog at all?

Priority Scoring dashboard showing the Priority Ranking table with RICE scores, color-coded badges, sprint and assignee columns, and a Board Health score panel

Priority Scoring dashboard — ranked RICE scores across all backlog issues, with Board Health panel on the right

Going further: AI-assisted scoring with Rovo

Scoring every issue manually is the biggest adoption barrier for RICE. Priority Scoring includes a Rovo AI Agent that reads issue content — summary, description, acceptance criteria — and suggests dimension values with written reasoning.

The agent operates in three modes, configurable in Settings:

To score an issue via Rovo, open a Rovo chat and type: "Suggest a RICE score for PROJ-123." Rovo fetches the issue, reasons through the dimensions, and returns suggested values. In confirm or auto mode, it can then write the score to the field directly.

Bulk scoring is available in auto mode: "Score all unscored issues in project PROJ with RICE." Rovo works through the list up to the configured bulk limit (default: 10 issues per call).

Note on AI suggestions: Rovo's suggestions are a starting point, not a verdict. Impact and Confidence in particular benefit from human review — the AI doesn't know your company's strategic bets or which customer segments matter most this quarter. Use Suggest or Confirm mode until you've calibrated the agent's judgment against your own.

Keeping scores consistent across your team

RICE scoring only works if the team uses the same definitions. A Reach of "1,000 users" means nothing if one person estimates monthly users and another estimates total registered users. Before you start, agree on three things:

The Score Audit Trail in Priority Scoring (Settings → Score Audit Trail) logs who set what value and when, which helps teams spot outliers and recalibrate over time.

Frequently asked questions

What is RICE scoring in Jira?

RICE is a product prioritization framework that scores each backlog item using four factors: Reach (how many users are affected), Impact (how much it moves the needle), Confidence (certainty of estimates), and Effort (total team time). The score formula is (Reach × Impact × Confidence) ÷ Effort. Jira doesn't include RICE natively — you add it with an app like Priority Scoring for Jira.

How do I calculate a RICE score?

Multiply Reach × Impact × Confidence, then divide by Effort. Example: 200 users/quarter × 1 (medium impact) × 0.8 (80% confidence) ÷ 2 person-months = RICE Score of 80. A higher score means higher priority. Compare scores across your backlog — the absolute value doesn't matter, only the ranking.

Does Jira have built-in RICE scoring?

No. Jira Cloud does not include RICE scoring out of the box. You add it through the Atlassian Marketplace. Priority Scoring for Jira adds native RICE custom fields (sliders per dimension, automatic calculation) directly to the issue view — no scripting or custom Automation rules needed.

What is a good RICE score?

RICE scores are relative, not absolute. What matters is the ranking across your backlog, not the number itself. A score of 50 might be top-tier in one backlog and average in another. Use color-coded thresholds (🟢🟡🔴) to highlight the top and bottom of your ranking at a glance — Priority Scoring for Jira lets you configure these thresholds per framework.

What is the difference between RICE, WSJF, and ICE in Jira?

RICE is best for product teams prioritizing features by user impact. WSJF is designed for SAFe engineering orgs and focuses on cost of delay. ICE is a faster, simpler framework for teams that need quick directional scoring without detailed data. All three are supported as native custom fields in Priority Scoring for Jira.

Can I use RICE scoring in Jira without scripting?

Yes. Priority Scoring for Jira adds RICE as a native custom field — no ScriptRunner, no Automation rules, no formulas to maintain. Install the app, add the field to your project screen, and the sliders are immediately available on every issue.

Add RICE scoring to your Jira backlog

Priority Scoring for Jira — native RICE, WSJF, and ICE fields, Rovo AI agent, Board Health score, and Portfolio view. Free trial included.

Try it free on Marketplace →