Get a first look at what's coming. The Developer Passport Australia Release Preview kicks off March 12. Dive in! 

amitbaweja
ServiceNow Employee

Introduction

Think about one of the most common questions we hear everyday: "We know AI Agents are the future, but how do we figure out the right use cases for our organisation?"

 

Most teams don't struggle with enthusiasm for agentic AI — they struggle with knowing where to start. Which process should be the first candidate? How much autonomy is too much? When is an AI Agent the right answer versus a better KB article or a Virtual Agent topic?

 

This article introduces the A.G.E.N.T. Method — a five-phase framework for identifying, designing, and operationalising AI Agent use cases on ServiceNow. It's designed to be memorable, practical, and adaptable regardless of where your organisation sits on the AI maturity curve.

 

Note: The A.G.E.N.T. Method is a practitioner framework developed from patterns observed across customer engagements. It is not an official ServiceNow product methodology, but rather a structured approach that complements existing resources like the AI Gallery, Process Mining, and the Group Action Framework.


Before the Framework: Is This Even an Agentic Problem?

Before applying any framework, you need a filter. Not every process needs an AI Agent. From what we've seen across engagements, there are four sweet spots where agentic AI consistently delivers value:

  1. Complex decision-making — Where you need nuanced judgement, not just if–then logic. Processes with context-sensitive exceptions and longer SLAs.
  2. Rules that are difficult to scale — Extensive, intricate rule sets that become costly and error-prone every time a new iteration is added.
  3. Heavy reliance on unstructured data — Scenarios involving natural language interpretation: parsing free-text tickets, extracting meaning from emails, conversational interactions.
  4. Flexibility on inputs and methods — Where the steps may change per logic path decided at runtime, and the agent needs to adapt its approach dynamically.

If a process doesn't hit at least one of these, you may be better served by a standard workflow, an improved Knowledge Base, or a Virtual Agent topic. The most expensive mistake we see is teams building an AI Agent when a well-structured catalog item would have handled 80% of the volume.


The A.G.E.N.T. Method

Each letter maps to a phase. Each phase builds on the one before it. And the method loops; once you've completed the fifth phase, you cycle back to the first for your next use case, faster each time.


A — Assess the Landscape

Before you open any tooling, understand the terrain.

The Assess phase answers foundational questions:

  • What's the business problem? Define it clearly. What's the pain, the volume, and who's affected?
  • How does it tie to KPIs? Connect the use case to metrics your leadership already tracks. If you can't tie it to an existing KPI, getting executive sponsorship becomes much harder.
  • What already exists in your instance? Audit your KB articles, catalog items, VA topics, and existing workflows. You'd be surprised how often partial solutions already exist.
  • Is this feasible today? Assess your data readiness, technological maturity, and platform version. Not every great idea is buildable right now.

The most underrated activity in this phase is sitting with your L1 support teams. Even one hour with your frontline surfaces more actionable insight than a week of abstract workshops. Ask them: What's repetitive? What takes disproportionate time? Where do you wish the system could just handle it?

 

Complementary tools: Process Mining Evaluation Projects and the Group Action Framework are excellent companions to the Assess phase; they give you data-driven confirmation of where the real pain lives.

Best Practice: Don't skip the filter. The best agentic projects are the ones where a simpler solution genuinely wouldn't work.


G — Ground in Real Data

25 real records beat 100 assumptions.

This is the phase that separates projects that fly from projects that crash. The principle is straightforward: pull at least 25 real production records that represent the problem you're trying to solve.

Why 25? It's the minimum to get a meaningful cross-section — different categories, priorities, and edge cases. Maybe 5 high-priority, 10 medium, 10 low. For each record, ask:

  • What would I have wanted the AI Agent to do here?
  • What data would it need as input?
  • What should the output look like?
  • What decision logic would drive the path?

This phase also forces you to make concrete decisions about:

  • LLM selection — ServiceNow models (NowLLM or other OEM models) or third-party options (bring your own key). Match the model to the task complexity.
  • Memory requirements — Does the agent need short-term memory within a session, or long-term memory across interactions?
  • Functional scope — What is explicitly in scope and what is explicitly out?

Best Practice: Start with a cross-section of at least 25 actual production records you would have wanted the AI Agent to address. If you can't identify 25 records, the problem may not be frequent enough to justify agentic AI as the solution.


E — Engineer the Experience

If the human can't understand what the agent did, it's not working.

This is the design phase — the most creative and often the most underestimated. You're not just building backend logic; you're designing a collaboration between humans and AI.

 

Interaction Design: What does the agent's output look like for the person using it? On ServiceNow, this often manifests through the Now Assist Panel — skill cards showing summarization, resolution recommendations, and KB matches that the fulfiller can accept, reject, or modify. The UX is the product.

Autonomy Levels: Set these per use case based on risk appetite. For read-heavy operations like triage and classification, fully autonomous may be appropriate. For actions that modify production data — resolving incidents, updating CIs, closing changes — keep a human in the loop. Most organizations start advisory and expand autonomy as trust builds.

Override Mechanisms: Humans must always be able to pause, reject, or modify agent actions. This isn't just a safety feature, it's a trust feature.

Governance, Ethics, and Compliance: GDPR, HIPAA, industry-specific regulations, and your corporate AI governance policy all need to be addressed in this phase. Involve security, legal, and compliance stakeholders early, not as an afterthought.

Hallucination Mitigation: Ground agent responses in Knowledge Base content rather than open generation. Set confidence thresholds. Define escalation paths for low-confidence outputs. Consider output transformation strategies (e.g., "Summary for search results") to filter and retain only relevant information.

Mind-Map the Toolkit: Document the agent's toolkit — Skills, Workflows (scripted, AI, conversational), Knowledge Base sources, output controls, and backout plans. Tools like MindMeister or Miro work well for this.

Best Practice: Mind-map the agent's full toolkit before writing a single prompt or instruction. Include backout plans for when things go wrong — because at some point, they will.


N — Navigate the Pilot

Working and useful are two different things.

You've designed the agent. Now prove it works — and more importantly, prove it's useful.

Testing Scenarios: Cover happy paths, edge cases, error conditions, and adversarial inputs. Each test should be independently verifiable.

Accuracy Thresholds: The pattern we see across most organizations is an 75–80% accuracy rate before they'll move to UAT or a controlled pilot. But accuracy alone isn't the whole story.

Holistic Pass/Fail Criteria: An agent can technically "work" but still fail. Your criteria should include:

  • Did the agent invoke the right tools?
  • Was the output understandable to the human?
  • Was the latency acceptable?
  • Did the agent follow its instructions sequentially without skipping steps?

Pilot KPIs Tied to Business Outcomes: Map your pilot metrics back to the business problem from the Assess phase. If you're tackling incident resolution, measure MTTR, first-contact resolution rate, and SLA compliance, not just "the agent ran without errors."

Scope Management: This is critical. Major changes mid-pilot essentially reset the agent's learning. Small iterative tweaks are fine and expected. Full redesigns mean going back to the Engineer phase.

Best Practice: Align on what constitutes a pass versus a fail before testing begins. There are many parts to an agent. The agent might work correctly but produce output that's hard to interpret, which should still be flagged for improvement.


T — Tune & Scale

Production is the beginning, not the end.

The agent is live. The work isn't over — it's entering a new phase.

Observability via Activity Logs: ServiceNow's AI Agent activity logs provide a full reasoning trace; why the orchestrator chose certain agents, what tools were invoked, what data passed between steps, and the final output. This is your primary diagnostic tool when quality degrades.

Performance Analytics: Use PA dashboards to track whether the agent is actually moving the needle on your KPIs over time. MTTR trending down? First-contact resolution improving? SLA compliance holding steady? This data justifies the investment and builds the case for scaling.

Continuous Improvement Loops: Feed edge cases, failures, and user feedback back into prompt refinement, tool configuration, and instruction tuning. The agent should get better over time, not just maintain baseline performance.

Versioning Discipline: Distinguish between major updates (fundamentally change agent behavior → re-pilot through Navigate) and minor tweaks (refine existing behavior → lighter validation). Document every change. Create a technology stack document detailing every tool, framework, and service involved.

Scale to the Next Use Case: Take your learnings back to "A." The second time through the method is faster because you've built organizational muscle. You know how to assess, you have a testing framework, you understand your governance model.

Best Practice: The method loops. Once you've Tuned your first use case, go straight back to Assess for the next one. Each cycle gets faster.


Where to Start then?

If you're reading this and want a concrete first step:

  1. Pick one high-volume, read-heavy use case. Incident triage is the classic starting point — it hits all four sweet spots and carries low risk since the agent is advising, not acting.
  2. Pull 25 production records. Filter your incident list for a mix of priorities and categories. For each one, write down what you'd have wanted the agent to do.
  3. Sit with your L1 team for an hour. Ask what's repetitive, what's painful, what they wish the system could just handle.
  4. Run it through A.G.E.N.T. You'll have a solid use case design within a couple of weeks.

For pre-built agentic workflows you can deploy and customize today, explore the ServiceNow AI Gallery.


The A.G.E.N.T. Method at a Glance

Phase Name Core Question
A Assess the Landscape Is this the right problem, and is it truly agentic?
G Ground in Real Data Can we define this with 25 real production records?
E Engineer the Experience What does the human–agent interaction look like?
N Navigate the Pilot Does it work and is it useful?
T Tune & Scale Is it improving, and what's next?

Related Resources


This article is part of a series from the AI Center of Excellence team at ServiceNow. We work directly with enterprise customers to accelerate AI adoption, and we share practical guidance from the field with the broader community. If you found this useful, leave a comment or share it with your team.

Version history
Last update:
2 hours ago
Updated by:
Contributors