Business Process Management¶
Summary¶
Treats the organization as a system of processes that IS supports and improves: BPMN modeling, re-engineering, workflow automation, RPA, and continuous improvement (Lean, Six Sigma).
Role in the course: Show that IS work is process work: before you can automate or analyze, you must model the process the system serves.
Concepts Covered¶
This chapter covers the following 19 concepts from the learning graph:
- Process Modeling
- BPMN Notation
- Pool and Lane
- Process Activity
- Gateway
- Process Event
- As-Is Process
- To-Be Process
- Process Reengineering
- Lean Methodology
- Six Sigma
- DMAIC
- Workflow Automation
- Robotic Process Automation
- RPA Tools
- Process Mining
- Continuous Improvement
- Value Stream Mapping
- Business Rules Engine
Prerequisites¶
This chapter builds on concepts from:
Welcome back
Welcome to Chapter 5. So far we have talked about what information systems are and how they get built. This chapter is about what they actually do: support and improve the processes by which an organization gets work done. By the end you will be able to read a BPMN diagram the way you read a recipe, tell the difference between a workflow engine and an RPA bot without squinting, defend a Lean recommendation against a Six Sigma purist, and — perhaps most importantly — recognize the moment when "let's automate this" is the worst possible answer to a problem that needed redesigning instead. The skill of seeing an organization as a network of processes is one of the highest-leverage skills in IS. Let's collect it.
Why Processes, and Why Now¶
Every organization, whether it knows it or not, is a collection of processes — repeatable sequences of activities that turn inputs into outputs that someone, somewhere, actually wants. The marketing department runs a lead-qualification process. The finance department runs a month-end close process. Even the cafeteria runs a sandwich-assembly process, and if you have ever watched the line back up because somebody could not decide between Swiss and provolone, you have witnessed a process bottleneck in real time.
The reason processes matter to information systems is structural: an information system without a clearly understood process is an information system serving nobody in particular. You cannot design a database schema without knowing what events the schema must record. You cannot build an application without knowing what the user is trying to accomplish. You cannot automate something you cannot describe. The process is the thing the system is for. Skipping process analysis to get to "the technology" is one of the most reliable ways to ship a beautiful system that nobody uses.
There is also a long historical reason. The discipline of Business Process Management (BPM) emerged in the 1990s out of the convergence of three older traditions: industrial engineering (which had been studying factory processes since Frederick Taylor), total quality management (which spread from Toyota into Western manufacturing in the 1980s), and the early enterprise software wave (which made it suddenly possible to instrument and automate office work the way factories had been instrumented for decades). BPM is what you get when you take the rigor of factory process improvement and apply it to white-collar work. It is older than this textbook, and frankly, older than most of the textbooks you will encounter in IS.
Process Modeling: Drawing the Thing Before You Touch It¶
Process Modeling is the disciplined activity of representing a business process as a formal diagram, with explicit notation for the steps, the decisions, the actors, the inputs, and the outputs. The model is not the process — the process keeps happening whether or not anyone draws it — but the model is the artifact the team can argue about, mark up, share with auditors, and feed into a workflow engine. A process that has never been modeled is a process that lives entirely in the heads of the people who do it, and that is a fragile place for a process to live.
Process modeling buys an organization three things at once. First, shared understanding: the moment you draw a process for a room of people who all run pieces of it, you discover they disagreed about what the process actually was. (This is annoying, but it is the entire point.) Second, defect surfacing: latent problems — duplicate work, missing handoffs, decisions made by nobody — become visible when you put them on a wall. Third, automation readiness: a clean model is the first step toward automating, monitoring, or instrumenting the process with software.
The footgun in process modeling is over-modeling. A team that produces a sixty-page BPMN specification before it has redesigned anything has not modeled a process; it has produced expensive wallpaper. The model is a means; the improved process is the end. A useful process model is one detailed enough to support the next decision and no more.
BPMN Notation: The Lingua Franca¶
The dominant notation for process modeling is BPMN — Business Process Model and Notation — currently at version 2.0, governed by the Object Management Group (OMG). BPMN exists for the same reason musical notation exists: so that a process drawn by an analyst in Stuttgart can be read by an analyst in São Paulo without translation. Before BPMN there were dozens of competing process-diagram dialects (flowcharts, IDEF0, EPCs, ad-hoc whiteboard scribble), none of them universally understood. BPMN became the standard the way QWERTY became the keyboard layout: not because it was perfect, but because it was good enough and everyone agreed.
BPMN Notation consists of a small vocabulary of graphical elements with strict meanings. The four most important categories are flow objects (activities, gateways, events), connecting objects (sequence flows, message flows), swimlanes (pools and lanes), and artifacts (data objects, annotations). Most BPMN diagrams you will see in industry use perhaps fifteen of the notation's roughly hundred-plus symbols — the rest are reserved for specialists.
Let's introduce the four elements you must know in order to read any BPMN diagram, in the order you should learn them.
Pool and Lane: Who Does What¶
A Pool and Lane structure organizes the diagram by actor. A pool is a rectangular container representing a single participant in the process — typically an organization, system, or major role boundary. Within a pool, lanes subdivide responsibility further: usually departments, teams, or roles inside that organization. The convention is that a pool reads horizontally left-to-right, and any activity drawn inside a particular lane is performed by the actor that lane represents.
The pedagogical superpower of swimlanes is that they make handoffs visible. Every time a sequence flow crosses from one lane to another, work has changed hands — and handoffs are where processes most often break, slow, or lose information. A diagram with five lane crossings is doing more handoffs than a diagram with two; whether that is appropriate depends on the work, but you cannot have the conversation without the diagram.
Process Activity: The Actual Work¶
A Process Activity is a unit of work that someone (or some system) performs as part of the process. Activities are drawn as rounded rectangles with a verb-phrase label: Receive Order, Validate Credit, Pick Items, Ship Package. Two flavors matter early:
- A task is an atomic activity — work that is not further decomposed in this diagram. "Approve invoice" might be a task in the AP-clerk lane.
- A sub-process is an activity that itself expands into a smaller BPMN diagram. "Onboard new employee" might be a sub-process containing twenty smaller activities elsewhere.
The discipline of activity labeling matters more than students typically expect. Verb + noun labels ("Validate Credit", "Generate Invoice") survive contact with users. Vague labels ("Process") survive nothing. If you find yourself naming an activity "Process the Thing", you have not finished modeling.
Gateway: The Decisions¶
A Gateway is a diamond-shaped element that controls how sequence flows split or merge. Gateways are how a BPMN diagram represents decisions and parallelism. The gateway types you will see most are:
- Exclusive (XOR) gateway — exactly one outgoing path is taken, based on a condition. Drawn as a plain diamond or a diamond with an "X". "Is the order over $10,000? If yes, route to manager review; if no, auto-approve."
- Parallel (AND) gateway — all outgoing paths are taken simultaneously. Drawn as a diamond with a "+". Used when work can proceed in parallel: "Notify shipping AND notify accounting."
- Inclusive (OR) gateway — one or more outgoing paths are taken, based on conditions. Drawn as a diamond with an "O". Used when multiple conditions can be true at once.
- Event-based gateway — the next path is chosen by which event happens first (e.g., customer responds, or 48 hours elapses).
The footgun with gateways is using the wrong type. An exclusive gateway when work should run in parallel forces the team into needless serialization. A parallel gateway where conditions are mutually exclusive creates phantom branches that the merge gateway then has to wait for forever. Picking the gateway is picking the logic; get the gateway wrong and the diagram lies.
Process Event: The Things That Happen To You¶
A Process Event is something that happens during the process, drawn as a circle. There are three positions and several types:
- A start event (thin-bordered circle) marks where the process begins — typically a triggering condition like "customer submits order" or "month ends".
- An intermediate event (double-bordered circle) marks something that happens during the process — a message arrives, a timer fires, an error is raised.
- An end event (thick-bordered circle) marks where the process terminates — successful completion, escalation, cancellation.
Events are the parts of a process the actors don't do — they're the parts that happen to the actors. A timer firing, a message arriving, an exception being raised. Many BPMN beginners under-use events because tasks are easier to think about, but a process diagram with no events is usually a process that has not modeled how it begins, how it ends, or what to do when something goes wrong.
| BPMN element type | Shape | Represents | Common pitfall |
|---|---|---|---|
| Pool | Outer rectangle | One participant (org, system) | Mixing two orgs in one pool |
| Lane | Horizontal strip in a pool | Role/team/department | Lanes that don't match real responsibility |
| Activity (task) | Rounded rectangle | Unit of work performed | Vague verb-noun labels |
| Gateway | Diamond | Decision or parallelism | Wrong gateway type for the logic |
| Event | Circle (single/double/thick) | Trigger, occurrence, or termination | Forgetting end events |
| Sequence flow | Solid arrow | Order of work within a pool | Crossing pool boundaries (use msg flow) |
| Message flow | Dashed arrow | Communication between pools | Drawn as a sequence flow by mistake |
Diagram: Anatomy of a BPMN Order-to-Cash Process¶
Anatomy of a BPMN Order-to-Cash Process
Type: interactive-infographic
sim-id: bpmn-order-to-cash
Library: vis-network
Status: Specified
A vis-network BPMN diagram showing a simplified order-to-cash process inside a single Customer-Facing-Company pool with three lanes: Sales, Credit, and Fulfillment. The process begins with a start event labeled "Order received" in the Sales lane, flows through a "Validate order" task, then crosses into the Credit lane via an exclusive gateway labeled "Order > $10,000?". The yes path runs a "Manager credit review" task; the no path runs an "Auto-approve credit" task. Both paths re-merge at a converging exclusive gateway, then cross into the Fulfillment lane via a parallel gateway that triggers two simultaneous tasks: "Pick and pack" and "Generate invoice". A converging parallel gateway joins them, followed by a "Ship order" task and an end event labeled "Order shipped". A second pool labeled "Customer" sits below, with a message flow connecting the Customer to the start event ("places order") and another message flow back from "Order shipped" ("delivery confirmation"). To work around the vis-network horizontal-edge label rendering bug, sequence flows use a slight y-offset (e.g., 480 to 490) so labels render on initial load.
Color palette: pool boundaries in slate-gray, lanes in alternating soft-blue tints, activity rounded rectangles in mascot-emerald with white text, exclusive gateways in amber with the "X" marker, parallel gateways in coral with the "+" marker, events in mascot-magenta gradients. Sequence flows in dark teal; message flows dashed in muted gray.
Interactivity: hovering each element reveals (a) the BPMN element type, (b) its formal definition, and (c) a concrete real-world example. A "trace execution" button animates a token traveling through the diagram following one of two paths (large order vs. small order), pausing briefly at each gateway to show the routing decision. A toggle highlights every "lane crossing" — the visible handoffs in the process — to reinforce the systems-thinking lesson that handoffs are where processes break.
Layout: hierarchical, left-to-right, full canvas width, height ~560px. Canvas resizes on window resize.
Learning objective (Bloom: Applying): Students can read a BPMN diagram, name each element type by its shape, identify lane handoffs, and trace a token through both paths of an exclusive gateway.
Implementation: vis-network, deployed at /information-systems/sims/bpmn-order-to-cash/.
Iris's Tip
When you read someone else's BPMN diagram, count the lane crossings before you read anything else. Each crossing is a handoff — a place where work leaves one person's hands and enters another's, where context can be lost, where forms can sit in queues, where finger-pointing can begin. Diagrams with many crossings are not always wrong, but they are always worth interrogating. A pro will look at a five-lane diagram and ask, "Could this be a three-lane diagram?" — because every lane you can dissolve is an organizational seam you no longer have to maintain.
As-Is and To-Be: The Two Diagrams Every Project Needs¶
Once a team has BPMN in its pocket, the most useful single move in process work is to draw two diagrams of the same process: the As-Is Process and the To-Be Process.
The As-Is Process is the process as it actually is today, warts and all. Not the way the policy manual describes it, not the way the executive sponsor thinks it works, but the way it actually runs in the field on a Tuesday afternoon when the nice manager is on vacation. Drawing the as-is is humbling work. The team always discovers undocumented steps ("oh, Karen always emails the file to herself first, that's just how we do it"), redundant approvals, decisions that were originally added to handle a 2014 audit and now serve no purpose, and at least one workaround for a bug in a system nobody on the call has the password to anymore.
The To-Be Process is the process as it should be after improvement: simpler, faster, less error-prone, better instrumented, sometimes radically different in shape. The to-be is not a wish list — it is a designed alternative that the team has thought through, validated against constraints, and committed to as the target state.
The reason to draw both — instead of just leaping to the to-be — is that the as-is is where the real lessons live. The as-is reveals which steps exist for legitimate compliance reasons (those have to survive into the to-be) versus which exist only because nobody has questioned them in eight years (those can be deleted). The as-is also surfaces the political reality of the change: every redundant approval has somebody whose self-image is wrapped up in performing it.
| As-Is Process | To-Be Process |
|---|---|
| Captures what is actually happening | Captures what should happen after change |
| Often messy, with workarounds | Often simpler, with workarounds removed |
| Built from interviews and observation | Built from analysis, design, and tradeoff |
| Reveals legitimate vs vestigial steps | Targets a measurable improvement |
| Politically uncomfortable to draw | Politically uncomfortable to defend |
Process Reengineering: When Improvement Is Not Enough¶
Sometimes the gap between the as-is and the to-be is small — a few redundant steps removed, a handoff streamlined. Sometimes the gap is enormous, and what the organization needs is not improvement but a clean-sheet redesign. That is the territory of Process Reengineering, often capitalized as Business Process Reengineering (BPR), the radical-redesign movement that swept through corporate America after Michael Hammer's 1990 Harvard Business Review article "Reengineering Work: Don't Automate, Obliterate."
Hammer's central insight has aged well: the worst thing you can do with a broken process is automate it faster. Automating a bad process produces a bad process that runs more efficiently — which is, on a long enough timeline, worse than the original. Reengineering asks the deeper question: if we were starting from scratch today, with the technology we have now, would we design this process at all? Sometimes the answer is no. Sometimes a fifteen-step approval workflow gets reengineered into a one-step workflow plus an exception path, because fourteen of the steps existed only to compensate for limitations of a 1980s mainframe.
Reengineering is high-risk and high-reward. The successful cases of the 1990s — Ford's accounts payable, IBM Credit's quote-to-cash, Mutual Benefit Life's policy issuance — produced 5x to 10x improvements in cycle time. The unsuccessful cases — and there were many — produced expensive consulting bills, layoffs, and demoralized survivors who learned that "reengineering" was a word their leadership used when they wanted to fire people. The footgun is reengineering as a euphemism; the structural fix is to be honest about goals (efficiency? compliance? customer experience? cost?), to involve the people who actually run the process, and to manage the human side of the change as carefully as the technical side.
Pause here
There is a hierarchy of leverage in process work that every IS professional should internalize. From lowest to highest leverage: automate the existing steps, streamline the existing steps (eliminate sub-steps within a task), eliminate entire steps that no longer earn their keep, and reengineer the process so the steps themselves are different. Most of the time, the highest-impact move is several rungs above the one the team instinctively reaches for. When somebody says "let's automate this" — pause and ask whether eliminating it would be even better. The cheapest, fastest, most reliable step is the one you don't have to do.
Lean: Eliminate the Waste¶
Process improvement has two dominant intellectual traditions. The first is Lean Methodology, born at Toyota in the post-war decades and codified in the West by Womack, Jones, and Roos in The Machine That Changed the World (1990). Lean's central concept is waste — anything in a process that consumes resources without adding value the customer would pay for. Lean's job is to see waste and eliminate it.
Lean classically catalogs eight wastes (the original seven Toyota wastes plus an eighth added later for knowledge work):
- Defects — work that has to be redone because it was wrong the first time.
- Overproduction — making more than is needed, sooner than it is needed.
- Waiting — work sitting in a queue, waiting for the next step or the next person.
- Non-utilized talent — people doing work below their skill level (the eighth waste).
- Transportation — moving things (or files, or tickets) between locations unnecessarily.
- Inventory — work-in-progress piling up between steps.
- Motion — people moving around, or clicking around a UI, more than necessary.
- Extra processing — doing more to a thing than the customer requires.
The mnemonic DOWNTIME (Defects, Overproduction, Waiting, Non-utilized talent, Transportation, Inventory, Motion, Extra processing) is the standard memory aid; if you remember nothing else from Lean, remember DOWNTIME and you can spot waste in any process you walk into.
Lean's deepest contribution to process work is its cultural posture: continuous, distributed, ground-up improvement, owned by the people who do the work. The Japanese term kaizen — small, ongoing, employee-driven improvements — captures the philosophy. Lean is not primarily about big projects; it is about a thousand tiny improvements compounding over years. That is also why it is so easy to fail at: most organizations do not have the patience, the cultural humility, or the management courage to let frontline workers rewrite their own work.
Six Sigma and DMAIC: Reduce the Variation¶
The second great process-improvement tradition is Six Sigma, developed at Motorola in the 1980s, popularized at GE under Jack Welch in the 1990s, and built around statistical process control. Where Lean focuses on eliminating waste, Six Sigma focuses on reducing variation. The "Six Sigma" name refers to a defect rate of 3.4 defects per million opportunities — six standard deviations from the mean — which is a lot of decimal places, but the underlying idea is simple: a process that produces consistent output is more valuable than one that produces good-on-average output with bad surprises.
Six Sigma's signature methodology for improving an existing process is DMAIC (pronounced "duh-MAY-ik"), a five-phase cycle:
- Define — clarify the problem, the customer, and the goals. Charter the project, name the deliverable, agree on what "improvement" will mean numerically.
- Measure — instrument the as-is process. Establish baselines for the metrics you intend to move. You cannot improve what you have not measured.
- Analyze — find the root causes of the defects or variation. Tools include Pareto charts, fishbone diagrams, hypothesis tests, and statistical process control.
- Improve — design and implement the change. Pilot it, validate it against the baseline, and confirm the improvement is real (not just regression to the mean).
- Control — institutionalize the improvement. Update procedures, train staff, instrument ongoing monitoring, and set up alarms so the process doesn't drift back to the old behavior.
Six Sigma also famously borrowed martial-arts ranks for its practitioners — Yellow Belt, Green Belt, Black Belt, Master Black Belt — which lent it a corporate-cult flavor that aged less gracefully than its statistical methods. The methods themselves are entirely sound and broadly applicable; the cultural packaging was a product of its time. (You can keep the DMAIC and skip the belts.)
| Dimension | Lean | Six Sigma |
|---|---|---|
| Primary target | Waste | Variation and defects |
| Cultural posture | Bottom-up, continuous, kaizen | Top-down, project-based, statistical |
| Signature tools | Value stream mapping, 5S, kanban | DMAIC, control charts, fishbone |
| Data intensity | Light to moderate | Heavy — statistical analysis required |
| Best fit | Repeatable office/factory work | High-volume processes with measurable defects |
| Common combination | "Lean Six Sigma" — both at once | "Lean Six Sigma" — both at once |
In practice, the two traditions are usually combined into Lean Six Sigma, a hybrid that uses Lean tools for waste elimination and Six Sigma tools for variation reduction. Most modern process-improvement programs draw from both toolboxes without much fuss about lineage.
Value Stream Mapping: Seeing the Whole Flow¶
The most useful single Lean tool, and the one most worth mastering, is Value Stream Mapping (VSM). A value stream map is a specialized process diagram that shows the entire flow of value — from raw input to delivered output — annotated with the time spent at each step (process time) and, crucially, the time spent between steps (wait time). The result is usually shocking: in most office processes, work spends 2-5% of its lifecycle being worked on and 95-98% sitting in a queue waiting to be worked on. Once you have seen this on a wall, you cannot unsee it.
Value stream mapping is what a Lean practitioner does first, before reaching for any other tool, because it makes the whole flow visible. The discipline forces the team to walk the actual process — physically or virtually — and document what really happens, not what the documentation says happens. (Lean people sometimes call this gemba — going to the real place where the work happens. It is the antidote to designing improvements from a conference room.)
The standard VSM symbology shows process boxes (with cycle time, change-over time, uptime), inventory triangles (with quantities), information flows (manual vs electronic), customer demand, supplier relationships, and a timeline at the bottom that distinguishes value-added time (the time work is actually being transformed) from non-value-added time (everything else). The eye-watering ratio between those two is the headline finding of nearly every VSM project.
Diagram: A Value Stream Map of an Invoice-Approval Process¶
A Value Stream Map of an Invoice-Approval Process
Type: interactive-infographic
sim-id: value-stream-map-invoice-approval
Library: p5.js
Status: Specified
A p5.js value stream map showing an invoice-approval process from end to end: invoice received → AP-clerk data entry → manager approval → director approval (over $10K) → AP-clerk payment scheduling → payment issued. Each process box is rendered as a labeled rectangle with three sub-fields (cycle time, change-over time, uptime). Between each pair of process boxes is an inventory triangle showing the typical queue depth (e.g., "23 invoices waiting"). At the bottom of the canvas runs a time ladder: the upper rung shows non-value-added time (the wait time at each queue), the lower rung shows value-added time (the actual processing time). A summary box on the right reports total lead time, total value-added time, and the value-added ratio (typically 2-4%). Information flows (manual vs electronic) are shown above the process boxes with appropriate arrow styles.
Color palette: process boxes in mascot-emerald, inventory triangles in coral (red triangles to evoke "stop, pile-up here"), value-added time in green on the timeline, non-value-added time in red. Information flow arrows in slate-gray (manual = jagged line, electronic = lightning bolt).
Interactivity: hovering each process box shows the activity description and the real-world cycle time. A "to-be" toggle redraws the map after Lean improvements (e.g., eliminating the duplicate-approval step, batching reviews, auto-approving small invoices) and shows the updated value-added ratio. The dramatic before/after comparison — typically a jump from 3% to 25% value-added ratio — is the pedagogical payoff. A "wait time" slider lets students stress-test the model under different queue conditions to see how lead time explodes when WIP is unbounded.
Layout: full canvas width, height ~520px. Canvas resizes on window resize. setup() calls updateCanvasSize() first. Canvas parented with canvas.parent(document.querySelector('main'));.
Learning objective (Bloom: Analyzing): Students can read a value stream map, compute a value-added ratio, identify the largest sources of waste, and propose a sequence of Lean interventions that improves the ratio.
Implementation: p5.js, deployed at /information-systems/sims/value-stream-map-invoice-approval/.
Workflow Automation: The Engine Underneath¶
Once a process is modeled and improved, the next step in many cases is Workflow Automation: software that executes the process directly, routing work between human actors, triggering system actions, enforcing rules, and recording every step in a way that auditors can replay. A workflow automation engine takes a process model — increasingly, the BPMN diagram itself — and runs it as code. Cases enter the engine, the engine routes them to the right person or system at the right time, the engine logs what happened, and the engine raises exceptions when work stalls.
The strategic value of workflow automation is twofold. First, consistency: the workflow engine performs the routing exactly the same way every time, so the process behaves predictably regardless of which staff are working that day. Second, visibility: because every step is logged in a structured way, the organization can see at any moment how many cases are in which step, which steps are slow, and where bottlenecks have formed. The same instrumentation that makes the process auditable also makes it improvable.
Modern workflow platforms — Camunda, IBM Process Server, Pega, Appian, ServiceNow Workflow, Microsoft Power Automate, Salesforce Flow — all share roughly the same architecture: a process designer (often graphical, often BPMN-based), a process engine (the runtime), a task list (where humans pick up work), an admin console (for monitoring and intervention), and an integration layer (for talking to other systems). What distinguishes them is the depth of integration with their host platform, the sophistication of their rules engines, and the ergonomics of their designers.
The footgun in workflow automation is the same one that haunts process work generally: automating a bad process locks the bad process in. Once the workflow engine is running the process, the cost of changing the process goes up, because changing the diagram now means redeploying production software. Teams that automate before they reengineer end up with a beautifully instrumented version of exactly the process they should have replaced.
Robotic Process Automation: When Integration Costs More Than a Bot¶
Sometimes the systems in a process are simply not integrate-able by normal means. The system of record is a thirty-year-old mainframe with a 3270 terminal interface. The vendor portal has no API. The acquired company runs a legacy Windows application with no programmable interface. In all three cases, the proper engineering solution is API integration — and the proper engineering solution is unavailable, unaffordable, or scheduled for retirement next year. Enter the bots.
Robotic Process Automation (RPA) is a category of software that automates business processes by mimicking the actions of a human user at the user-interface level: opening applications, reading screens, filling fields, clicking buttons, copying values from one system to another. An RPA "bot" is essentially a script that drives the GUI the way a human would, except 24/7 and without coffee breaks.
RPA's appeal is straightforward. It can automate work across systems that were never designed to be integrated, without modifying any of those systems, without waiting for an API roadmap, and without involving the IS team that owns each underlying application. For routine, high-volume, rule-based, screens-based work — claims processing, invoice routing, account reconciliation, employee onboarding — RPA can deliver automation in weeks where a proper API integration would take quarters.
RPA's footgun is its core mechanism. Because the bot interacts with the user interface, every UI change in any of the underlying systems can break the bot — silently, with no error visible to anyone except the queue of work backing up behind the silently-failing bot. This is a textbook footgun: silent (the bot does not throw an error you would notice), easy to trigger (any vendor patch can move a button), and damaging in delayed and invisible ways (work piles up downstream until somebody asks why the queue is so long). The structural fix is not "more careful UI selectors"; it is honesty about the tradeoff. RPA is the right answer when the alternative is "no automation at all, ever." When proper API integration is achievable on a reasonable timeline, RPA is usually the wrong answer dressed in a tuxedo.
RPA Tools that dominate the market include UiPath, Automation Anywhere, Blue Prism, Microsoft Power Automate Desktop, and (increasingly) AI-augmented platforms that combine traditional RPA with computer vision and large-language-model reasoning. The general shape of an RPA tool is: a recorder (which captures a human performing the task), a designer (which lets you edit the recorded script visually), an orchestrator (which schedules and supervises bots in production), and an analytics layer (which reports on bot health and savings).
| Dimension | Workflow Automation | Robotic Process Automation |
|---|---|---|
| Integrates via | APIs, queues, native connectors | The user interface (screen scraping) |
| Modifies underlying apps | Often (deeper integration) | Never — that's the point |
| Brittleness | Low — APIs are versioned | High — UI changes break bots silently |
| Time to first automation | Weeks to months | Days to weeks |
| Right answer when | Apps have APIs and you control them | Apps have no APIs, no roadmap, no choice |
| Footgun | Locks in a bad process | Bot fails silently after a UI patch |
| Long-term posture | Strategic — the durable foundation | Tactical — bridge to proper integration |
Diagram: As-Is vs To-Be — Manual, RPA, and Workflow Automation¶
As-Is vs To-Be — Manual, RPA, and Workflow Automation
Type: interactive-infographic
sim-id: as-is-to-be-automation-comparison
Library: p5.js
Status: Specified
A p5.js side-by-side comparison diagram showing three versions of the same employee-onboarding process. Panel 1 (As-Is, Manual): five lanes — HR clerk, IT, Facilities, Manager, Payroll — each performing manual handoffs by email; total cycle time labeled at the top (e.g., "9 days"). Panel 2 (To-Be with RPA): same lane structure, but two RPA bot icons replace the HR-clerk's repetitive data-entry steps and the IT account-creation steps; cycle time labeled (e.g., "3 days"); a small caution icon notes "fragile to UI changes". Panel 3 (To-Be with Workflow Automation): collapsed to three lanes (HR, Manager, Payroll) with a workflow-engine icon orchestrating system-to-system integration via APIs; cycle time labeled (e.g., "1 day"); a small "strategic foundation" badge.
Color palette: manual-only lanes in slate-gray, RPA bot icons in coral with caution-stripe borders, workflow-automation engine in mascot-emerald, integration arrows in teal. Cycle-time bar at the top of each panel scales proportionally so students can see the time difference visually.
Interactivity: hovering each step in any panel shows duration and actor. A "show waste" toggle highlights the eight Lean wastes wherever they occur in the as-is panel (especially Waiting and Motion). A "switch panel" tab lets the student bring any of the three panels into a primary focused view. A side panel shows total cycle time, value-added ratio, and number of human handoffs for the panel currently in focus.
Layout: three vertical stacked panels, full canvas width, height ~640px. Canvas resizes on window resize. setup() calls updateCanvasSize() first. Canvas parented with canvas.parent(document.querySelector('main'));.
Learning objective (Bloom: Evaluating): Students can compare a manual, RPA-automated, and workflow-automated version of the same process; quantify the cycle-time and handoff differences; and articulate the tradeoffs of each approach.
Implementation: p5.js, deployed at /information-systems/sims/as-is-to-be-automation-comparison/.
Heads up
There is a special circle of IS purgatory reserved for organizations that have both an enthusiastic RPA program and a long backlog of API integrations the bots are working around. Each successful bot deployment lowers the perceived urgency of fixing the underlying integration, the integration backlog grows, and a few years later the company has 300 bots, no one remembers what each of them does, every vendor UI update is a fire drill, and the original integration backlog is still untouched. The structural fix is to track every RPA deployment with a paired ticket: "Replace this bot with a proper API integration when X system is upgraded." Without that paired ticket, RPA is not a bridge to anywhere — it is the destination, and the destination is a maintenance nightmare.
Process Mining: Listening to the Process¶
For decades, process modeling required a human analyst to interview people and construct the as-is diagram by hand. The diagram that emerged was, inevitably, an idealized version of reality — a composite of how everyone thought the process worked, with the inconvenient bits sanded off. Then, around 2010, a Dutch computer scientist named Wil van der Aalst pointed out that nearly every modern enterprise system already records a log of every transaction it processes — and that those logs, if interpreted correctly, are a literal record of how the process actually runs. The discipline that emerged from that observation is Process Mining.
Process mining is the application of data-mining techniques to event logs from enterprise systems (SAP, Oracle, Salesforce, ServiceNow) to automatically reconstruct the as-is process diagram from the data. Tools like Celonis, UiPath Process Mining, IBM Process Mining, and Microsoft Process Insights ingest the event log, identify the case ID, the activity, and the timestamp on each event, and produce — within minutes — a process map that reflects what actually happens, including all the variants, exceptions, and shortcuts that nobody mentioned in the interviews.
The first time an organization runs process mining on a real process, the experience is disorienting. The "single, well-understood process" turns out to have 47 distinct variants. The 12-step process documentation turns out to describe a path that only 8% of cases actually take; the remaining 92% are a long tail of exceptions, rework loops, and undocumented shortcuts. The "fast" process executives believed in turns out to have a small number of cases that take 30 days because of a vendor integration nobody knew was failing.
This is process modeling's superpower update for the data era. Where the analyst-built as-is is an opinion, the mined as-is is a measurement. The two should agree; when they don't, the data wins. Process mining also enables continuous monitoring — once the mining model is in place, the organization can watch the process in something like real time, alert on emerging deviation, and quantify the impact of every improvement intervention.
The Business Rules Engine: Pulling the Logic Out¶
A subtle but powerful piece of the BPM toolbox is the Business Rules Engine (BRE). A business rule is a declarative statement of business logic: "Auto-approve any expense under $50." "Charge sales tax based on shipping ZIP code." "Reject loan applications where credit score is below 580 unless the applicant has co-signer documentation on file." Historically, business rules lived hard-coded inside application code — buried in stored procedures, sprinkled through user-interface validation, written into RPA bot scripts. Every rule change required a developer, a code review, a deployment, and a release window.
A business rules engine is a piece of software that externalizes the rules from the application code into a separate, declarative artifact (often a decision table, a rules language, or a DMN — Decision Model and Notation — diagram, which is BPMN's sibling standard) that business analysts can edit directly. The application or workflow engine queries the BRE at runtime ("here's an expense report; what's the decision?") and the BRE returns the answer based on the current rule set. Vendors include FICO Blaze Advisor, IBM ODM, Drools (open-source), Camunda DMN, and the rules engines embedded in larger BPM platforms.
The strategic payoff is that policy changes — which are constant in finance, healthcare, insurance, and government — become configuration changes rather than code changes. A new tax rate, a new compliance check, a new approval threshold can be edited and deployed without a software release. The footgun, of course, is that the rule set itself becomes a software artifact in disguise: thousands of rules can become a tangle that nobody can reason about, and the people editing them often do not have the test discipline of professional developers. The structural fix is to treat the rule set with the same engineering rigor as code — version control, peer review, automated tests, staged rollout — even when the editors are non-developers.
Continuous Improvement: The Engine That Doesn't Stop¶
We close on the cultural concept that ties all of the previous tools together: Continuous Improvement. The Japanese word kaizen — a compound of kai (change) and zen (good) — captures it: change for the better, ongoing, owned by everyone, never finished. Continuous improvement is not a project, a phase, a one-time initiative, or a six-month engagement. It is a disciplined organizational habit of always looking for the next small improvement and acting on it.
The engineering of continuous improvement has three load-bearing structural elements. Measurement — without baseline metrics, you can't tell whether a change is an improvement or a regression. Visibility — improvement opportunities have to be surfaced where leadership and front-line staff can both see them, often via process mining dashboards or simple kanban-style improvement boards. Capacity — staff must have time to work on improvements, distinct from their delivery commitments. The organizations that fail at continuous improvement are almost always the ones that asked for it on top of an already-100%-loaded delivery commitment, which is the corporate equivalent of asking someone to mow the lawn while they are running a marathon.
The systems-thinking insight beneath continuous improvement is that processes do not stay improved on their own. Every process has drift — staff turnover changes who knows what, regulations change the rules, vendor systems change their behavior, the business changes the goals. Without an active improvement loop, every improved process gradually decays back toward its prior state, often within a year or two of the project that improved it. The continuous-improvement loop is a balancing feedback that fights the drift. Without it, the doom loop wins. With it, processes can actually get better year over year, which is one of the most underrated competitive advantages an organization can develop.
Putting It All Together¶
Business process management is the discipline of treating the organization as a system of processes that can be modeled, redesigned, automated, instrumented, and continuously improved.
- Process modeling is the foundational activity; BPMN notation is its lingua franca; pools, lanes, activities, gateways, and events are the elements you must master to read or write a diagram.
- The as-is process captures reality; the to-be process captures the target; process reengineering is what you do when the gap is large enough to require a clean-sheet redesign rather than incremental improvement.
- Lean methodology eliminates waste (DOWNTIME); Six Sigma reduces variation; DMAIC is Six Sigma's five-phase improvement cycle; the two traditions combine into Lean Six Sigma in most modern programs. Value stream mapping is the most useful single tool from the Lean side, because it makes wait time visible.
- Workflow automation runs the process as code, executing the BPMN diagram directly and instrumenting it for analysis. Robotic process automation mimics human users at the UI level — fast to deploy, brittle to maintain, useful when proper integration is unavailable. RPA tools like UiPath, Automation Anywhere, and Blue Prism dominate the category.
- Process mining reconstructs the actual as-is from event logs, replacing analyst opinion with measurement. The business rules engine externalizes business logic from code, making policy changes configurable rather than deployable. Continuous improvement is the cultural and structural commitment that keeps all of this from slowly decaying back to its starting state.
A graduate of this chapter walking into their first process-improvement engagement should be able to ask, in order: What process are we improving? Have we modeled the as-is honestly? Have we mined the event logs to validate the model? What's the to-be? Are we eliminating, streamlining, or just automating? If we're automating, is it workflow automation or RPA, and have we acknowledged the tradeoffs? Where do the business rules live, and who can change them? And how will the organization keep improving this process after we leave? That is a more sophisticated set of questions than most consultants ask in their first decade of work. You can ask all of them after a single chapter.
Nicely done
You just learned the working vocabulary of business process management — BPMN swimlanes, gateways, events, the as-is/to-be discipline, Lean's eight wastes, Six Sigma's DMAIC, value stream mapping, the workflow-versus-RPA tradeoff, process mining, business rules engines, and the continuous-improvement loop. Next time somebody pitches "let's just automate this," you will hear they have not modeled the process yet. Next time somebody describes a six-month consulting engagement to "reengineer" something, you will hear what's the as-is, what's the to-be, and who owns the change? Next time somebody is excited about their new RPA bot, you will hear what happens when the vendor patches the UI? And next time a process actually gets better and stays better, you will recognize it as the rare and valuable thing it is. Onward to Chapter 6, where we step back from processes and ask the deeper question of how data — the substance every process is shoveling around — should be modeled, stored, and trusted.