Skip to content
Paul Sherman

Designing a Copper Ore Refinery Monitoring App

Three-weeks. Two onsite discovery and design sprints that produced validated wireframes, user flows, and an interactive prototype for a mobile-first copper ore refinery monitoring application.

3
Weeks
2
Design sprints
1
Wireframe and user flow deliverable

The Problem

Rio Tinto's Kennecott copper mine in Utah operates a concentrator facility where raw ore is processed to extract copper. The concentrator is a complex system of circuits and rows, each containing dozens of instrumented components: frother flow, collector setpoints, froth depth, air setpoints, pH levels, and more. The percentage of copper recovered from the ore depends on how well these components are tuned and maintained across each shift.

At the time this project began, shift planners estimated recovery percentages before each shift based on their own experience and a small set of upstream variables like ore type and grind size. Rio Tinto's PACE Analytics unit had built a predictive model that weighed hundreds of variables to generate more accurate recovery predictions. The model was more accurate than human estimates. But it hadn't been adopted. The planners and shift supervisors who needed the predictions had no way to access them in their day-to-day workflow.

This was a classic gap between what a data science team builds and what operational users can actually use. The model existed. The interface didn't. And without the interface, the model's value was theoretical.

What Existed Before

Before we began designing the application, the only way shift supervisors and planners could view the concentrator's real-time performance was through a live Excel spreadsheet. The spreadsheet displayed row-level recovery predictions, shift plan targets, and variance calculations across a set of tabs. It was not intuitive or easy to parse at a glance, and it did not provide actionable information that operators could use to adjust the ore recovery process in real time.

The predictive model itself produced a feature importance chart showing which of hundreds of variables were driving recovery predictions up or down. This was valuable to the data science team, but it was unintelligible to the shift supervisors and planners who needed to act on the information.

Excel spreadsheet showing the concentrator's live monitoring view with Monitor, Tornado, and Heatmap tabs. Row 7 through Row 10 display Prediction and Shift Plan percentages with variance calculations, presented in a dense tabular format that is difficult to scan quickly.
The existing Excel-based monitoring view. Shift supervisors and planners had to interpret raw prediction and variance numbers across rows and tabs with no visual hierarchy or actionable framing.
Horizontal bar chart showing SHAP feature importance values for the predictive model. Dozens of variables like frother flow, collector setpoints, and air setpoints are ranked by their positive or negative impact on recovery prediction, displayed in red and green bars.
The predictive model's feature importance output. While analytically powerful, this visualization was designed for data scientists, not for the shift supervisors and planners who needed to act on the predictions.

How We Worked

Rio Tinto's PACE Analytics team engaged me and a collaborator to close that gap in three weeks. The engagement ran from late November through late December 2019, structured as a compressed discovery-and-design sprint with four phases: project kickoff and discovery, two sprint weeks onsite, and a deliverables documentation phase.

I traveled to the mine site at Kennecott to conduct the research directly. During the first week, I conducted stakeholder and subject matter expert interviews with Rio Tinto representatives both remotely (Brisbane-based staff) and in-person at the Kennecott corporate office. Then I went to the concentrator facility itself to interview and observe the primary users: shift supervisors and shift planners/schedulers.

The onsite sessions were participatory. I didn't just interview people about their work. I worked alongside them and the analytics data science team to brainstorm workflows, requirements, and design concepts in real time. The artifacts generated during these sessions included stakeholder goals and target user needs, design principles and high-level workflows, target user personas and user stories, and low-fidelity sketched workflows and wireframes.

During the second week, I created higher-fidelity wireframe and workflow representations and then brought them back to the mine for follow-up validation sessions with stakeholders, subject matter experts, shift supervisors, and planners. These sessions tested where the designs met user needs and where they didn't, and surfaced latent requirements that hadn't emerged during the first round. After each session, I iterated the workflows and wireframes based on what I'd learned.

The design process was deliberately rapid and iterative. Concepts progressed from hand-sketched whiteboard work in the first stakeholder session, to annotated wireframes after the first supervisor and planner sessions, to mid-fidelity mobile wireframes validated in the second round, to a final annotated wireframe set with desktop variants.

Four photographs showing design artifacts from each onsite session arranged on a timeline. From left to right: hand-sketched wireframes from Stakeholder Session 1, a whiteboard layout with colored annotations from Shift Supervisor and Planner Session 1, a mid-fidelity digital wireframe with trend charts and deviation bars from Shift Supervisor and Planner Session 2, and a detailed annotated wireframe from Stakeholder Session 2. Sprint Week 1 and Sprint Week 2 are labeled below.
Design progression across four onsite sessions. Concepts evolved from rough whiteboard sketches to validated mid-fidelity wireframes over two sprint weeks.

What We Found

Two Personas, Two Time Horizons

The discovery research identified two primary personas with distinct but complementary needs.

Shift supervisors are responsible for improving recovery within their area during each shift. They're walking the floor, monitoring equipment, and troubleshooting when components drift out of spec. Their core questions are tactical and time-sensitive: What affected recovery during last shift? What's affecting recovery right now? Which components are functioning nominally and which need attention?

Their pain points were specific: it was nearly impossible to improve recovery within a shift because they lacked real-time visibility into component performance, and it was difficult to tell whether a component was undergoing slow degradation, intermittent failure, or something else entirely.

Planners and schedulers operate on a different time horizon. Their job is to create shift plans that yield the best possible recovery given the upstream inputs. They possess deep expert knowledge of the recovery process, but the upstream information they rely on isn't always accurate. They need to understand the relationship between planned and actual recovery at a granular level so they can make better plans. Their aspirational need was the ability to run "what if" scenarios using historical upstream data and actual recovery data.

Persona card for the Shift Supervisor. A silhouette of a worker in a hard hat holding a mobile device. Responsibilities include improving recovery, ensuring components function nominally, and monitoring for deviations from planned settings. Pain points: nearly impossible to improve recovery within a shift, difficult to determine if a component is degrading slowly, failing intermittently, or experiencing another type of failure. Key questions listed under Perspectives focus on understanding which settings affected recovery, how components have performed historically, and which row components are out of range in real time.
Persona 1: Shift Supervisor. Tactical, mobile, and time-sensitive — needs real-time visibility into component performance while walking the concentrator floor.
Persona card for the Planner/Scheduler. A silhouette of a worker at a desk with a computer. Responsibilities include creating shift plans that yield the best recovery given upstream inputs. Pain points: upstream information is not always accurate, and it is difficult to make detailed use of actual recovery data at a granular level. Perspectives focus on understanding planned versus actual recovery to improve future plans, with a future aspiration to run what-if scenarios using historical data.
Persona 2: Planner/Scheduler. Strategic and desktop-oriented — needs granular historical data to create better shift plans and understand the gap between planned and actual recovery.

Design Principles

Based on the discovery work, I generated five design principles that framed the V1 scope and future roadmap. These included: provide value to target users through minimum viable data visualizations for row recovery components, make it mobile-friendly since supervisors need data while walking the operations floor, support improved recovery planning, and leverage both extrinsic and intrinsic motivation among operators and supervisors.

The mobile-first requirement was critical. Shift supervisors need to access performance data on their phones while walking the concentrator floor. Planners and schedulers work at desktops for deeper analysis. The application needed to serve both contexts.

What We Designed

The wireframes I produced covered mobile and desktop viewports with three core views.

The performance overview showed the current shift's row recovery data, comparing planned, model-predicted, and actual recovery percentages across circuits and rows. Users could toggle between three time frames: current shift, last four shifts, and a custom date range. Below the trend chart, a "Top 5 Settings Deviations" visualization showed which component settings had deviated most from their targets, immediately surfacing the factors most likely to be dragging down recovery.

The settings detail view let users tap on any deviation bar to drill into that component's historical performance. A horizontal bar chart showed the component's deviation from setpoint across multiple shifts, with a red threshold line making it easy to spot when deviations crossed from normal into concerning territory. Users could select date ranges and sort order.

The desktop viewport reorganized the same data into a wider layout with a persistent left-side navigation. The larger screen allowed more historical data to be visible at once, which served the planners' need for longer-term pattern recognition.

Three mobile wireframes side by side showing the Row Recovery Performance screen. The left wireframe shows the Current Shift view with circuit and row selectors, a trend chart comparing Plan, Model, and Actual recovery percentages, and a Top 5 Settings Deviations bar chart. The center wireframe shows the same layout with Last 4 Shifts selected, displaying updated trend and deviation data. The right wireframe shows a custom Date Range picker with calendar controls and the corresponding trend data.
Mid-fidelity mobile wireframes showing the three time-frame views: Current Shift, Last 4 Shifts, and custom Date Range. Each view displays the recovery trend chart and top settings deviations for the selected period.

I also recommended the inclusion of a feedback mechanism in the mobile menu to encourage continued data collection from users post-launch, and a share-via-email feature so supervisors could quickly send a performance snapshot to a colleague.

Annotated wireframe showing two mobile screens with callout annotations. The left screen displays the app's hamburger menu with navigation items including Dashboard, Row Recovery, Grind Size, Settings, and Feedback, with a callout noting that most menu items are future functionality and recommending a feedback link for continued data collection. The right screen shows the main Row Recovery Performance view with a share button in the header, accompanied by a callout explaining the share-via-email feature that generates a PDF and sends it through the user's default email client.
Annotated wireframes detailing the mobile menu structure and share-via-email feature. Callouts document recommended future functionality and the rationale for including a user feedback mechanism.

What I Learned

This project reinforced something I've seen consistently across domains: the hardest part of deploying a predictive model isn't building the model. It's designing the interface that lets the people who need the predictions actually use them in their workflow. The analytics team had built a model that was demonstrably more accurate than human estimates. But accuracy doesn't matter if the predictions sit in a data science notebook that shift supervisors never open.

The onsite research was essential for this. You cannot design an effective industrial monitoring application from a conference room. The shift supervisors' workflow is physical: they walk the concentrator floor, they check equipment, they make decisions in real time based on what they see and hear. Designing a mobile app that fits into that workflow required watching them do it, understanding the rhythm of their shifts, and building the information hierarchy around their actual decision sequence rather than around the data science team's model architecture.

The compressed timeline was a feature, not a constraint. Three weeks of rapid iteration with real users produced a more grounded design than months of remote specification work would have. Each of the four onsite sessions brought the design closer to something that fit the operational reality. The hand-sketched whiteboard concepts from session one looked nothing like the validated wireframes from session four, and that's exactly the point.