Enterprise AutoML Workflow Redesign

The configuration workflow for dotData's AutoML platform was technically correct but impossible to use. I redesigned it from a linear wizard into a canvas workspace — cutting setup time in half and letting business analysts run models without a data science team.

Role

Senior Product Designer

Product

dotData Enterprise AutoML Platform

Team

2 designers, engineering

Focus

Workflow Design, Complex Systems UX

dotData Enterprise AutoML configuration workspace overview

The Problem

The primary users were business analysts — people who understood their domain well but had limited hands-on experience with machine learning. Most of them were introduced to dotData by their organization, not by choice. Before the platform, many had relied on Excel or leaned on their data science team to run analyses.

That dependency didn't go away when dotData arrived. During configuration, analysts would typically work through settings in regular check-ins with a dotData data scientist — and in many cases, the DS would end up configuring the model on their behalf. The tool existed to give business users independence, but in practice, it still required expert support to operate.

The cost of a misconfiguration was hours. Model training could take one to two hours depending on data size. If a user set the wrong data type, they'd have to re-upload the dataset and start over from scratch. If the model ran but accuracy came back low, they often had no idea why. Most of dotData's customers didn't have an internal data science team — that's exactly why they were using the platform. But every time something went wrong, they were filing it into our DS team's calendar.We were absorbing the cost of a UX problem as customer support hours.

Expert Trap

The interface was graphical, but the concepts were still framed in data science terminology. A business analyst had no way to know what “prediction target time” meant or why getting it wrong would break everything downstream.

Invisible Ripple Effects

Configuration changes silently altered the dataset used for modeling. The only signal was a low accuracy score at the end — with no explanation attached.

Rigid Workflow

The step-based wizard forced users into a linear sequence. If something went wrong, there was no way to adjust a single parameter — users had to restart from the beginning.

Understanding the System

Going into this project, I assumed the problem was mostly terminology. The interface used data science language, and business analysts weren't data scientists — so the fix should be a language problem, right?

An engineer pushed back on that early. He told me to take a real dataset and try to configure a task myself. See if I could get it right without help.

I couldn't.

That changed how I thought about the problem. I started reading documentation, asking engineers to walk me through the pipeline logic, and using ChatGPT to translate technical concepts into plain language with concrete examples. Slowly I built up a mental model of what the system was actually doing. And that's when I realized the terminology wasn't the real issue. The issue was that the interface gave users no way to understand the consequences of their decisions before it was too late to fix them.

Time column → Prediction Target Time → Lead Time → Feature Window

These four parameters are tightly coupled. Each one determines how the system slices data for training. The interface never communicated that — and without understanding it, designing something helpful was impossible.

From Wizard to Workspace

Our first instinct wasn't to scrap the wizard. We spent time exploring how to make it better — clearer section labels, easier navigation between steps, more explicit definitions for each parameter. But the more we dug in, the more something felt off.

At some point I realized: what users were actually doing, across all these steps, was defining a single flat table. That table was the training dataset. Every parameter — time column, prediction target time, lead time, feature windows — was just determining what that table would look like. The wizard was hiding that. It was breaking one coherent decision into a sequence of isolated steps, so users never had a mental model of what they were building toward.

That's when the canvas idea came in. If the real task was constructing a dataset, the interface should make that visible — not walk users through it one field at a time.

Linear Wizard

Simpler navigation, but delays feedback about configuration mistakes. Users can't see how steps connect.

Pros: simpler navigation

Cons: delayed error discovery, hidden dependencies

Decision Workspace

Higher information density, but users can immediately observe system behavior and iterate without restarting.

Pros: faster iteration, visible dependencies

Cons: more complex layout

We went with the workspace because early feedback was more valuable than reducing interface complexity.

Design Details

Prediction Target Time

One of the trickiest parameters to design around was Prediction Target Time — the reference point the system uses to align historical data with training samples. It appears in multiple steps, and users almost never understood why it kept showing up.

The fix: instead of asking users to navigate back to configure it, we let them set it inline wherever it appeared. A small edit icon opens a lightweight panel — the workflow never breaks.

Prediction Target Time inline configuration

Auto-Connecting Tables

Connecting tables was another drop-off point. Business analysts aren't database engineers — manually selecting join keys for each table connection caused a lot of errors.

We introduced auto-connect: when a user adds a table, the system scans the schema and suggests a likely relationship based on column matches. Users can override it, but the default is usually right. That alone eliminated a whole category of silent failures.

Validation

We tested with three external participants — master's students in business analytics — given a real dataset and a concrete problem to solve: predict which servers were likely to fail in the next seven days. A PM briefed them on the scenario beforehand. During the sessions, I sat back and observed, tracking behavior rather than what they said.

One thing stood out immediately. All three skipped every tooltip and description text without reading it — not because they were rushing, but because they were focused on getting through the task. They weren't learning the system; they were trying to use it. That confirmed that any solution relying on user education would fail in practice.

One tester flagged something that directly shaped a later design decision: she found herself going back and forth between the description text and the configuration panel trying to understand prediction target time, and suggested the concept should be interactive rather than explained in static text. That observation led directly to the timeline visualization.

“The canvas flow is so much smoother than before.”

dotData DS team, post-launch

Impact

50%

reduction in configuration time

10 min → 5 min, in benchmarks where users knew the task and hit no errors

dependency on the data science team for setup

business analysts could complete configuration independently

configuration errors from silent failures

smart defaults and inline validation caught mistakes before training ran

The bigger change was around iteration. Users could now adjust any part of the configuration without starting over — something that wasn't possible in the original workflow. The CS team flagged this as one of the most noticeable improvements after launch.

Reflection

When I first looked at this system, it felt like a black box. Users put data in, waited an hour or two, and got a result they couldn't interpret. My assumption was that the fix was simple — just show people more.

That turned out to be wrong. The question I kept coming back to was: does the user actually need to know this? Or is it just going to make them feel like they're doing something wrong? Some information helped. Some just made people second-guess themselves. Most of the real design work was figuring out which was which — and I'm still not sure we drew the line in exactly the right place.

One thing I wanted to push further but didn't get to: closing the loop between configuration and model performance. Right now users set things up, wait, and get a result — but they can't connect what they configured to what came out. That gap is still there, and I think it's the most important thing to go after next.

Concept exploration

If I were designing this today

The workspace redesign solved the visibility problem — users could finally see what their decisions were doing. But it still assumed something that wasn't always true: that users already knew what problem they wanted to solve.

In practice, a lot of business analysts came to dotData with data but without a clear use case. They needed the system to help them figure out what was worth building — before they could configure anything.

With agentic AI, that's now a solvable problem. Instead of handing users a blank configuration form, the system can analyze their data first, identify what's possible, and guide them toward a model that actually fits what they have.

Three design decisions that changed

01

From form-first to conversation-first

The original workspace started with a blank configuration panel — users had to know what they wanted before they could do anything. The new flow starts with a question: what problem are you trying to solve? The system uses the answer to recommend a use case, explain why the data supports it, and pre-fill everything it's confident about. Users only make decisions where human judgment actually matters.

02

Data quality as a conversation, not a gate

The original system surfaced data errors late — sometimes only after training ran. The new flow detects issues immediately after upload and handles them one at a time, showing the user exactly what changed and why. Missing values, inconsistent date formats, ambiguous column types — each gets surfaced with a concrete before/after and a plain-language explanation, not a technical error message.

03

Two-layer sampling for faster iteration

Enterprise datasets are often large. Waiting 20 minutes to discover a configuration mistake is expensive. The new flow separates two distinct decisions: how much data to use for configuration (where a 10% sample is almost always enough), and how much to use for the actual model run (where you can do a quick run first to validate direction before committing full compute).

I built an interactive prototype to explore this direction. It uses real retail data — 45 stores, 79 departments, 143 weeks of sales history — and walks through the full flow from upload to model configuration.

Try the prototype →

Click "Load sample data" to start