Reimagining Petavue’s Data Hub to make structural and semantic data changes predictable, traceable, and safe.
Petavue is a B2B AI analytics platform built for RevOps and data teams. At its core, the product helps companies trust their reporting, but that trust depends entirely on how well the underlying data is configured. The Data Hub is where that configuration happens. It's where admins define what data exists, what it means to the business, and how it connects to every report, metric, and dashboard in the product. When it works well, it's invisible. When it doesn't, reports break in ways that are hard to trace and harder to explain.
By the time this project started, the Data Hub was functional but fragile. Admins could make changes that looked routine but caused failures they wouldn't discover until something downstream broke. The work was about closing that gap and making the system's consequences visible before damage occurred.
Team
2 Designers
1 Product Manager
2 Developers
Product Design, Visual Design, Design System, User Flows, Prototyping
Tech Stack
Figma

Re-design the Data Hub as a reliable system layer where changes to data structure and meaning are predictable, explainable, and safe.
At its core, it answers two questions:
What data exists in the system
What that data means to the business
These questions are handled by two equal pillars.
The work focused on:
Making the impact of Data Dictionary changes visible before they break reports
Turning Key Definitions into inspectable, shared representations of business logic
Reducing hidden dependencies between structure, meaning, and reporting
Helping admins move fast without causing silent failures downstream
The Data Hub allowed admins to make powerful structural and semantic changes, but the system gave little visibility into their downstream impact. As a result, small configuration updates could unexpectedly break reports across the product.

We saw this pattern in how admins described their workflow. They'd make a change, assume it went through, and only find out something broke when a report failed hours later. The action and the consequence were separated by time and context.
Disabling a column, removing a relationship, or updating a definition were treated as routine UI actions.
In reality, these changes could affect multiple downstream reports, drill-downs, and metric definitions.
Because the system did not surface these dependencies beforehand, admins often discovered the impact only after reports failed.

This came up repeatedly when engineers got pulled in to debug. The troubleshooting process always started with the same question: what changed recently? The system had no answer.
When a report failed, users saw generic execution errors.
There was no clear path back to the change that triggered the failure. Troubleshooting meant manually investigating tables, columns, and definitions to identify the issue.
This made the system feel unreliable even when the underlying data was correct.
Admins managing definitions often didn't know which tables or columns they were drawing from. That information lived in the underlying query, readable only if you knew what to look for.
The Data Dictionary controlled structure, while Key Definitions controlled business meaning.
However, these surfaces were disconnected. Admins could update one without clearly seeing how it affected the other.
This separation hid critical dependencies between:
tables and definitions
columns and metrics
paths and drill-downs
The biggest risk in the Data Hub was not incorrect data. It was invisible impact.
Correct changes could still break reports.
Valid updates could silently disable definitions.
Failures appeared far away from their cause.
We focused on making consequences visible before damage occurred.
We approached the Data Hub as a system, not a collection of screens.
Instead of redesigning individual pages in isolation, we focused on how changes in one part of the Data Hub affect the rest of the product.
Every decision was evaluated based on how clearly it surfaced dependencies and how predictable the outcome felt to the user.
The approach centered on three principles:
01
Surface impact before execution
High-impact actions like disabling tables, columns, or paths should never behave like silent toggles.
The system now shows what reports, definitions, and drill-downs depend on a change before the admin confirms it.
02
Make dependencies inspectable
Relationships between sources, tables, columns, and definitions are now visible inside the interface.
Admins can trace how data flows from structure to meaning and understand what a change will affect.
03
Treat structure and meaning as one system
The Data Dictionary controls data availability.
Key Definitions control business interpretation.
This was a two-designer project with a PM and two developers. Most structural decisions, what to surface, where to intervene, how much friction to add, were made collaboratively with the PM and senior developer who understood the edge cases better than anyone. My role was translating those constraints into interface decisions, finding the design pattern that matched the technical reality without overwhelming the admin with complexity they didn't need to see.
The Data Dictionary is where admins manage everything the system knows about data — sources, tables, columns, relationships. Every report, chart, and metric in Petavue depends on decisions made here. It is not a settings page. It is the foundation.
The existing screen did a reasonable job of presenting information. Tables were organized in a left-side nav, columns were listed with data types, descriptions, and sample data, and a toggle on each row controlled whether a column was available for analysis. On the surface, it looked functional.
Dictionary list page






