Build To Solve: 2025 Power BI Competition

An image of the three Support dashboards created during this competition.

Overview

Problem: The Kahua Support team lacked proper visualization into data on key aspects of their process. These included the volume of tickets per application category and the number of Service Level Agreement (SLA) response and resolution violations over time.

Solution: Using data from interviews with four members of support, my team and I built out a comprehensive dashboard. With multiple pages, each built around a central research question, we investigated relationships between resolution times and ticket type, categories and ticket volume, and the influx of tickets over the past three years. This allows the Support team to measure the effectiveness of their current approach and make a case for improvements across the company (such as enhanced documentation and support materials).

Role: UX Research, UI Design, Data Analysis
Team: 3; Jennifer Tran, Sofia Torres Moreta, and myself
Tools: Power BI, Figjam, Microsoft Teams
Duration: July 9th - July 24th
Note: The data displayed in these dashboards is synthetic and does not reflect real Kahua Support statistics.
Screenshot of the SLA Compliance Overview dashboardScreenshot of the Scheduling Insights DashboardScreenshot of the Ticket Flow and Resolution dashboard.

Project

We divided our project into four phases:

  1. Preparation & Planning
  2. Research
  3. ETL (Extract, Transform, Load) & Mockup
  4. Development

These are explained in detail below.

Phase 1: Preparation & Planning

Throughout this project we used the ETL (extract, transform, load) process to pull and display our data. In addition, we conducted loosely structured interviews with key stakeholders. We wanted our stakeholders to be able to not only easily comprehend and use our dashboards, but to gain valuable information from them. These interviews served as the foundation of our project and informed our design and development decisions.

Phase 2: Research

The research phase was by far the largest in this effort, making up more than half of our time working on the project.

We began by conducting introductory research on our own. Then, we met with multiple members of support and posed questions such as those listed below.

  1. How does your team use the data currently available to you? How does it benefit or not benefit your team?
  2. What tools do you currently use to track tickets — resolution times, response rates, status, etc — if any? Are these tools effective or ineffective to your current workflow?
  3. How do you currently determine which tickets or issues need attention first? Can you walk me through that process?
  4. What does a successful resolution look like for your team, and how do you measure that today?

After our interviews we created an affinity map of our meeting notes in order to uncover key patterns. We discovered that the support team was interested in delving into...

User interview affinity maps done in Figjam.

With this knowledge, we paired down our research questions to the following four.

With the intent of creating the maximum three pages, we began the next phase: extracting, transforming, and analyzing the data and mocking up our dashboards.

Phase 3: Extract, Transform, Analyze

As Jennifer transformed the data, we began sketching mockups.

Preliminary dashboard layout sketches.

We accounted for space on the page, volume of data, usability, and visual appeal when choosing our graphs and tables. We aimed to present the data in a way that directly answered our research questions, thus meeting the Support team’s needs to the best of our ability.

Once we agreed on the layout of our dashboards, we scheduled a follow-up meeting with members of the support team we had previously interviewed. With their feedback in mind, we made changes to what visuals we utilized and how we presented the data. Then, we were ready to start developing.

Final dashboard mockups.

Phase 4: Development

The development phase was our last and fastest phase. To stay consistent and grounded, we continued asking the Support team for clarification of the support process. As we gained more knowledge, we updated our design to ensure we answered our research questions and portrayed the data in unique, helpful, and readable ways.

Finally, we met with our advisor for final feedback before submitting our dashboards. Taking her thoughts into consideration, we changed a few visuals and added some filters for further clarity. 

Screenshot of the SLA Compliance Overview dashboardScreenshot of the Scheduling Insights DashboardScreenshot of the Ticket Flow and Resolution dashboard.

Conclusion

The competition judges applauded the visual design and usability of our dashboards. However, it was the foundation of research that impressed them the most. This was a primary factor in our award of first place. 

Interviews with stakeholders kept our vision aligned with our users’ goals and points of view throughout the process. Feedback from our advisor offered a view into the conventions of making data presentable and comprehensible. Thus, we could not only start strong, but continue to improve our product along the way. 

Takeaways