From Static Spreadsheets to Actionable Insights:
Longitudinal Performance Analytics

How interactive dashboards transformed multi-year assessment data into actionable insights for school leadership.

Snapshot Overview

  1. 👨‍💻My Role: Sole analyst

  2. 🔨Tools: Tableau, Excel

  3. 🔢 Data: Multi-year IA & GMAS assessment data (row-level, teacher-grain)

  4. 📊Records: 64,000+ student test scores | 400+ teachers | 4 school years

  5. 👥Stakeholders: District leadership, instructional coaching teams

  6. 🎯Outcome: Replaced manual spreadsheet review with interactive dashboards that enabled trend-based decision-making and prioritized targeted support. This shifted leadership discussions from isolated score review to trend-based coaching decisions, enabling earlier and more targeted instructional support.


The Problem

Problem summary: Leadership had access to years of assessment data, but the format prevented meaningful trend analysis and slowed decision-making.

Georgia Cyber Academy, a K–12 online charter school, had collected multiple years of internal assessment and state standardized testing data across all grade levels. While the data itself was technically complete, it was primarily delivered as raw exports or large Excel pivot tables.

As a result, leadership struggled to quickly answer key questions such as:

  • How individual teachers performed over time, not just in a single assessment

  • Whether changes reflected meaningful trends or short-term fluctuation

  • How performance compared across grade bands, departments, or assessment types

  • Which teachers stood out — positively or negatively — without extensive manual filtering

In practice, answering even basic questions required time-consuming spreadsheet work and subjective interpretation, limiting the data’s effectiveness for leadership-level decision-making.

Example of pre-aggregated reporting that masked longitudinal trends.

The Objective

The objective of this project was to design an interactive analytics solution that enabled district leaders to move beyond static reports and analyze performance trends over time.

Specifically, the solution was designed to:

  • Visualize longitudinal performance at the teacher level, rather than isolated assessment results

  • Classify performance patterns as increasing, stable, or declining

  • Enable dynamic filtering by grade band, department, course, and school year

  • Shift analysis from static numbers to clear, decision-ready visual narratives

The goal was not just to report results, but to support more informed, trend-based decision-making.

The Approach

Approach summary: Built a structured, repeatable analytics workflow using row-level assessment data to support longitudinal, teacher-level analysis.

To address the limitations of pre-aggregated reporting, I worked directly with raw IA and GMAS datasets rather than relying on Excel pivot tables. This enabled deeper longitudinal and trend-based analysis across multiple school years.

My approach included:

  • Data preparation and cleaning

    • Requested and worked directly with raw datasets spanning multiple years.

    • Verified record completeness across teachers, assessments, and school years

    • Cleaned the data by removing duplicate records and resolving inconsistencies in teacher identifiers, course labels, and assessment naming

    • Developed calculated fields to normalize assessment performance across multiple testing windows, enabling accurate longitudinal trend comparisons.

  • Data modeling for longitudinal analysis

    • Structured the raw data to support year-over-year and assessment-to-assessment comparisons

    • Created calculated fields to track performance changes across IAs within a school year and across multiple years

    • Established consistent sorting and ordering logic to preserve instructional timelines and assessment sequences

  • Visualization and interaction design

    • Built interactive Tableau dashboards that allowed leadership to filter by grade band, department, course, and individual teacher

    • Developed trend classifications (increasing, flat, decreasing) to quickly surface teachers whose student performance was improving, stagnating, or declining

    • Designed views that made growth patterns, outliers, and long-term trajectories immediately visible without requiring technical interpretation

This approach transformed raw, multi-year assessment data into an executive-friendly tool that supported exploration, comparison, and data-informed decision-making at the district level.

Raw assessment data required validation and standardization before longitudinal analysis was possible.

Skills Applied

  1. Data Cleaning & Validation
    Standardized identifiers, resolved inconsistencies, and validated multi-year assessment records to ensure accurate longitudinal analysis.

  2. Longitudinal Data Modeling
    Structured row-level data to support year-over-year and assessment-to-assessment comparisons at the teacher level.

  3. Trend & Pattern Analysis
    Classified performance trajectories as increasing, stable, or declining to distinguish sustained trends from short-term fluctuation.

  4. Dashboard Development (Tableau)
    Built interactive dashboards with multi-dimensional filtering to support exploration by grade band, department, course, and teacher.

  5. Stakeholder-Focused Analytics
    Translated complex performance data into clear, decision-ready visual narratives aligned to leadership decision-making workflows.

Early iteration (high visual density)

Refined iteration (clarity & trend visibility)

GMAS Dashboard

The GMAS dashboard focuses on longer-term outcomes, providing a summative view of performance across school years. Because GMAS is administered annually, this view emphasizes year-over-year trajectories rather than short-term fluctuation.

Key features include:

  • Multi-year performance trend analysis at the teacher level

  • Consistent trend logic aligned with the IA dashboard

  • Contextual filtering to compare patterns across grade bands and departments

Together, these dashboards connect short-term instructional signals with long-term achievement outcomes, enabling leadership to interpret performance trends with greater clarity and confidence.


Example of teacher-level IA trends highlighting short-cycle instructional response.

The Solution

To address these challenges, I developed two complementary Tableau dashboards—one focused on Interim Assessment (IA) data and the other on GMAS data—using a shared analytical framework and consistent design logic.

Design rationale: Interim Assessments (IA) and GMAS serve different analytical purposes. Separating the dashboards preserved contextual accuracy while allowing consistent trend logic across short-cycle instructional signals and long-term outcomes.

Although both dashboards analyze performance trends over time, this separation allowed each view to remain easier to interpret while still supporting meaningful comparisons across school years.

Interim Assessment (IA) Dashboard

The IA dashboard was designed to capture short-cycle instructional trends within and across school years. With multiple assessment points per year, this view allows leadership to see how teacher performance shifts over time in response to instruction, pacing, and intervention.

Key features include:

  • Assessment-to-assessment trend visualization within each school year

  • Trend classification to surface increasing, stable, or declining performance

  • Filtering by grade band, department, course, and teacher for targeted analysis

Key Challenges

  • Pre-aggregated reporting limited longitudinal insight
    Initial access to the data relied heavily on Excel pivot tables that summarized performance at a single point in time. While technically complete, these summaries obscured year-over-year patterns and made it difficult to distinguish sustained trends from short-term fluctuation.

  • Inconsistent identifiers across datasets and years
    Teacher names, course labels, and assessment structures varied across school years and data sources. Without careful validation and standardization, these inconsistencies introduced significant risk of inaccurate comparisons over time.


Design Iteration: From Density to Clarity

Initial dashboard iterations revealed challenges related to visual density when displaying hundreds of longitudinal trajectories simultaneously. Iterative refinement focused on reducing noise, improving readability, and surfacing meaningful patterns without sacrificing analytical depth.

Key Insights & Impact

The dashboards surfaced patterns that were difficult to detect using static reports or single-year views. By visualizing performance longitudinally, leadership was able to shift focus from isolated results to sustained trends over time..

Key Insights

  • Performance trajectories were more informative than single data points

    • Teachers with similar scores in a given assessment often showed very different trajectories across assessment cycles. Viewing growth patterns over time provided a clearer picture of instructional impact than snapshot comparisons.

  • Short-cycle IA trends often not aligned with long-term GMAS outcomes

    • In many cases, trends observed in Interim Assessment (IA) data did not directly translate to GMAS outcomes. Teachers who showed short-term gains or declines across IA windows often demonstrated different patterns in summative GMAS results.

      This highlighted an important distinction between short-cycle instructional signals and long-term performance outcomes, reinforcing the need to interpret IA trends as contextual indicators rather than direct predictors of summative result

  • Trend-based views helped prioritize support more effectively

    • Rather than reacting to isolated low scores, leadership could identify teachers showing sustained decline and differentiate them from those experiencing temporary fluctuations. This allowed for more targeted coaching and resource allocation

  • Context mattered

    • Filtering by grade band and department revealed that performance patterns were not uniform across instructional contexts. This helped leadership frame conversations more accurately and avoid one-size-fits-all conclusions.

Impact

As a result of this work:

  • Leadership moved from manual spreadsheet review to visual, repeatable analysis

  • Conversations shifted from “How did this assessment go?” to “What is the trend over time?”

  • Data was used more consistently to inform instructional planning, coaching, and decision-making

  • The dashboards became a scalable framework that can be reused and extended in future assessment cycles

Reflection & What’s Next

This project reinforced the importance of designing analytics around decision-making, not just reporting. While the dashboards successfully enabled longitudinal insight, future iterations could focus on automating data ingestion, introducing alerts for emerging negative trends, and expanding segmentation to better support proactive instructional planning. If extended further, the next phase of this work would focus on:

  • Incorporating additional contextual variables (such as enrollment changes or course placement) to better explain variation in trends

  • Exploring more formal statistical methods to quantify the strength and stability of observed trajectories

  • Future iterations could include automated ingestion and SQL-based transformations to support near–real-time monitoring of performance trends..

Most importantly, this project demonstrated how thoughtful data modeling and visualization can turn underutilized historical data into a practical decision-support system. Rather than replacing professional judgment, the dashboards were designed to inform better questions, support more targeted conversations, and encourage evidence-based action.