Assessment Applications Design & Development

Screen capture of Scenario-Based Task ‘Bicycle Materials’ for NAEP grade 12 Science. 1 of 14 SBTs I developed.

Context

Large-scale assessment programs required interactive digital experiences capable of measuring complex skills in authentic contexts. Traditional item formats could not capture multi-step reasoning, tool use, or applied decision-making.

Each task functioned as a self-contained web application while needing to integrate with platform constraints, scoring systems, and telemetry pipelines in a regulated, high-stakes environment.

Role & Scope

As an individual contributor, I designed and delivered 14 full-scale scenario-based assessment tasks, each implemented as an interactive application.

Each task included:

  • Storyboard defining the content and user flow through the task
  • Custom UI components and multi-state interaction logic
  • Embedded tools and task-specific workflows
  • Scoring rules and response modeling
  • Metadata schemas and API integration
  • Telemetry capture for machine scoring and analysis

Approach

I owned the work end-to-end across design, specification, and delivery:

  • Collaborated with subject matter experts (SMEs) to translate domain content into interactive, assessable experiences
  • Developed and maintained detailed storyboards for each task, guiding design, development, and client review through approval
  • Designed user experiences through interaction modeling, storyboarding, and UX design
  • Translated assessment intent into structured, machine-scorable interactions
  • Authored detailed functional specifications defining system states, transitions, navigation, scoring logic, and edge cases
  • Served as product owner through vendor development, managing sprint cycles, clarifications, and change control
  • Coordinated QA, platform integration, and user acceptance testing to ensure fidelity, usability, and correctness

This work required tightly coupling interaction design, system behavior, and measurement goals so that student actions could be captured as valid, interpretable evidence.

Results

  • Delivered 14 production-grade interactive assessment applications
  • Enabled measurement of complex, multi-step skills through simulation-based tasks
  • Established consistent patterns for integrating interaction design, scoring logic, and telemetry capture
  • Reduced ambiguity in implementation through detailed, system-level specifications

Key Takeaway

Interactive assessment design is fundamentally a systems problem—effective solutions require aligning user interaction, technical implementation, and how evidence of student thinking is captured and scored.

Show and Tell (What We Built)

The main type of application I worked on were NAEP Scenario Based Tasks. The examples below are 2 SBTs that I personally worked on and were released to the public: Clear Water and Bicycle Materials. While most SBTs I worked on were from ideation through production release. When I took these two over, they had already been designed and released to pilot with test takers. My work on these two included revising design and content following pilot results and modernizing features and functions, such as improving accessibility.

The early days of SBT (Scenario Based Taks) design were characterized by creative exploration to develop engaging and authentic contexts, designs, and user interface elements to meet assessment measurement goals. These early SBTs had a wide range of presentations.

​SBTs begin with establishing an authentic and engaging context for the task.

Test taker introduced to the ‘Clear Water’ task by Mr Boyd with an image showing a muddy river flowing into a lake.
Test taker introduced to the ‘Bicycle Materials’ task introducing the scenario of identifying strong light-weight materials suitable for building a bicycle frame.

Test takers would be introduced to the context and guided through an activity solve a problem (engineering) or conduct an investigation (science). This format allowed the test taker to apply their knowledge and skills in an authentic context and allowed the SBT developers (myself as designer and assessment specialists) to develop item content that was more integrated with the larger task.

For example, in ‘Clear Water’ students were able to use a simulated apparatus to test the turbidity of water and collect data.

After the test taker is introduced to the Secchi disk, they are asked to collect data, by using the UI to lower the Secchi disk into the animated water and record the depth where the water is too deep to see the disk in the provided data table.
Test takers use the interface to lower the Secchi disk into the turbid lake water to collect data on turbidity. This concept and analysis is used later in the task.

The SBTs also supported novel item compositions. Not all compositions were ideal, but NAEP gathered valuable data from student play-testing and usability studies as well as direct data from test takers during pilot assessment administrations.

Multiple-select multiple-choice item in ‘Bicycle Materials’. Item stimuli provided by the in-task guide as spoken text with speech bubble and playback control. A periodic table wraps around the item response options. It worked, but not sure this would get approved if designed today.
Three-part composite item includes a zone item where test taker selects the appropriate electron in the atomic model, answers a single-select multiple-choice item, then proves an explanation in a text constructed response field. This item provided a challenge for keyboard navigation so as to not cue the test-taker regarding the correct possible answers.

While most of my SBT designs are confidential assessment content, this overview captures some of the flavor of the design and development for these first of their type large-scale assessment experiences. I’ll continue expanding this section with detailed case studies as I publish more of my portfolio.