top of page

2026

Creating a Question on RightOn

Redesigning how teachers create questions for games on RightOn’s platform, focusing on improving speed to question completion by introducing a more controlled and transparent AI-assisted workflow.

Project Focus

MVP Redesign

My Role

UX/UI Intern

Timeframe

70 hrs

Tools

Figma

Hero 01.jpg
Hero 02.jpg
Hero 03.jpg

Situation

My Focus

Impact

RightOn is a platform where teachers can create questions for games to help students learn from mistakes and misconceptions.​

The question creation experience had recently been redesigned to improve the manual workflow, but creating high-quality questions was still time-consuming and effort-heavy.

Design the AI-assisted question creation flow to reduce effort and improve clarity, while fitting into the redesigned manual workflow.

Product

Improved control, transparency, and usability of AI-assisted content creation.

Final design Cards Image.jpg

The Problem

Teachers needed a fast way to create questions that fit the context of their classroom, but the existing workflow was confusing, time-consuming, and inefficient. AI was introduced to help, but lacked clarity and control, reducing trust and usefulness. As a result, teachers often defaulted to manual workflows or dropped off the platform altogether.

Creating questions wasn’t just tedious, it was frustrating and unpredictable.

Key Insight

After the initial redesign without AI, teachers no longer had trouble creating questions, but they needed a faster way to complete them, especially generating answer options and explanations.

  • AI had the potential to help, but only if it fit into their existing workflow

  • The existing AI tool felt unclear or unpredictable, adding friction instead of reducing it

How might we help teachers save time creating questions that fit the needs of their classrooms?

Sketches 1.jpg

What if we improved the existing AI generation tool that was meant to improve speed to question completion.

KEY DECISIONS

Flexible & guided AI entry point

A prompt-based entry point was created that allows different levels of generation depending on user intent when creating a question.

Instead of forcing a single workflow, the system adapts to different needs.

Before

Decision 1 Before.jpg

After

Decision 1 after.jpg

Controlled regeneration

I designed a system that allows users to regenerate individual answer options, regenerate selected groups, or regenerate everything.

This shifts AI usage from all or nothing to iterative and controllable.

Before

Regenerate Before.jpg

After

Decision 2.jpg

Make AI behavior more transparent

Users had difficulty understanding what would happen and when it would happen when using the old AI tool, so I refined the interaction patterns and language with more intentional placement of AI controls.

This made the system predictable and not feel like rolling a dice.

Before

Regenerate Before.jpg

After

Decision 3.jpg

TESTING & ITERATION

Early versions of the AI flow were tested with teachers using realistic classroom scenarios.

What did I learn?

While the initial workflow was directionally strong, usability testing revealed gaps in clarity that required refinement rather than a full redesign.

Key insight 1

Validated direction

  • Users were able to complete key tasks

  • The structure of the flow made sense.

  • AI-assisted creation showed promise for teacher needs.

Key insight 2

Lack of clarity

  • The scope of AI generation was still not fully clear.

  • There was confusion around what certain signifiers trigger, such as “Regenerate”.

What changed after testing?

Clarifying the scope of AI

Two unique AI prompt entry points were created for two separate actions: completing a question or creating a full question from scratch.

This removes ambiguity on what an entered prompt is for and what information is needed from users before AI performs a task.

Before

Prompt entry Before 02.jpg

After

Answers only or full question.jpg

Improved signifiers

UI iterations were made to improve clarity on critical action triggers for question answer option iteration.

Before

Controls pre testing v2.png

After

Controls Final v2.png

Before

Unlocked or locked before.jpg

After

Unlocked or locked After.jpg
Full window image.jpg

REFLECTION

Postponing possible solutions

Features like short-phrase “filters” for prompting were explored but ultimately deprioritized due to:

  • Cognitive load concerns

  • Implementation complexity

  • Time constraints

Prompt card with filters.png

Impact

Early feedback indicated that the experience felt more intuitive and aligned with how teachers actually work.

  • Made AI-assisted creation more structured and predictable

  • Gave teachers clearer control over what gets generated and updated

  • Reduced confusion around how AI behaves within the system

What's Next?

There are still plenty of opportunities to improve the experience. One area explored but not included in the MVP was short-phrase filters for AI prompting.

 

These could:​

  • Speed up repetitive prompt parameters

  • Improve onboarding for AI interactions

  • Further refine generation transparency

bottom of page