UI / UX Design

Brainly AI-Powered Test Prep

Executed a data-driven design strategy by combining qualitative interviews, quantitative surveys, and competitive audits to build a zero-input, adaptive testing engine that reduced prep time by 40%.

Year :

2025

Industry :

EdTech

Employer :

Brainly

Project Duration :

8 weeks

Overview

Students shouldn't have to study how to study.

I led design end-to-end on Brainly Test Prep, an AI feature that turns the homework students already do into a personalized, adaptive study plan. No flashcards to build. No "where do I even start" panic the night before an exam.

First 60 days post-launch:

  • +30% weekly engagement vs. control

  • −40% self-reported study setup time

  • 4.5 / 5 average CSAT

(Brainly serves 15M+ MAU, so small UX wins compound fast.)


The problem

Brainly's own research told the story: 28% of US students name exams as their #1 source of school anxiety, and 68% admit they don't give themselves enough time to study. The gap wasn't motivation. It was knowing where to start.

Tools like Quizlet and StudySmarter only made it worse. They assume students will sit down, organize materials, and build their own study set. Cognitive load on top of cognitive load.

Our hypothesis: if we removed the setup work entirely and met students where their materials already lived (Brainly already had their homework and Q&A history), we'd turn studying from a planning problem into a doing problem.

Goals & KPIs

Three things we committed to before any pixels got pushed:

  1. Engagement: +20% weekly returns from Test Prep users vs. non-users

  2. Time-to-first-quiz: under 60 seconds from upload

  3. Satisfaction: ≥ 4.0 CSAT in the first month


Research

20 interviews (ages 13–19) and 500+ surveys. The qual surfaced an emotional layer I hadn't expected — students were carrying expectations from parents, peers, and teachers, and the planning of study time was where it all collapsed. Multiple kids said some version of "I open my notebook and just stare at it."

The quant confirmed it: 80% preferred automation over building their own flashcards, 60% felt overwhelmed during exam season, and 55% specifically asked for daily reminders.

Students didn't want more features. They wanted something to tell them what to do next.


Competitive landscape

A full audit showed a clear white space: nobody had nailed zero-input ingestion plus mistake-driven adaptive review.


Capability

Brainly

Quizlet

StudySmarter

Khan

Zero-input homework ingestion

AI-generated quizzes

🔸

Adapts to student mistakes

🔸

Study plan tied to exam date

🔸

Designing for AI uncertainty

This is where AI product design gets interesting. Model outputs aren't deterministic, and a 7th grader can't spot a hallucinated question. Two principles shaped every screen:

Make AI confidence legible. Every generated question shows topic and difficulty up front, so students know what's being tested.

Always give an out. Every quiz has a "this seems off" flag. Hides the question for that student, feeds back into retraining. Two birds.


The pivot

V1 had separate spaces for homework, quizzes, study plans, and past reviews. Five participants into a 15-person usability test, I had my answer. Students were treating the quiz library and the study plan as different products. One 9th grader said "wait, this is the same thing?" and she was right.

Killed it. Consolidated into a single dashboard organized around upcoming exams, with quizzes nested under the exam they prep for. Next round of testing finished in half the time.

The lesson I keep coming back to: AI features tempt you to expose the machinery. Resist it. Students don't care about the model pipelines. They care about the test on Friday.

The shipped flow

Homework done in Brainly → tap "Add to Test Prep" → pick exam date → AI generates study plan + practice questions → daily nudges scheduled around the student's school calendar → performance feeds the next session (wrong answers come back harder, mastered topics space out).






Reflections

Empathy is structural, not decorative. When your user is anxious, every design decision either adds to that load or removes it. There's no neutral.

Restraint beats power. The temptation with AI is to show off everything the model can do. The better move is to let the model recede so the student feels competent.

For v2: tighter feedback loop on flagged questions (retraining lagged behind flag volume), and a proper parent surface — they're a real influence on study behavior and we skipped them in v1.

UI / UX Design

Brainly AI-Powered Test Prep

Executed a data-driven design strategy by combining qualitative interviews, quantitative surveys, and competitive audits to build a zero-input, adaptive testing engine that reduced prep time by 40%.

Year :

2025

Industry :

EdTech

Employer :

Brainly

Project Duration :

8 weeks

Overview

Students shouldn't have to study how to study.

I led design end-to-end on Brainly Test Prep, an AI feature that turns the homework students already do into a personalized, adaptive study plan. No flashcards to build. No "where do I even start" panic the night before an exam.

First 60 days post-launch:

  • +30% weekly engagement vs. control

  • −40% self-reported study setup time

  • 4.5 / 5 average CSAT

(Brainly serves 15M+ MAU, so small UX wins compound fast.)


The problem

Brainly's own research told the story: 28% of US students name exams as their #1 source of school anxiety, and 68% admit they don't give themselves enough time to study. The gap wasn't motivation. It was knowing where to start.

Tools like Quizlet and StudySmarter only made it worse. They assume students will sit down, organize materials, and build their own study set. Cognitive load on top of cognitive load.

Our hypothesis: if we removed the setup work entirely and met students where their materials already lived (Brainly already had their homework and Q&A history), we'd turn studying from a planning problem into a doing problem.

Goals & KPIs

Three things we committed to before any pixels got pushed:

  1. Engagement: +20% weekly returns from Test Prep users vs. non-users

  2. Time-to-first-quiz: under 60 seconds from upload

  3. Satisfaction: ≥ 4.0 CSAT in the first month


Research

20 interviews (ages 13–19) and 500+ surveys. The qual surfaced an emotional layer I hadn't expected — students were carrying expectations from parents, peers, and teachers, and the planning of study time was where it all collapsed. Multiple kids said some version of "I open my notebook and just stare at it."

The quant confirmed it: 80% preferred automation over building their own flashcards, 60% felt overwhelmed during exam season, and 55% specifically asked for daily reminders.

Students didn't want more features. They wanted something to tell them what to do next.


Competitive landscape

A full audit showed a clear white space: nobody had nailed zero-input ingestion plus mistake-driven adaptive review.


Capability

Brainly

Quizlet

StudySmarter

Khan

Zero-input homework ingestion

AI-generated quizzes

🔸

Adapts to student mistakes

🔸

Study plan tied to exam date

🔸

Designing for AI uncertainty

This is where AI product design gets interesting. Model outputs aren't deterministic, and a 7th grader can't spot a hallucinated question. Two principles shaped every screen:

Make AI confidence legible. Every generated question shows topic and difficulty up front, so students know what's being tested.

Always give an out. Every quiz has a "this seems off" flag. Hides the question for that student, feeds back into retraining. Two birds.


The pivot

V1 had separate spaces for homework, quizzes, study plans, and past reviews. Five participants into a 15-person usability test, I had my answer. Students were treating the quiz library and the study plan as different products. One 9th grader said "wait, this is the same thing?" and she was right.

Killed it. Consolidated into a single dashboard organized around upcoming exams, with quizzes nested under the exam they prep for. Next round of testing finished in half the time.

The lesson I keep coming back to: AI features tempt you to expose the machinery. Resist it. Students don't care about the model pipelines. They care about the test on Friday.

The shipped flow

Homework done in Brainly → tap "Add to Test Prep" → pick exam date → AI generates study plan + practice questions → daily nudges scheduled around the student's school calendar → performance feeds the next session (wrong answers come back harder, mastered topics space out).






Reflections

Empathy is structural, not decorative. When your user is anxious, every design decision either adds to that load or removes it. There's no neutral.

Restraint beats power. The temptation with AI is to show off everything the model can do. The better move is to let the model recede so the student feels competent.

For v2: tighter feedback loop on flagged questions (retraining lagged behind flag volume), and a proper parent surface — they're a real influence on study behavior and we skipped them in v1.

UI / UX Design

Brainly AI-Powered Test Prep

Executed a data-driven design strategy by combining qualitative interviews, quantitative surveys, and competitive audits to build a zero-input, adaptive testing engine that reduced prep time by 40%.

Year :

2025

Industry :

EdTech

Employer :

Brainly

Project Duration :

8 weeks

Overview

Students shouldn't have to study how to study.

I led design end-to-end on Brainly Test Prep, an AI feature that turns the homework students already do into a personalized, adaptive study plan. No flashcards to build. No "where do I even start" panic the night before an exam.

First 60 days post-launch:

  • +30% weekly engagement vs. control

  • −40% self-reported study setup time

  • 4.5 / 5 average CSAT

(Brainly serves 15M+ MAU, so small UX wins compound fast.)


The problem

Brainly's own research told the story: 28% of US students name exams as their #1 source of school anxiety, and 68% admit they don't give themselves enough time to study. The gap wasn't motivation. It was knowing where to start.

Tools like Quizlet and StudySmarter only made it worse. They assume students will sit down, organize materials, and build their own study set. Cognitive load on top of cognitive load.

Our hypothesis: if we removed the setup work entirely and met students where their materials already lived (Brainly already had their homework and Q&A history), we'd turn studying from a planning problem into a doing problem.

Goals & KPIs

Three things we committed to before any pixels got pushed:

  1. Engagement: +20% weekly returns from Test Prep users vs. non-users

  2. Time-to-first-quiz: under 60 seconds from upload

  3. Satisfaction: ≥ 4.0 CSAT in the first month


Research

20 interviews (ages 13–19) and 500+ surveys. The qual surfaced an emotional layer I hadn't expected — students were carrying expectations from parents, peers, and teachers, and the planning of study time was where it all collapsed. Multiple kids said some version of "I open my notebook and just stare at it."

The quant confirmed it: 80% preferred automation over building their own flashcards, 60% felt overwhelmed during exam season, and 55% specifically asked for daily reminders.

Students didn't want more features. They wanted something to tell them what to do next.


Competitive landscape

A full audit showed a clear white space: nobody had nailed zero-input ingestion plus mistake-driven adaptive review.


Capability

Brainly

Quizlet

StudySmarter

Khan

Zero-input homework ingestion

AI-generated quizzes

🔸

Adapts to student mistakes

🔸

Study plan tied to exam date

🔸

Designing for AI uncertainty

This is where AI product design gets interesting. Model outputs aren't deterministic, and a 7th grader can't spot a hallucinated question. Two principles shaped every screen:

Make AI confidence legible. Every generated question shows topic and difficulty up front, so students know what's being tested.

Always give an out. Every quiz has a "this seems off" flag. Hides the question for that student, feeds back into retraining. Two birds.


The pivot

V1 had separate spaces for homework, quizzes, study plans, and past reviews. Five participants into a 15-person usability test, I had my answer. Students were treating the quiz library and the study plan as different products. One 9th grader said "wait, this is the same thing?" and she was right.

Killed it. Consolidated into a single dashboard organized around upcoming exams, with quizzes nested under the exam they prep for. Next round of testing finished in half the time.

The lesson I keep coming back to: AI features tempt you to expose the machinery. Resist it. Students don't care about the model pipelines. They care about the test on Friday.

The shipped flow

Homework done in Brainly → tap "Add to Test Prep" → pick exam date → AI generates study plan + practice questions → daily nudges scheduled around the student's school calendar → performance feeds the next session (wrong answers come back harder, mastered topics space out).






Reflections

Empathy is structural, not decorative. When your user is anxious, every design decision either adds to that load or removes it. There's no neutral.

Restraint beats power. The temptation with AI is to show off everything the model can do. The better move is to let the model recede so the student feels competent.

For v2: tighter feedback loop on flagged questions (retraining lagged behind flag volume), and a proper parent surface — they're a real influence on study behavior and we skipped them in v1.