Teach Edge
HomeHow It WorksPricingTrialFAQsContactBlog
Ai In Education

What students actually think about AI marking and feedback

King's School, Chester ran an independent student voice survey on TeachEdge after around 18 months of use. Here's what 148 students said about clarity, speed, progress, and trust in marking.

19 February 2026•3 min read•
Ai In EducationStudent VoiceFeedbackAssessmentTeacher WorkloadEconomicsBusiness Studies

Quick Summary

  • •Students rated clarity of feedback highest (91.2% agree/strongly agree).
  • •Timely feedback scored strongly (84.4%), which matters for acting while questions are still fresh.
  • •Trust in marking accuracy was the toughest item, but still positive (78.4%).
  • •These results come after sustained use (around 18 months), not a one-off novelty demo.
  • •Independent surveys like this move the AI-in-schools conversation beyond hype and fear.

King's School, Chester recently shared something I wish more schools did.

Stephen Walton (Head of Business and Economics) ran a student voice survey asking students what they actually think about AI marking and feedback after using TeachEdge for around 18 months.

A key point up front: TeachEdge had no involvement in designing, administering, or collecting this survey. Stephen and the school ran it independently and shared the results with us afterwards.

There were 148 student responses across GCSE and A Level Business and Economics.

The headline results

Across six statements, the proportion who agreed or strongly agreed was:

  • 91.2% – "AI feedback is clear and understandable"
  • 87.1% – "Overall I am satisfied by the AI feedback I have received"
  • 84.5% – "AI feedback has helped me understand how to improve my answers"
  • 84.4% – "AI feedback is timely"
  • 83.1% – "AI feedback is helping me to make progress"
  • 78.4% – "I trust the accuracy of the AI marking"

Bar chart showing the student voice survey results: percent who agree or strongly agree across six statements about TeachEdge AI marking and feedback Figure 1: Student voice survey results from King's School, Chester (n=148): the percentage of students who agreed or strongly agreed with each statement.

The bit I think matters most

Three things stand out.

1) Clarity is the biggest win

The strongest result is that students find the feedback clear and understandable.

That matters because "feedback" only helps if a student can actually use it under real exam pressure. Clarity beats quantity.

2) Speed is not just convenient, it changes behaviour

Timeliness came through strongly too.

Students often receive feedback when the class has already moved on. Fast feedback makes it more likely they will act on it while the question is still fresh.

3) Trust in accuracy is the hardest question, and it still landed well

The lowest score (though still strong) was trust in marking accuracy.

That makes sense. Students are rightly sceptical about any marking system, human or otherwise. If anything, the fact this was not 95% makes the survey feel more believable. Trust is earned slowly, especially with assessment.

Why the "18 months" detail matters

This was not a novelty effect.

These students were not reacting to a one-off demo lesson. They have had AI-supported feedback in their routine practice for well over a year. That is what makes this genuinely useful.

What we take from this (and what we'll keep improving)

For me, the message from students is pretty clear:

  • Make feedback easy to understand
  • Get it back quickly
  • Be transparent about marking decisions

On our side, the work is ongoing, especially around trust and accuracy. We're not claiming perfection. What we are building is a workflow where AI does the heavy lifting, but teachers stay in control.

That includes:

  • exam-board aligned prompts
  • assessment objective breakdowns where relevant
  • teacher review and editing before anything is released to students

Thank you

Finally, a genuine thank you to Stephen Walton for running the survey and for being happy to share the results publicly. It's generous, and it helps move the conversation about AI in schools onto something more useful than hype or fear.

Related Posts

Straight from the Source: What GCSE Business Students REALLY Think About AI Essay Feedback (and why it matters)

A GCSE Business teacher asked students for honest, unfiltered views on TeachEdge's AI feedback. The themes were clear: speed, detail, confidence — plus a few sharp suggestions on how we can improve.

Gcse BusinessStudent VoiceAi In Education

Navigating AI Essay Marking and Feedback

AI can save hours on essay marking, but the real impact comes when a teacher reviews, tweaks, and stands behind the feedback. This post explains the 'human-in-the-loop' approach that makes AI feedback feel trustworthy and genuinely useful to students.

Ai In EducationEssay MarkingFeedback

Stop Trying to Catch Them: Why AI Detection is a Dead End for UK Secondary Schools

AI detection tools cannot reliably prove whether a GCSE or A Level student used generative AI. Schools will get further by modelling good AI use and protecting supervised writing time.

Ai In EducationAssessmentAcademic Integrity

Ready to transform your marking workflow?

View Pricing•Contact Us

Product

  • How it works
  • Pricing
  • Compare TeachEdge to other AI Marking Tools

Resources

  • Blog
  • FAQs
  • Copyright FAQs

Company

  • Contact

Legal

  • Privacy policy
  • Data security
  • Child-Friendly Privacy Policy
  • Terms

TeachEdge is a web-based education application used by teachers in UK secondary schools. Disclaimer: TeachEdge.ai is independent of, and not endorsed by, any examination board.

TeachEdge.ai is a UK-built platform that helps secondary teachers give better marking and feedback on essays and longer exam-style questions, without the copy-and-paste admin. As an AI teaching tool, it supports AI assessment and AI feedback in a way that keeps teachers firmly in control: teachers set a task for a class, students submit in their own portal (typed or handwritten, including diagrams), and Teach Edge produces accurate draft marks and feedback calibrated to the relevant exam board (Edexcel, OCR, AQA, CAIE, WJEC/Eduqas). Teachers review and adjust that feedback before anything is released. Nothing is released to students until the teacher approves it.

It currently supports GCSE and A Level practice across: Economics, Business, History, English Language, English Literature, Sociology, Politics, Geography, Law, Philosophy, Music, Media, Film Studies, Biology, Maths, Physics, Chemistry, French, Spanish, Criminology, Psychology and HSC. The aim is simple: reduce marking load while making feedback clearer, more consistent, and more useful for students to act on — practical AI for teachers that fits normal classroom routines.

Teach Edge also includes personalised tutoring. Teachers set the topic and students work through a one-to-one conversation that starts with a short baseline check and then proceeds in a Socratic, scaffolded way. Crucially, teachers can review full conversations and see summaries of student understanding or misconceptions, including class-level patterns, so tutoring feeds directly back into teaching — another way Teach Edge supports AI feedback that teachers can trust and act on.

© 2026 TeachEdge.ai. All rights reserved.