PowerMark · Aircury × Smartgrade (Powermark)

PowerMark

Executive summary of the case study.

Raúl Álvarez Hinojosa
2 min read

Within Smartgrade, Powermark is their AI auto-marking product: per Smartgrade’s UK marketing site, it targets a sharp drop in marking time (customers report 65–90 % savings vs traditional marking), real-time marking for HeadStart Primary Reading and GPS or custom online papers, typically 94–97 % accuracy, no personally identifiable information sent to AI models, aggregated class and cohort feedback, and teacher moderation and override (low-confidence answers surfaced first). At Aircury I designed and oversaw the PowerMark stack (OCR + NLP) integrated with Smartgrade for handwritten responses and tests—inference, data quality, and rollout with pedagogical teams.

OCRNLPAI-assisted markingTeacher moderationPrivacy & complianceEdTech integrationObservability

Screenshots — Powermark (Smartgrade UK)

Smartgrade UK Powermark hero: headline on marking time and four value cards (accuracy, workload, secure AI, personalised feedback).
Value proposition — marking in moments, real-time accuracy (94–97% as stated on-page), time savings (65–90% per customer reports), no PII to AI models, class and cohort reporting.
Smartgrade UK “How does Powermark work?” with four steps: marketplace assessment, online submission, AI with low-confidence flags, aggregated teacher reports.
Flow — AI-enabled assessment from the Marketplace, online attempt, auto-marking with doubtful answers surfaced for review, aggregated and misconception-focused reports.

A long-form case study (our own metrics, model evaluation, and classroom operations) is in preparation. Partnerships or a walkthrough: hola@raul-alvarez.es.

FAQ#

Are Powermark and PowerMark the same thing?

Powermark is Smartgrade’s product name for AI auto-marking. PowerMark is Aircury’s OCR+NLP work integrated with that platform.

What accuracy does Smartgrade publish?

They aim for 94%+ where marking guidance is detailed, and cite a large HeadStart Reading and GPS sample (~9,500 responses) at 97% AI accuracy vs 94% for teachers. Accuracy can fall into the 80–90% range when questions or guidance are weaker.

How do they mitigate common AI risks?

They describe per-answer confidence scoring, priority teacher review, random sampling with expert checks, and tight mark schemes; they state student data is not used for model training and PII is not passed to external LLMs.

What was your role on PowerMark?

I owned inference pipeline design and oversight, data quality, and safe rollout with pedagogical stakeholders; the full article will spell out scope and limits.

What about paper assessments?

Smartgrade UK positions scanned paper marking as the next Powermark phase and invites schools/MATs to join a pilot via their sales channel.

Raúl Álvarez Hinojosa

Raúl Álvarez Hinojosa

AI Architect · AI Solutions Engineer · Head of AI @ Aircury

Profile, projects, and technical case studies on raul-alvarez.es — Granada, EU remote.

More about the author →
© 2026 Raúl Álvarez Hinojosa. All rights reserved.|Privacy