Ministry Of AIMinistry Of AI
  • Home
  • Courses
  • About
  • Blog
  • Login
  • Register
Back
  • Home
  • Courses
  • About
  • Blog
  • Login
  • Register
  • Home
  • Blog
  • Blog
  • Detecting AI Is Broken—So Let’s Build Smarter Assignments Instead

Blog

03 Apr

Detecting AI Is Broken—So Let’s Build Smarter Assignments Instead

  • By Stephen Smith
  • In Blog
  • 0 comment

Detecting AI Is Broken—So Let’s Build Smarter Assignments Instead

The explosion of generative AI tools like ChatGPT is shaking up classrooms around the world—and not always in a good way. While students now have AI at their fingertips to draft essays, summarize notes, and brainstorm arguments, educators are left scratching their heads: how do you keep learning real when machines can ace your assignments?

The go-to response for many institutions has been to combat AI use with AI detectors. But here’s the uncomfortable truth: these tools don’t really work. They misclassify human work as AI and vice versa. Worse, with a little paraphrasing or translation, almost anyone can easily sneak past them.

So what’s the fix? According to a new study by Muhammad Sajjad Akbar, we’ve been asking the wrong question. Instead of obsessing over detection, we should be focusing on design. Specifically, designing AI-resilient assessments that actually challenge students to think critically, solve problems creatively, and produce work that generative AI can’t replicate.

Let’s break down the key ideas from this groundbreaking research—and what it means for students, educators, and the future of learning.


Why AI and Education Are at a Crossroads

Generative AI isn’t just a tech trend—it’s mashed its way into education like a wrecking ball. Tools like ChatGPT can spin up essays, explain concepts, and even mimic student voices in a matter of seconds. For students, that can feel like an academic shortcut. For professors, it’s a potential academic integrity time bomb.

But beyond questions of cheating, there’s something more serious at stake: Are we trading away deeper learning for convenience?

Educators are noticing that when students offload too much thinking to AI, they’re missing out on developing key cognitive skills like problem solving, analysis, and creativity. In modern education, especially at the postgraduate level, these higher-order thinking skills are essential. They help graduates adapt, think independently, and contribute meaningfully in any industry.


Current AI Detection Tools: Not the Silver Bullet

Most colleges and universities have turned to AI detectors in a desperate attempt to catch machine-written content. But this study found that most of these tools are, frankly, not very good:

  • Poor accuracy: Even popular commercial tools like Turnitin and GPTZero often misidentify both AI and human-written text. Some had false positive rates as high as 50%.

  • Easy to trick: Students can confuse detectors just by paraphrasing, translating, or modifying AI-generated content.

  • Lack of transparency: Many tools give vague reports (“This text is 70% likely written by AI”) without solid evidence.

The result? Educators may wrongly accuse students or let AI-authored content slip through—not because of laziness or malice, but because the tools they’re using simply aren’t up to the job.


Introducing a Better Approach: AI-Resilient Assignment Design

Instead of trying to catch AI after it’s been used, the researchers suggest beating AI at its own game—by creating assessments that AI tools struggle to complete effectively.

To do this, they built a smart, web-based tool that helps educators analyze how “AI-solvable” their assignments are. It measures whether a task falls into the easy-for-AI category (think: summarizing, defining terms) or if it requires complex thinking like analyzing data, evaluating arguments, or creating something original.

This system works by combining:

  • Bloom’s Taxonomy – a classic model that categorizes learning objectives from basic to advanced.

  • Natural Language Processing (NLP) – using GPT-3.5 Turbo, BERT semantic analysis, and TF-IDF to gauge how complex and nuanced the assignment questions really are.

  • Python automation – to quickly scan PDFs of assignments and return an “AI-solvability score” with real-time feedback.

The lower the score, the harder it is for AI to generate a satisfactory answer. And that’s exactly what we want!


What the Research Found: Most Assignments Are AI-Too-Friendly

When the team tested 50 real-world assignments from various computer science courses, the results were sobering:

  • 22 assignments had AI-solvability scores between 65–74% — meaning AI could probably handle them with ease.

  • Only 4 assignments scored below 50%, indicating true resistance to AI-generated responses.

These high-solvability assignments were mostly definitions, factual explanations, or basic applications—the kinds of things ChatGPT gobbles up. On the other hand, the low-solvability ones required designing unique solutions, interpreting specific scenarios, and thinking critically, creatively, and contextually.

Put simply: the more thinking an assignment requires, the harder it is for AI to fake it.


How Students Use AI: Help or Hindrance?

The team also surveyed postgraduate students at the University of Sydney to understand how AI is actually showing up in their learning. The results paint a mixed picture:

  • 70% of students said AI tools helped them understand material better. (Nice!)

  • But over 60% admitted to making only minimal edits to the AI text before submitting. (Uh-oh.)

  • 75% believed the AI answers were more accurate than theirs.

  • 60% said AI tools actually made them less creative.

This suggests a potential over-reliance on AI—not just for support, but as a crutch. While the tools can boost confidence and clarify tough concepts, they can also kill creativity and reduce critical engagement when used passively.


The Real Impact: Smarter Assessment Design Can Fix This

By using the AI-solvability tool before finalizing an assignment design, educators can ask themselves key questions:

  • Does this task really challenge students to think independently?
  • Can ChatGPT realistically generate a viable response here?
  • Am I valuing originality and synthesis over regurgitation?

Armed with this insight, instructors can rethink their prompts—pulling back on the easy-to-automate stuff and leaning into unique, personalized, and cognitively demanding formats. Think open-ended projects, scenario-based analysis, or unfamiliar case studies—things AI can’t fake.

Even better, the tool gives clear feedback so assignments can be revised before they’re given out. That means assessments not only become fairer, but more effective at encouraging actual learning.


Real-World Applications: Where This Can Make a Difference

Here’s how this AI-aware assignment design approach can be practically used:

  • Instructors can automatically check if their assignment is too easy for ChatGPT and tweak it on the spot.

  • Education departments can integrate the tool into Learning Management Systems to flag at-risk assessments early.

  • Faculties can train professors to apply Bloom’s Taxonomy more deliberately when crafting course tasks.

  • Students benefit from assignments that reward insight and creativity rather than formulaic responses, making room for real skill-building.


Key Takeaways

✅ AI detectors are unreliable. They produce too many false positives/negatives and are easy to circumvent.

✅ Stop trying to catch AI—start designing around it. Move from detection to prevention by crafting better, higher-order assessments.

✅ Use Bloom’s Taxonomy as your secret weapon. The more your task pushes into analysis, evaluation, and creation, the harder it is for AI to fake it.

✅ Students should be taught AI literacy. Knowing how to use AI tools responsibly—as support, not a substitute—builds better thinkers.

✅ Proactive assessment design helps everyone. Educators reduce cheating, and students engage more deeply with their learning.

✅ This is scalable and ready now. A Python-powered tool exists today that gives measurable, actionable feedback on how AI-susceptible your assignments really are.


In a world where machines can crank out essays in seconds, it’s tempting to react with suspicion, lockdowns, or blame. But this research shows us a smarter path forward: design assessments that AI can’t do—but students must.

Let’s raise the bar not just for what AI can do, but for what students need to learn.


Want to rethink your assignments or understand how AI-solvable your tasks are? Keep this question handy next time you write a course prompt:

“Could ChatGPT answer this well without understanding the course or the context?”

If the answer is “yeah, probably”… it might be time for a redesign.


Author Note: Based on research by Muhammad Sajjad Akbar, Beyond Detection: Designing AI-Resilient Assessments with Automated Feedback Tool to Foster Critical Thinking.


If you are looking to improve your prompting skills and haven’t already, check out our free Advanced Prompt Engineering course.

This blog post is based on the research article “Beyond Detection: Designing AI-Resilient Assessments with Automated Feedback Tool to Foster Critical Thinking” by Authors: Muhammad Sajjad Akbar. You can find the original article here.

  • Share:
Stephen Smith
Stephen is an AI fanatic, entrepreneur, and educator, with a diverse background spanning recruitment, financial services, data analysis, and holistic digital marketing. His fervent interest in artificial intelligence fuels his ability to transform complex data into actionable insights, positioning him at the forefront of AI-driven innovation. Stephen’s recent journey has been marked by a relentless pursuit of knowledge in the ever-evolving field of AI. This dedication allows him to stay ahead of industry trends and technological advancements, creating a unique blend of analytical acumen and innovative thinking which is embedded within all of his meticulously designed AI courses. He is the creator of The Prompt Index and a highly successful newsletter with a 10,000-strong subscriber base, including staff from major tech firms like Google and Facebook. Stephen’s contributions continue to make a significant impact on the AI community.

You may also like

Unlocking the Future of Learning: How Generative AI is Revolutionizing Formative Assessment

  • 30 May 2025
  • by Stephen Smith
  • in Blog
Unlocking the Future of Learning: How Generative AI is Revolutionizing Formative Assessment In the evolving landscape of education, the...
Navigating the Coding Classroom: How Peer Assessment Thrives in the Age of AI Helpers
30 May 2025
Redefining Creative Labor: How Generative AI is Shaping the Future of Work
29 May 2025
Guarding AI: How InjectLab is Reshaping Cybersecurity for Language Models
29 May 2025

Leave A Reply Cancel reply

You must be logged in to post a comment.

Categories

  • Blog

Recent Posts

Unlocking the Future of Learning: How Generative AI is Revolutionizing Formative Assessment
30May,2025
Navigating the Coding Classroom: How Peer Assessment Thrives in the Age of AI Helpers
30May,2025
Redefining Creative Labor: How Generative AI is Shaping the Future of Work
29May,2025

Ministry of AI

  • Contact Us
  • stephen@theministryofai.org
  • Frequently Asked Questions

AI Jobs

  • Search AI Jobs

Courses

  • All Courses
  • ChatGPT Courses
  • Generative AI Courses
  • Prompt Engineering Courses
  • Poe Courses
  • Midjourney Courses
  • Claude Courses
  • AI Audio Generation Courses
  • AI Tools Courses
  • AI In Business Courses
  • AI Blog Creation
  • Open Source Courses
  • Free AI Courses

Copyright 2024 The Ministry of AI. All rights reserved