AI Assessments: SMU ChatGPT Research | AcademicJobs SG

Singapore Management University Leads AI-Resistant Assessment Revolution

New0 comments

Be one of the first to share your thoughts!

Add your comments now!

Have your say

Engagement level

See more Higher Ed News Articles

a close up of a typewriter with a paper on it
Photo by Markus Winkler on Unsplash

Understanding the Rise of Generative AI in Higher Education Assessments

In the rapidly evolving landscape of higher education, generative artificial intelligence (GAI) tools like ChatGPT have transformed how students approach learning and assessments. Since its public release in late 2022, ChatGPT has demonstrated remarkable capabilities in generating human-like text, code, and even complex calculations, prompting educators worldwide to rethink traditional evaluation methods. In Singapore, a global leader in educational innovation, universities are at the forefront of integrating AI responsibly while safeguarding academic integrity.5253

Singapore Management University (SMU), one of the nation's top institutions, exemplifies this proactive stance. With its emphasis on experiential learning and industry relevance, SMU has developed comprehensive frameworks for GAI use, allowing students to leverage tools like ChatGPT for idea generation and research under strict guidelines. This approach aligns with national efforts, where the Ministry of Education (MOE) and institutions like National University of Singapore (NUS) and Nanyang Technological University (NTU) permit AI in assignments provided proper citation and transparency.63

However, the challenge lies in assessments. Traditional exams and assignments, once reliable indicators of student mastery, are now vulnerable to AI assistance, raising concerns about critical thinking and originality. A 2025 Higher Education Policy Institute (HEPI) study revealed that 88% of students use GAI for assignments, underscoring the urgency for redesign.64

Spotlight on SMU's Groundbreaking Research

Leading the charge is Dr. Michelle L. F. Cheong from SMU's School of Computing & Information Systems. Her seminal paper, "ChatGPT's Performance Evaluation in Spreadsheets Modelling to Inform Assessments Redesign," published in the Journal of Computer Assisted Learning on May 5, 2025 (DOI: 10.1111/jcal.70035), provides empirical evidence on ChatGPT's limitations.62 This open-access study fills a critical gap by testing ChatGPT version 3.5 on real-world spreadsheet modeling tasks—skills essential in business and data analytics courses.

Cheong's work is particularly relevant for Singapore's higher education sector, where quantitative skills are prized. SMU, known for its business-focused programs, integrates spreadsheets extensively, making this research directly applicable. The study not only benchmarks AI performance but offers actionable redesign strategies, positioning SMU as a thought leader.41

Dr. Cheong emphasizes, "Educators need new information on how well ChatGPT performs to redesign future assessments in this new paradigm." Her findings reveal that while AI excels in routine tasks, it falters in higher-order thinking, empowering instructors to target those areas.64

Methodology: Rigorous Testing with Bloom's Taxonomy

Cheong categorized assessment questions using the revised Bloom's Taxonomy, a framework outlining six cognitive levels: (1) Remembering, (2) Understanding, (3) Applying, (4) Analyzing, (5) Evaluating, and (6) Creating. Two original quizzes were developed:

  • Financial calculations quiz: Scenarios like computing discounts and payments for buying computers.
  • Monte Carlo simulation quiz: Estimating COVID-19 infection risks in queues using probabilistic modeling.

These multi-step, linked items mimic authentic coursework. ChatGPT was prompted in four engineering variations:

  • Zero-Shot-Baseline: Direct question, no guidance.
  • Zero-Shot-Chain-of-Thought (CoT): Adds "Let's think step by step."
  • One-Shot: Includes one solved example.
  • One-Shot-CoT: Example plus step-by-step reasoning.

Performance was scored objectively on numerical accuracy and logical correctness, repeated for consistency.6264

Bloom's Taxonomy levels illustrating ChatGPT performance decline in higher cognitive tasks

Key Findings: Where ChatGPT Succeeds and Fails

The results were telling. With baseline prompts, ChatGPT achieved high accuracy up to Level 3 (Applying), correctly handling basic formulas and data setup. Errors emerged at Level 4 (Analyzing), often miscalculating despite identifying the right method—e.g., wrong probability distributions in simulations.

Prompt enhancements extended capabilities: Zero-Shot-CoT reached Level 5 (Evaluating) reliably, while One-Shot variants boosted mid-levels. However, no configuration mastered Level 6 (Creating), where novel model design is required. Overall, accuracy dropped sharply with complexity, highlighting AI's brittleness in advanced reasoning.64

Bloom's LevelBaseline AccuracyBest Prompt AccuracyExample Error
1-3 (Low-Mid)HighHighRare
4 (Analyzing)LowMediumIncorrect linkages
5-6 (High)Very LowLowFailed innovation

These insights validate anecdotal educator concerns and provide data-driven baselines.

Redesign Strategies Tailored to Cognitive Levels

Cheong proposes level-specific redesigns to mitigate GAI while fostering skills:

  • Levels 1-2: In-class ChatGPT sessions to generate answers, then critique errors. Students learn prompt crafting and verification.
  • Levels 3-4: Collaborative peer projects, e.g., building EV adoption spreadsheets, emphasizing analysis where AI stumbles.
  • Levels 5-6: AI-prohibited tasks focusing on creation/evaluation, like designing original simulations from real data.

SMU's Centre for Teaching Excellence (CTE) echoes this with AI-resistant practices: process-tracked drafts, personal reflections, in-class viva voce, and authentic tasks like mock client consultations.SMU CTE Guidelines63

For spreadsheets specifically, require code explanations, error debugging, or integration with class-specific datasets.

Singapore's Broader Higher Education Landscape

Singapore universities report few AI misconduct cases—less than a handful at SMU over three years—thanks to transparent policies.52 NUS's interim AI policy mandates disclosure, while NTU and SMU use Turnitin AI detectors judiciously, recognizing false positives. Educators like those at SMU advocate shifting from detection to design: "Trying to catch AI use is a lost cause," notes a computing lecturer.53

An OECD survey highlights Singapore teachers' high AI adoption—3 in 4 use it, double the global average—driving innovative pedagogies.28 This positions the nation to lead in AI-augmented education.

Explore academic career advice for roles in AI-enhanced teaching.

Stakeholder Perspectives: Educators, Students, and Industry

Educators praise Cheong's work for empowering redesign without bans. SMU students, surveyed informally, appreciate guided AI use for brainstorming but stress human oversight for depth. Industry partners value graduates skilled in AI collaboration, aligning with Singapore's Smart Nation initiative.

Challenges include equity—prompt engineering favors advanced users—and detection tool reliability. Solutions: Faculty training via SMU CTE workshops and student AI literacy modules.

SMU students collaborating on AI-integrated spreadsheet projects

Real-World Case Studies and Statistics

In Cheong's financial quiz, ChatGPT nailed discount calcs (Level 2) but bungled multi-scenario analysis (Level 4). The Monte Carlo task exposed flaws in risk modeling, mirroring real analytics pitfalls.

Stats: 92% of students use AI per HEPI; Singapore's low cheating (ST report) vs. global rises. Cheong's prior study on permitted ChatGPT assessments showed improved learning when integrated thoughtfully.39

  • AI boosts productivity but risks rote reliance.
  • Redesigned assessments raise critical thinking by 20-30% in pilots.

Read the full paper for datasets.62

Future Outlook: AI as Ally in Singapore Higher Ed

Looking ahead, Singapore's Research, Innovation and Enterprise 2030 (RIE2030) plan invests S$37 billion in AI, including education. SMU's research paves the way for hybrid models: AI for basics, humans for innovation.

Implications: Enhanced employability via AI fluency; global benchmarks for unis. Challenges like evolving AI (GPT-4o) require ongoing research.

Discover Singapore higher ed jobs or global opportunities.

an apple sitting on top of a stack of books

Photo by - Landsmann - on Unsplash

Actionable Insights for Educators and Institutions

To implement:

  • Audit assessments via Bloom's mapping.
  • Pilot Cheong's prompts for baselines.
  • Integrate AI literacy in curricula.
  • Use tools like Peerceptiv for feedback.

Rate professors on Rate My Professor for AI-savvy insights. For career growth, check higher ed career advice.

Discussion

0 comments from the academic community

Sort by:
You

Please keep comments respectful and on-topic.

Frequently Asked Questions

🤖What does SMU research say about ChatGPT's performance in assessments?

Dr. Cheong's study shows ChatGPT excels in lower Bloom's levels (1-3) but fails at higher ones (4-6), especially creation tasks. Read the paper.

📊How can educators redesign assessments based on this research?

Use in-class AI critique for basics, peer projects for mid-levels, and AI-free creation for advanced. Align with career advice.

📚What is Bloom's Taxonomy in AI assessment context?

Revised framework: Remembering to Creating. ChatGPT handles recall/application but not analysis/evaluation.

🇸🇬Are Singapore universities banning ChatGPT?

No, SMU/NUS/NTU allow it with disclosure. Focus on redesign, few misconduct cases.

💡What prompts improve ChatGPT in spreadsheets?

Chain-of-Thought ('step by step') and One-Shot examples extend to Level 5.

⚖️How does SMU handle AI academic integrity?

Clear guidelines, process-focused evals, detectors like Turnitin. See CTE resources.

📈Stats on student AI use in Singapore?

88-92% use GAI; 75% teachers adopt AI (OECD). Low cheating reported.

🛡️Examples of AI-resistant tasks?

Personal reflections, in-class vivas, novel model creation, debugging AI outputs.

🔮Future of AI in Singapore higher ed?

RIE2030 invests S$37B; hybrid models for AI fluency. Check jobs.

How to rate AI-savvy professors?

Use Rate My Professor for insights on innovative teaching.

🔬Other SMU AI research?

Cheong's work on permitted ChatGPT assessments shows learning gains.