This research represents the first comprehensive examination of whether artificial intelligence can design assessments capable of passing New Zealand's national moderation system. Conducted in partner...
📰 Content
Research
The "AI Adoption Playbook for UK Awarding Organisations" offers invaluable guidance for responsibly integrating AI into educational assessments. Featuring 37 practical strategies and expert advice, it...
📰 Content
Content
The research article explores ChatGPT's performance on multiple-choice question (MCQ) examinations in higher education. It finds that GPT-3.5 versions perform better than random guessing but often fai...
📰 Content
Report
BMC Medical Education's study compared AI-generated and clinician-designed multiple-choice questions (MCQs) for emergency medicine exams. AI-generated questions, while generally easier and associated ...
📰 Content
Report
The study evaluates ChatGPT-4o's effectiveness in generating medical examination multiple-choice questions (MCQs) compared to humans. AI-generated questions were quicker to produce and had comparable ...
📰 Content
Report
Sergio Araneda from Caveon will present research on "item pre-exposure" - a new test security risk from using Generative AI for item construction. His experiments show 40% overlap in AI-generated ques...
📰 Content
Content
Silk Data explores AI-powered question generation for businesses, covering automated question creation from text, audio, and video inputs. The article discusses applications in e-learning, HR recruitm...
📰 Content
Content
Research paper presenting an AI-powered system that automates question paper generation for educational institutions. The system uses Gemini-1.5-Pro and rule-based algorithms to create balanced, sylla...
📰 Content
Research
This research presents a human-in-the-loop framework for automatic item generation using AI to create multiple exam variants whilst maintaining psychometric rigor. The study demonstrates how educators...
📰 Content
Research
UNESCO MGIEP explores how artificial intelligence can transform educational assessment practices, addressing challenges in measuring complex cognitive skills needed for solving wicked problems like cl...
📰 Content
Content
Comprehensive analysis of how AI is transforming UK awarding organisations, assessment practices, and skills development. Examines regulatory frameworks, quality assurance innovations, learner engagem...
📰 Content
Content
Taylor Educational Consulting explores ethical considerations for using generative AI in qualification design and assessment development. The article covers legal compliance, copyright issues, GDPR re...
📰 Content
Content
A new EY-Parthenon–FICCI report reveals that over 60% of Indian higher education institutions now allow students to use AI tools, with 53% using generative AI for learning materials. The study of 30 l...
📰 Content
News
California enacted three new pieces of legislation to strengthen oversight of the state's bar exam following technical problems with the February 2025 attorney licensing test. The laws require advance...
📰 Content
News
The Standards and Testing Agency is piloting AI-generated questions in SATs moderator standardisation tests to reduce costs and school workload. The trial explores whether large language models can cr...
📰 Content
News
Tim Burnett discusses AI adoption strategies for UK awarding organisations, covering deployment options from consumer chat UIs to on-premise solutions. The article explores API routes, enterprise plat...
📰 Content
Content
Academic research examining bias and fairness in automated scoring systems used in educational testing. The paper surveys predictive methods that can lead to biased results, provides definitions of fa...
📰 Content
Research
Research paper presenting a multi-task generalized linear model with BERT features to estimate test item difficulties for adaptive language assessments. The method rapidly improves difficulty estimate...
📰 Content
Research
Generative AI is revolutionising assessment item generation, offering 10x to 100x production gains while maintaining psychometric quality comparable to human-authored items. However, success requires ...
📰 Content
Research
Comprehensive research review examining AI's role in generating assessment items across academic, professional, and psychometric domains. Shows AI can produce high-quality content comparable to human-...
📰 Content
Research
TTS Talent examines the scientific and ethical risks of using AI-generated psychometric assessments, comparing them to rigorous test development standards. The article highlights concerns about lack o...
📰 Content
Content
ACRP and The Academy of Clinical Research Professionals are conducting a study to evaluate AI-assisted item writing for certification exams. The research compares AI-generated questions with human-wri...
📰 Content
Research
This study evaluated 80 multiple-choice questions created by ChatGPT-4 for undergraduate psychology education. Results showed AI-generated items had reasonable content validity but limitations in asse...
📰 Content
Research
Pearson VUE explores how generative AI can assist in developing test items for driving theory tests, presenting research findings on AI-generated content quality compared to human-written items, while...
📰 Content
Content
A comprehensive study evaluating AI-generated exam questions across 91 college classes with nearly 1,700 students. Researchers developed an iterative refinement strategy using large language models to...
📰 Content
Research
Research study examining ChatGPT's ability to create physics concept inventory items. After careful prompt engineering and expert evaluation, ChatGPT-generated items showed medium difficulty and discr...
📰 Content
Research
Research paper exploring deep learning approaches for automated test item generation using recurrent neural networks, presenting an alternative to traditional human-written assessment items by impleme...
📰 Content
Research
This comprehensive review examines 60 studies on using large language models (LLMs) like T5, BERT, and GPT for automatic item generation in educational assessment. The research reveals that whilst LLM...
📰 Content
Research
This comprehensive review examines 60 studies on using large language models (LLMs) like T5, BERT, and GPT for automatic item generation in educational assessment. The research reveals that while LLMs...
📰 Content
Research
Research paper introducing STAIR-AIG, a human-in-the-loop framework that integrates expert judgment to optimize AI-generated assessment items for critical thinking. The study compares evaluations by h...
📰 Content
Research
Research paper exploring automated generation of distractors for math multiple-choice questions using large language models. The study compares various LLM approaches including in-context learning, fi...
📰 Content
Research
Research exploring GPT-4's potential to automate generation of personality situational judgment tests in Chinese. Two studies found that optimised prompts with temperature 1.0 produced creative, accur...
📰 Content
Research