Discussion about this post

User's avatar
Wendy Winnard's avatar

Really insightful for my research. Up to now there is little evidence that LLM can reliably grade structured response questions in A level chemistry exams. Human in the loop examiners could comparatively grade 1000s, but one of the bugbears as a teacher is being able to decipher handwriting. Has anyone got a tool that can do this?

Expand full comment
Mark Aveyard's avatar

Thank you, that was informative. But I don't understand why AI avoids some of the hallucination errors under CJ prompts/tasks? Is ChatGPT ordinarily doing some kind of CJ process on a massive scale or is does it behave differently within a CJ process? And if so, how does it "know" to behave differently?

Expand full comment
3 more comments...

No posts