Exactly. But not just "error prone", rather: eloquently and confidently incorrect. And that in itself is a problem. When I evaluate students' work, I implicitly do so from a mental model of the student - aptitude, ability, experience, language skills etc. That's useful for summative assessment, since it helps efficiency - but that won't work anymore. I see a need to assess much more carefully, require fine-grained referencing, check every single fact ... and that won't scale. And then there is also the spectre of having to decide when this crosses the line to "concoction" - i.e. an actual academic offence ...
Best, Boris > On 2022-12-19, at 03:58, Milan Glacier <n...@milanglacier.com> wrote: > > [You don't often get email from n...@milanglacier.com. Learn why this is > important at https://aka.ms/LearnAboutSenderIdentification ] > > On 12/18/22 19:01, Boris Steipe wrote: >> Technically not a help question. But crucial to be aware of, especially for >> those of us in academia, or otherwise teaching R. I am not aware of a >> suitable alternate forum. If this does not interest you, please simply >> ignore - I already know that this may be somewhat OT. >> >> Thanks. >> ------------------------------------------------------ >> >> You very likely have heard of ChatGPT, the conversation interface on top of >> the GPT-3 large language model and that it can generate code. I thought it >> doesn't do R - I was wrong. Here is a little experiment: >> Note that the strategy is quite different (e.g using %in%, not duplicated() >> ), the interpretation of "last variable" is technically correct but not what >> I had in mind (ChatGPT got that right though). >> >> >> Changing my prompts slightly resulted it going for a dplyr solution instead, >> complete with %>% idioms etc ... again, syntactically correct but not giving >> me the fully correct results. >> >> ------------------------------------------------------ >> >> Bottom line: The AI's ability to translate natural language instructions >> into code is astounding. Errors the AI makes are subtle and probably not >> easy to fix if you don't already know what you are doing. But the way that >> this can be "confidently incorrect" and plausible makes it nearly impossible >> to detect unless you actually run the code (you may have noticed that when >> you read the code). >> >> Will our students use it? Absolutely. >> >> Will they successfully cheat with it? That depends on the assignment. We >> probably need to _encourage_ them to use it rather than sanction - but >> require them to attribute the AI, document prompts, and identify their own, >> additional contributions. >> >> Will it help them learn? When you are aware of the issues, it may be quite >> useful. It may be especially useful to teach them to specify their code >> carefully and completely, and to ask questions in the right way. Test cases >> are crucial. >> >> How will it affect what we do as instructors? I don't know. Really. >> >> And the future? I am not pleased to extrapolate to a job market in which >> they compete with knowledge workers who work 24/7 without benefits, vacation >> pay, or even a salary. They'll need to rethink the value of their investment >> in an academic education. We'll need to rethink what we do to provide value >> above and beyond what AI's can do. (Nb. all of the arguments I hear about >> why humans will always be better etc. are easily debunked, but that's even >> more OT :-) >> >> -------------------------------------------------------- >> >> If you have thoughts to share how your institution is thinking about >> academic integrity in this situation, or creative ideas how to integrate >> this into teaching, I'd love to hear from you. > > *NEVER* let the AI misleading the students! ChatGPT gives you seemingly > sound but actually *wrong* code! > > ChatGPT never understands the formal abstraction behind the code, it > just understands the shallow text pattern (and the syntax rules) in the > code. And it often gives you the code that seemingly correct but indeed > wrongly output. If it is used with code completion, then it is okay > (just like github copilot), since the coder need to modify the code > after getting the completion. But if you want to use ChatGPT for > students to query information / writing code, it is error proning! ______________________________________________ R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.