No use crying over...: a teaching strategy to tame ChatGPT into being your students’ and your best friend.

Streamed Session

Session Materials

Brief Abstract

Our objective is to provide high-level understanding of LLMs, such as ChatGPT, to make instructors comfortable enough to start using these tools with students. We help instructors identify the assignments that had already been disrupted, and show how to use systematic approaches to remake them for this new reality.
 

Presenters

Hello, I'm George Jura, and for the last decade I have been is the Director of Academic Technology at the U of Wisconsin-Madison School of Nursing (SoN). I lead a small (but fantastic!) team in helping faculty identify, explore, and implement technology-enhanced teaching and learning strategies and course designs that best support and complement the School's undergraduate and doctoral programs, and offer our students the best possible learning experience. The driving force behind my work is the realization that even the most sophisticated technology is useless if people who could benefit from it don't know how to use it effectively; consequently, my passion is making information about technology easier to find and understand, and making technology easier to use. My primary long-term interests include evidence-based instructional design and course development, the use of student-response systems to increase student engagement in large undergraduate courses, and helping faculty develop generative learning activities that promote deeper understanding. My immediate present focus is on leading the UW-Madison School of Nursing academic technology team in the planning and design of courses for two new graduate (Doctor of Nursing Practice) 3-year programs launched in the fall of 2021. I am also fascinated by the LLM interaction tools (better known as AI, such as ChatGPT, Google Bard, or Craft AI Assistant) and how they can be used in university classrooms and programs. In the past, I lead campus-wide technology training initiatives for students and university staff at the U of Wisconsin-Whitewater, as its first Technology Advancement and Training Coordinator, including managing the UW-W TechQuest project, an innovative online technology training for incoming students (which we launched in 2011, and shared at the 2012 Regional Educause conference - see: https://events.educause.edu/archives/2012/educause-midwest-regional-conference-2012/proceedings/technology-literacy-on-campus-uw8211whitewater-techquests-best-practices-and-lessons-learned , and it's still being used today!). Before focusing entirely on instructional technology, I worked on the tenure-track at Iowa State U (Ames, IA), and at Lawrence U (Appleton, WI), where I taught several technology-enhanced courses on Spanish Mediaeval and Renaissance Literature; Spanish Art; Film and Narrative Theory; and Linguistics. I have shared my ideas on teaching with technology with audiences at numerous forums, including presentations and workshops at several EDUCAUSE National, and Midwest-Regional Conferences; ELI Annual Meetings; National Forums on Active Learning Classrooms; Distance Learning Conferences (Madison, WI) incl. 2020 virtual conference; several OLC Accelerate and Innovate conferences, TeachX (Northwestern U, IL), and - as faculty - a dozen-or-so Modern Language Association conferences. My current/recent OLC presentations include: -- the (forthcoming) 2023 OLC Virtual Innovate session (Apr. 6 at 11:30) titled 'Meet ACSL' - see more at: https://musingsontandl.com/2022/11/03/meet-acsl-a-low-prep-high-yield-teaching-strategy-that-produces-results/ This session is a much revised and improved (read: streamlined, more realistic, and more easily implemented) continuation and expansion of my ideas originally explored in my 2020 blog post + 3 years of real-world, practical experience. See the original 2020 blog post at: https://musingsontandl.com/2020/05/25/an-unexamined-video/ -- and 2022 OLC Innovate session is titled 'Nudge, Sludge, and Choice Architecture: Applying the Lessons of Behavioral Economics to Instructional Design.' Session info at: https://musingsontandl.com/2022/04/12/nudge-sludge-and-choice-architecture-applying-the-lessons-of-behavioral-economics-to-instructional-design/

Extended Abstract

Key objectives:

  •  to provide high-level, basic understanding of how LLMs operate
  • to place LLMs in the context of culture to understand the magnitude and weight of this advance 
  • to focus attention on ethical, epistemological, and methodological issues that each potential solution must consider
  • to help instructors develop a strategy to identify the assignments in their existing courses had already been disrupted (and should not be re-used "as is")
  • to offer instructors a simple and flexible but radical toolkit (available under CC BY-NC-SA license) that will allow them to transform their existing disrupted assignments into formats that can coexist with, and even benefit from a structured LLMs use in the new post ChatGPT reality.

Interactivity / Participant Engagement (note: wile not required for most participation, participants will be encouraged to have an active (free) Chat GPT account):

  • several polls are planned during the presentation:
    • know-the-context: a survey of instructors and instructional support staff on their pre-workshop approaches to identifying disruption in existing assignments;
    • pre-view: asking the audience to guess the main types of AI hallucinations and then grouping their answers and discussing in reference to established / usual classification
    • applying the knowledge: classifying several different examples of hallucinations

---
Summary
The email landed in our academic technology team mailbox, and it had been written with an obvious sense of urgency and apprehension: "I feel really naive," confessed one of our stellar instructors, an experienced NP (nurse practitioner) who treats dozens of patients in her clinical practice. “I just realized - she wrote - that my (graduate) students have been using Chat GPT. Some of them are pretty obvious. I do have them use Turnitin for the final paper, but I'm not sure if it checks for AI..."

First, let’s point out that by being able to detect the use of LLM at all, our instructor had already proven to be far from naive, and in fact well ahead of the crowd, but her email was a harbinger of a looming crisis that calls for a radical reconceptualization of many aspects of how we teach writing - and not only - in college courses.

The availability of LLM-based tools that use conversational, everyday language Q&A format to communicate, rather than computer code or complex specialized syntax, has made them immediately and widely accessible and unleashed them onto the world, ready or not.

GPT in Chat GPT (hence CG), stands for "Generative Pre-trained Transformer." And as GG itself explains (we asked for a sixth-grader level explanation), “it's a computer program that can do a lot of things with words, like writing stories or translating. The 'generative' part means that it can create new text that looks similar to the text it has seen before: a new story that sounds like it was written by a human. The 'pre-trained' means that it had learned a lot about language before it even starts working on a specific question, by reading a lot of text, so that it can understand how words are used and how sentences are structured. Finally, the 'transformer' part describes the way it's designed. It's like a big puzzle that can put together words in the right order to make sense. It's really good at understanding long sentences and paragraphs of text.“

In the first part of this presentation we provide a general, high-level overview of what Large Language Models (LLMs) are, and how they work. We then consider LLMs in the context of three key concepts related to technological change and its societal impact, that allow us to understand the magnitude of this technological advance: "disruptive innovation" (Christensen),  "exponential technological leap" (Moore and Kurzweil), and "paradigm shift" (Kuhn). We evaluate the potential of the technology to disrupt common, well-established educational teaching and learning practices (a lecture, a research paper, preparing for an exam, and so on).

We then introduce, explain, and operationalize (show how to implement in practice) the DI-AMP (disruptive impact analysis and mitigation process) that we developed specifically to address this formidable challenge. DI-AMP is a workflow (you can think of it as a step-by-step toolkit) that guides instructors through a process of identifying the assignments in their course(s) that have been disrupted by this technology: many instructors are not aware that the disruption had already taken place, and that the papers they are painstakingly correcting were not entirely written by a human. Once the disruption points are identified, we offer specific procedures and alternatives instructors can use to update the assignments to be a better fit for  the CG world in which we now all exist.

Our disruptive impact analysis and mitigation process (DI-AMP) considers the LLM challenge holistically, taking into account three aspects of using LLMs in college courses that are helpful in identifying and understanding the challenge, defining its boundaries, and developing viable, functionally sound solutions. The three aspects we address are: ethical, epistemological, and methodological

Ethical use considerations include creating well-reasoned guidelines for using LLM in course assignments (acknowledging, citing, limits on use, including full text of the base/source-chat sequence as an appendix, to mention just a few). We provide ready-to-use how-to instruction examples )copy-and-paste ready), course and assignment guideline sheets, and assignment checklists (all available under CC BY-NC-SA license). 

Epistemological questions include understanding the importance of the balance between what is already known and included in the prompt (question posed to LLM) and what output is produced, a challenge that is succinctly, if crudely expressed by the old acronym GIGA (garbage in, garbage out). We look at how to build conceptual and procedural knowledge to ensure best results, including prompting (aka prompt engineering).

Finally, in focusing on the methodological aspect we explore common risks of relying on LLM output (types of common errors aka hallucinations and identifying them, spotting biases that may reflect the training dataset), and ways to bring all three aspects (ethical, epistemological and methodological) bear on actual practices and strategies. 

At this stage we also recognize that DI-AMP is a transitional tool, a stop-gap measure created to let instructors adapt existing assignments to the new reality, before assignments that tacitly and automatically assume that LLMs are available and are being used become routine. DI-AMP is also a stop-gap measure for students that re-designs assignment to combine specific academic field subject matter with LLM skill development as simultaneous and parallel processes, until students are able to develop the basic LLM prompting (aka prompt engineering) skills at earlier stages of their education as a matter of course.

That skill set includes principles of prompt engineering (prompting) as well as the understanding of different categories of and methods for identifying LLM hallucinations (a term often used to designate several types of errors in LLM output). 

Finally, we conclude this presentation with real-world case studies of applying DI-AMP to three college-level assignments disrupted by the technology, and transform them into assignments that potentially produce similar learning outcomes  and  **deliberately, methodically, and reflexively** incorporate structured use of CG as a tool, not only to mitigate the negative effects of this disruptive technology, but also to help  students develop a new set of skills for the new post-CG world.