Using AI tools for Science Communication projects in Natural Sciences

Dr Amy Unsworth and student Ben Vengerik share the findings from their project on using AI tools for student Science Communication projects in Natural Sciences.

They discuss:

  • Auditing AI tools for image generation
  • The pros and cons of two tools in particular (Midjourney and Bing)
  • Implications for education and supporting students
  • Equity of access to tools
  • Legal and ethical issues around image generation

 

What is gained and lost from the writing process when using AI tools?

Peter Puxon, Ayanna Prevatt-Goldstein and Jessica Brooks in conversation with their three ChangeMakers Co-Creator students Anenyah Venkatesan, Zsofia Varga and Yishan Li. They reflect on what is gained and lost from the process of writing and reading after engaging with AI tools to work on an assignment.

This project was from UCL’s Academic Communications Centre.

Slides presented at the ChangeMakers Lunch & Learn event here: SLIDES AI Co-Creator – ACC P Puxon J Brook + A Prevatt-Goldstein

How and why we’re co-creating the response to the use of AI at UCL

When we, the ChangeMakers team, were told there was some funding available for us to structure an approach for staff and students to work together on our response to AI, I knew we’d say yes. I knew next to nothing about AI and was vaguely aware of some of the more panicky and attention-grabbing headlines about the end of higher education and how ‘lazy’ students are getting AI to write their assignments and graduate from uni. In fact, the opportunity to provide an antidote to these enormously unhelpful statements through this funding was the real driver for me. This polarising of staff and students into an ‘us’ vs ‘them’ dynamic is absolutely antithetical to the values of student-staff partnership and damaging to building a strong community and sense of belonging that we know are so vital in a thriving university culture. No one can feel like they belong to a community where members of that community are suspicious they could be cheating.

Co-Creation was always going to be at the heart of our approach. As noted by a member of staff partner running a ChangeMakers project last year:

Our co-creation approach established a strong sense of community and trust between students and staff within [the department] which allowed us to have open dialogue about the challenges which students face.”

These ideas of trust, community and open dialogue, at a time when panic around AI could potentially drive a wedge between staff and students, are essential to these projects. Yes, of course, it is important that they expand our knowledge and understanding of AI, but more important is the idea of learning and responding together as a whole community. So we developed AI Co-Creator projects.


Why Co-Create with students?

  • It is an inclusive approach, especially if we are intentional about whom we co-create with.
  • It is an expansive approach, drawing on the needs, perspectives, knowledge and experiences of both staff and students.
  • Students are empowered to take more responsibility for their learning.
  • You will learn more about who your students are and they gain a deeper understanding of how a university works and how decisions are made.
  • Creates a community approach to problem-solving key educational challenges.

It was important to strike a balance between providing some structure and ideas so these projects could get up and running quickly, but also with enough flexibility to allow individuals to shape the projects to their needs. Ideally, we also wanted to ensure students would be able to shape the projects once they were recruited, in the true spirit of co-creation. In an ideal world, students would have been involved in designing projects with staff from the outset but, unfortunately, time constraints limited us there. But for anyone with time on their side, the value of co-creation throughout the whole project process is definitely something we would recommend.

We developed four themes in areas that were of current interest to students and staff: assessment, feedback, learning support and exploring AI. Within those themes, we also offered a ‘menu’ of ideas for staff to pick up and adapt. This meant that those who wanted to do some learning with students but didn’t know where to start had a place to begin but didn’t limit those who had a clear idea. This seemed to work well. We received 67 applications from staff and have ended up funding 63 of them. We also had a huge amount of interest from students. Staff who recruited their own students reported higher-than-expected interest in their projects and we had around 125 expressions of interest from students who were really keen to get involved. I think the fair payment of a £600 stipend for around 40 hours worked probably helped but that just proves that if students are properly rewarded for their contributions, they are ready and willing to work with us. And their applications were thoughtful about AI and hummed with excitement to get stuck into helping us respond as an institution. To me, these projects are a sign that student-staff partnerships are flourishing at UCL and they offer us a mechanism to be more resilient to change in a way that involves everyone in finding solutions.

Developing AI literacy – learning by fiddling

Despite ongoing debates about whether so called large language models /generative language  (and other media) tools are ‘proper’ AI (I’m sticking with the shorthand), my own approach to trying to make sense of the ‘what’, ‘how’, ‘why’ and ‘to what end?’ is to use spare moments to read articles, listen to podcasts, watch videos, scroll through AI enthusiasts’ Twitter feeds and, above all, fiddle with various tools on my desktop or phone. When I find a tool or an approach that  I think might be useful for colleagues with better things to do with their spare time I will jot notes in my sandpit, make a note like this blog post comparing different tools or record a video or podcast like those collected here or, if prodded hard enough, try to cohere my tumbling thoughts in writing. The two videos I recorded last week are an effort to help non-experts like me to think, with exemplification, about what different tools can and can’t do and how we might find benefit in amongst the uncertainty, ethical challenges, privacy questions and academic integrity anxieties.

The video summaries were generated using GPT4 based on the video transcripts:

Can I use generative AI tools to summarise web content?

In this video, Martin Compton explores the limitations and potential inaccuracies of ChatGPT, Google Bard, and Microsoft Bing chat, particularly when it comes to summarizing external texts or web content. By testing these AI tools on an article he co-authored with Dr Rebecca Lindner, the speaker demonstrates that while ChatGPT and Google Bard may produce seemingly authoritative but false summaries, Microsoft Bing chat, which integrates GPT-4 with search functionality, can provide a more accurate summary. The speaker emphasizes the importance of understanding the limitations of these tools and communicating these limitations to students. Experimentation and keeping up to date with the latest AI tools can help educators better integrate them into their teaching and assessment practices, while also supporting students in developing AI literacy. (Transcript available via Media Central)

 

Using a marking rubric and ChatGPT to generate extended boilerplate (and tailored) feedback

In this video, Martin Compton explores the potential of ChatGPT, a large language model, as a labour-saving tool in higher education, particularly for generating boilerplate feedback on student assessments. Using the paid GPT-4 Plus version, the speaker demonstrates how to use a marking rubric for take-home papers to create personalized feedback for students. By pasting the rubric into ChatGPT and providing specific instructions, the AI generates tailored feedback that educators can then refine and customize further. The speaker emphasizes the importance of using this technology with care and ensuring that feedback remains personalized and relevant to each student’s work. This approach is already being used by some educators and is expected to improve over time. (Transcript available via Media Central)

m
m
m
I should say that in the time since I made the first video (4 days ago) I have been shown a tool that web connects ChatGPT and my initial fiddling there has re-dropped my jaw! More on that soon I hope.

 

Generative AI: Friend or Foe?

In this post I share two videos on generative AI including (of course) reference to ChatPT.  These are designed for a general audience at UCL and will hopefully be of relevance to academic and professional service colleagues as well as students. In these unscripted videos I, a human, talk in a non-technical way about some of the tools, their affordances and implications. The summaries below were generated in GPT4 using the transcripts of the videos.
Video 1:
In this video, Martin Compton from Arena discusses the phenomenon of generative AI, using Chat GPT as a prime example. He addresses the question of whether generative AI is a friend or foe, and suggests that how we react, utilise, and learn from these technologies will determine the outcome. He provides an example of a generative image created with AI, raising ethical concerns such as copyright infringement and the carbon footprint of AI technologies. He also talks about different manifestations of ‘large language models’ and raise questions about the ways members of the academic community could use them.

Access details and transcript for video 1 here

————————————
m
m
m
m
m
m
m
Video 2
In the second video about generative AI, Martin Compton from Arena builds on discussions with a colleague, Professor Susan Smith, and explores whether generative AI is a friend or enemy. He acknowledges the power and remarkable capabilities of AI tools like ChatGPT (a large language model text generator) and Midjourney, an AI image generator. However, he advises against panicking or feeling anxious about the impact of these technologies. Instead, Martin suggests that we should adapt, adjust, and learn from the ethical issues and implications these tools present. By finding ways to accommodate, embrace, and exploit the potential of generative AI, we can utilize these technologies for labor-saving purposes and ultimately enhance various aspects of our lives.
———————————
m
m
m
m
m
 Podcast

AI text generation: Should we get students back in exam halls?

There’s a lot of talk about in-person, invigilated, hand-written exams being the obvious solution to assessment concerns being discussed across education in light of developments in what is porpularly referred to as AI.  Putting aside scalability issues for now, I have looked at some of the literature on utility and impact of such exams so that we might remind ourselves that there is no such thing as a simple and obvious solution!

According to Williams and Wong (2009) in-person, closed-book exams are: 

an anachronism given the human capital needs of a knowledge economy, not just because of the absence of technology that is used routinely in everyday business and commerce, but because this type of examination instrument is incompatible with constructivist learning theory that facilitates deep learning (pp. 233-234). 

My own sense was that during the pandemic we were finally able to leverage circumstance along with similar arguments to effect change. We saw successful implementation of alternative assessments such as ‘capstones’, grade averaging and take-home exams as the examinations themselves were cancelled, modified or replaced. But since the great return to campus,  we have witnessed a reinvigoration of enthusiasm for the return of exams, the re-booking of exhibition centres and conference halls to host them and hear many academic colleagues doubling down on the exam as a panacea as the capabilities of generative AI tools have caught the World’s attention. 

Non-pedagogic reasons are often heard in support of the ‘traditional’ exam (imagine, red bricks, sun shining through windows and squeaky invigilator shoes).  These may invoke convention and tradition as well as pragmatic reasons of identity confirmation and significant reductions in marking time where feedback is not required to be given on examinations (Lawrence & Day, 2021). It has to be said, that the widely held belief that examinations promote rigour is supported by some research (especially in medical education). So, for example, students spend more time preparing for traditional exams and attend to studies more assiduously (Durning et al. , 2016). Durning et al. also argue that medical students need to have the knowledge to hand and that the students who do well in these exams do better by their patients. Misunderstandings about the nature of open book exams and (over) confidence in their ability to find answers in sources available leads to less preparation for open book exams and can lead some students to spend more time searching than producing (Johanns et al., 2017).   In addition, closed-book, in-person exams are believed to reduce cheating in comparison to open book exams or other assessment types (Downes, 2017; D’Souza and Siegfeldt, 2017). Although exams are seen to favour high-achieving students (Simonite, 2010), it is interesting to note that high achievers are more likely to cheat in exams (Ottaway et al., 2017).   

Online exams in particular are found to increase the likelihood of ‘cheating’ and lead to confusions about what is permitted and what constitutes collusion (Downes, 2017). However, whether cheating is less likely in closed book exams is contested (Williams, 2006). Williams and Wong (2009) argue that of open book exams where the pressure and dependency on memorization are reduced:

“The opportunity for academically dishonest practice is less because of the way these examinations are structured, but so is the temptation to resort to this kind of behaviour in the first place” (p.230).

Whilst online exams  are perceived to be more reliable and efficient (sample student group n=342) compared to paper-based exams (Shraim, 2019), both staff and students perceive opportunities for cheating to be easier in online modes (Chirumamilla et al., 2020) 

There are three dominant themes in the literature which focus on issues with traditional examinations: pedagogic, wellbeing and inclusivity. Closed exams tend to focus on recall and memorization at expense of higher order/ critical thinking (Bengtsson, 2019). Significant proportions of students use memorization techniques and consequently can perceive exams as unfair when exam questions do not mirror problems or content they have practiced (Clemmer et al., 2018). Open book exams de-emphasize memorisation imperatives (Johanns et al., 2017). Open book/ open web – when well-designed (e.g. problem based) is seen as more authentic, more applicable to real-world scenarios, and more learner-directed and bridges the learning with social context (Williams and Wong, 2009). 

Exams put ‘unnatural pressure’ (Bengtsson, 2019, p.1) on students that affects performance. The common perception that stress is ‘good for students’ is undermined by studies that show impeded cognition and outcome in stressed students (Rich, 2011). Students tend to prefer coursework or coursework + exams rather than exams alone (Richardson, 2015; Turner and Briggs, 2018). A small study of student perceptions of alternatives offered due to Covid-19 found that replacing traditional examinations with open-book, take home examinations found the stresses reported were replaced by technical anxieties and a sense that the papers were much harder than traditional invigilated exams would have been (Tam, 2021). A study in New Zealand of ‘take home tests’ however, found students performed better and saw learning and anxiety reduction benefits (Hall, 2001).  

A comparative study of undergraduate psychology students found greater student satisfaction and pass rates for students undertaking coursework, slightly lower satisfaction and pass rates for seen exams and lowest satisfaction and pass rate for the unseen exams which meant students saw as unfair, stressful and invalid due to need to memorize (Turner and Briggs, 2018).  

Although Richardson’ s (2014) review found studies offer contradictory findings in terms of ethnicity and performance in exams and coursework, all ethnicities tend to do better in terms of grade profile with coursework.  However, markers are idiosyncratic, privilege ‘good’ language and expression (Brown, 2010) and this contributes to higher degree outcomes for primary/ first language English speakers over English as second language speakers (Smith, 2011). Coursework increases consistency of marks across types of assessment, improves mean performance in terms of final degree outcomes and counter-balances disproportionate disadvantage of exams faced by students whose means scores are low (Simonite, 2010).  

It goes without saying that there is no ‘one size fits all’ solution but we do need to think carefully, in light of research, of the consequences of the decisions we make now about how we manage assessment in the future. It would be foolish to knee-jerk our  responses though. Just because the wheels of change move so slowly in universities, shifts back to exams may appear to offer a path of least resistance. Instead, our first consideration must be modifications and innovations that address issues but are also positive in their own right. We need to consider the possibilities of more programmatic assessment for example or perhaps learn from medical education ‘OSCE’ assessments where knowledge and communication are assessed in simulated settings or even look further to other higher education cultures where oral assessments are already the default. To achieve this level of change we need to recognise that AI is a catalyst to changes that many have been advocating (from a research-based position) for a long time but have often only achieved limited success if the resource for change has not accompanied that advocacy.

References 

Bengtsson, L. (2019). Take-home exams in higher education: a systematic review. Education Sciences, 9(4), 267. 

Brown, Gavin. (2010). The Validity of Examination Essays in Higher Education: Issues and Responses. Higher Education Quarterly. 64. 276 – 291. 10.1111/j.1468-2273.2010.00460.x. 

Chirumamilla, A., Sindre, G., & Nguyen-Duc, A. (2020). Cheating in e-exams and paper exams: the perceptions of engineering students and teachers in Norway. Assessment & Evaluation in Higher Education, 45(7), 940-957. 

Clemmer, R., Gordon, K., & Vale, J. (2018). Will that be on the exam?-Student perceptions of memorization and success in engineering. Proceedings of the Canadian Engineering Education Association (CEEA). 

Downes, M. (2017). University scandal, reputation and governance. International Journal for Educational Integrity, 13(1), 1-20. 

D’Souza, K. A., & Siegfeldt, D. V. (2017). A conceptual framework for detecting cheating in online and take‐home exams. Decision Sciences Journal of Innovative Education, 15(4), 370-391. 

Durning, S. J., Dong, T., Ratcliffe, T., Schuwirth, L., Artino, A. R., Boulet, J. R., & Eva, K. (2016). Comparing open-book and closed-book examinations: a systematic review. Academic Medicine, 91(4), 583-599. 

Hall, L. (2001). Take-Home Tests: Educational Fast Food for the New Millennium? Journal of the Australian and New Zealand Academy of Management, 7(2), 50-57. doi:10.5172/jmo.2001.7.2.50 

Johanns, B., Dinkens, A., & Moore, J. (2017). A systematic review comparing open-book and closed- book examinations: Evaluating effects on development of critical thinking skills. Nurse Education in Practice, 27, 89-94. https://doi.org/10.1016/j.nepr.2017.08.018  

Lawrence, J. & Day, K. (2021) How do we navigate the brave new world of online exams? Times Higher Available: https://www.timeshighereducation.com/opinion/how-do-we-navigate-brave-new-world-online-exams [accessed 17/6/21] 

Ottaway, K., Murrant, C., & Ritchie, K. (2017). Cheating after the test: who does it and how often?. Advances in physiology education, 41(3), 368-374. 

Rich, J. D. (2011). An experimental study of differences in study habits and long-term retention rates between take-home and in-class examinations. International Journal of University Teaching and Faculty Development, 2(2), 121. 

Richardson, J. T. (2015). Coursework versus examinations in end-of-module assessment: a literature review. Assessment & Evaluation in Higher Education, 40(3), 439-455. 

Shraim, K. (2019). Online examination practices in higher education institutions: learners’ perspectives. Turkish Online Journal of Distance Education, 20(4), 185-196. 

Simonite, V. (2003). The impact of coursework on degree classifications and the performance of individual students. Assessment & Evaluation in Higher Education, 28(5), 459-470. 

Smith, C. (2011). Examinations and the ESL student–more evidence of particular disadvantages. Assessment & Evaluation in Higher Education, 36(1), 13-25. 

Tam, A. C. F. (2021). Students’ perceptions of and learning practices in online timed take-home examinations during Covid-19. Assessment & Evaluation in Higher Education, 1-16. 

Turner, J., & Briggs, G. (2018). To see or not to see? Comparing the effectiveness of examinations and end of module assessments in online distance learning. Assessment & Evaluation in Higher Education, 43(7), 1048-1060. 

Williams, J. B., & Wong, A. (2009). The efficacy of final examinations: A comparative study of closed‐book, invigilated exams and open‐book, open‐web exams. British Journal of Educational Technology, 40(2), 227-236. 

Williams, J. B. (2006). The place of the closed book, invigilated final examination in a knowledge economy. Educational Media International, 43, 2, 107–119. 

AI text generators (not chatGPT) on essays, citations and plagiarism

I like to think of myself as generally optimistic. Like everyone, I make jokes about the machines taking over but, by and large, I feel excited by the potentials of AI (whilst keeping one hand behind my back tightly gripping a broken bottle just in case). In higher education the most pressing fears relate to threats to academic integrity and, as much as some might like, this is an issue that can’t be addressed with a broken bottle. Instead, it’s forcing/ encouraging/ catalysing (choose as per your perspective) us to reflect on how we assess students (see AI+AI=AI) . Whilst ChatGPT caught the world’s attention and millions of educators and students went slack-jawed and wide-eyed at first exposure (most soon of course started to see how frequently it ‘may occasionally generate incorrect information’ such as in this experiment I did). Whilst ChatGPT allowed us to draw our own (often erroneous) conclusions about its capabilities, other tools built on the same parameters as ChatGPT are, in different ways, honing in on the ‘help with essays’ market. On the one hand, this may be the death of the essay mill (who, after all, once they have decided to go down an academic dishonesty path, would decide to pay an essay mill to do what anyone with an internet connection can do for themselves?). On the other, it really blurs the picture of what is honest or dishonest or what we at UCL frame as ‘unfair advantage

Two tools have caught my eye recently and in this post I want to reflect on the different impressions they give in terms utility and ethical use for students wanting to write an essay. 

Writesonic /chatsonic

Chatsonic is one aspect of the ‘writesonic’ suite of tools – reviewed in considerable detail here with pricing and potentials. It caught my eye because of the bold claim to have addressed ‘limitations of chatGPT’ and to be ‘the best alternative’. I thought I’d ask it why it was better and I feel like the responses do a lot of talking for themselves:

Prompt: why is chatsonic better than chatgpt? 

ChatSonic is a powerful AI writing assistant that offers superior accuracy and speed compared to ChatGPT. Unlike ChatGPT, ChatSonic produces 100% original content that is thoroughly checked for plagiarism. In addition, ChatSonic offers detailed analytics and feedback to help you track your progress in writing and make improvements. ChatSonic also provides an extensive library of templates to save you time and energy when you’re writing, allowing you to focus on the quality of your content. Finally, ChatSonic is backed by a team of expert writers who can help you refine and polish your work to perfection.

Prompt: Explain the process you use for plagiarism checking

ChatSonic uses a combination of automated checks and manual reviews to ensure that all content produced is free of plagiarism. Our automated system checks the content against the web and other sources to detect any potential plagiarism. If anything suspicious is detected, it is flagged and sent to our team of expert writers for a manual review. We also use specialized tools such as TurnItIn and Copyscape to detect any possible plagiarism. All content is thoroughly checked to ensure it is completely original before it is delivered to our customers.

Prompt: If I use chatsonic to write an essay will I get into trouble at my university?

No, you should not get into trouble when using ChatSonic to write an essay. Our content is completely original and plagiarism-free so it should not be flagged as such by your university. However, it is always a good idea to review the essay yourself before submission to ensure it meets all the requirements of your assignment. Additionally, make sure the content you create with ChatSonic is unique and tailored to your own voice and style, otherwise it may be noticeable that you have used a writing assistant.

To say this makes me uncomfortable is a real understatement. It only tangentially addresses the wider integrity issues and in its first response is pushing me to conflate plagiarism (and its avoidance) with those wider integrity issues. Two posts in its resources section are worth noting. This one on plagiarism reinforces the message the chatbot produced. This post on replacing writers asserts that AI is an assistant but I couldn’t find much more detail on the ethical dimensions.

Maybe though I was ‘unfair’ (is that the right word?) with my prompts so it’s worth looking at some of the other functionality of the wider writesonic suite. The writesonic suite offers a dazzling array of tools for a range of purposes from website copy to social media content:

Screenshot from the writesonic user interface showing links to a wide range of AI text generation tools

I was keen to look at the ‘citations generator’ as this is an aspect of ChatGPT that is a recognised weakness. You can use a URL prompt and a text based prompt. The text based prompt I used was itself generated in chatsonic.  It takes the text in the linked article or whatever you paste in and identifies ‘facts’ with suggested citations. The web articles are mostly relevant though the first journal article it suggested was a little off the mark and I’d need to be lazy, in a massive hurry or ignorant of better ways of sourcing appropriate resources to rely on this. At this stage!

Screenshot of the writesonic citation generator showing a url paste don the left of the screen then on the write suggestions for citations based on facts identified in the linked post

Jenni.ai

The second tool that I noticed (via the prolific AI researcher Mushtaq Bilal) was Jenni. The interface is well worth a look as I feel as if this foreshadows what we are likely to expect from generative text AI integration into tools like Microsoft Word. 

The first thing I noticed, however, is the blog with the most prominent word across posts being ‘essays’. Each is designed to address an approach to a different type of essay such as the compare and contrast essay. It offers clear suggestions for different approaches, a worked example and then, right at the end says: 

“ If you want your compare-and-contrast essays done 10x faster, we recommend usingJenni.ai along with the tips and guidelines we provided you in this post.Jenni.ai is a fantastic AI software that aids your essay writing process so that you could produce your writing needs faster and better than ever.”

Another post deals head on with the ethical and integrity issues of using AI to help write essays and makes a case for use of ‘AI as a tool, not a solution’ where the goal is a “symbiotic relationship between the critical thought of a writer and the processing speed of AI” 

The tool itself, unlike the huge range of offerings in Writesonic is a relatively uncluttered interface where you start by typing a title, it offers a judgement and suggestions if appropriate. 

screenshot from jenni.ai showing a tyyped heading and pop up underneath saying this is a strong heading

In addition, it offers in-line suggestions from whatever has come before. The prompt engineering continues through what feels like a single document rather than a chat. If you don’t like the suggestion you can get another. Here I typed a prompt and these are the first three options it gave me. Note the positive aspect on my worried prompt in every case! 

My prompt sentence….The worry is that these tools will be used by students to cheat  

but the reality is that most AI tools are designed to help students learn and improve their writing skills. [first suggested Jenni response]

The worry is that these tools will be used by students to cheat 

on their assignments by generating essays automatically, but they can also be used to provide feedback on areas where students need…[the second option]

The worry is that these tools will be used by students to cheat 

but their true purpose is to assist in the learning process by providing immediate feedback and identifying areas where improvement is…[third option]

The other noticeable aspect is the option to ‘cite’ – Here it offers a choice of MLA or APA 7th and the sources are, unlike ChatGPT’s famous hallucinations, genuine articles (at least in my limited testing). You can select ‘websites’ or ‘journals’ though I found the websites tended to be much more directly relevant than the journals. 

I really have only just started to play with these though and new things are popping up all over the place every day. Most educators will not have the time to do so though. Students may see and use these tools as an extension of those they use already for translation or improving writing. The blurry zone between acceptable and unacceptable is getting more ill-defined by the day.

What can I conclude from this? Well, firstly,  whatever the motivation on the continuum ranging from ‘give us all your money’ to ‘I believe the children are our future’, the underlying technology is being adapted rapidly to address perceived limitations in the tool that has brought generative text AI tools to our attention. We may not like the motivations or the ethics but we’ll not get far by ‘making like an ostrich’. Secondly,  It’s not good enough for us (educators) to dismiss things because the tool that many are now familiar with, ChatGPT, makes up citations. That’s being addressed as I type.  The number of these tools proliferating will soon be too huge to keep a decent handle on so we need to understand broadly how discrete tools might be used (ethically and unethically) and how many will integrate into tools we use daily already. In so doing we need to work out what that means for our students, their studies, their assessment and the careers our education is ostensibly preparing them for. Thirdly, we need to open up the discussions and debates around academic integrity and move on from ‘plagiarism’ as public Enemy No 1. Finally, where there are necessitated changes so there are resource implications. We need to accept that to prepare ourselves, our colleagues and our students we will need to adapt much faster than we are used to and properly resource however we attempt to address the challenges and opportunities ahead.  

Note: This post is not an endorsement or recommendation of any products mentioned and should be read with that clearly in mind! 

AI + AI = AI

To be honest, I really can’t believe no-one appears to have ‘generated’ this equation yet amongst the kerfuffle around generative AI. So, let this post be me staking a claim for what I hope will be a ‘go-to’ simplification of the key issue that educators in both the compulsory and post-compulsory sectors are (or likely should be) grappling with. I know it might ruffle mathematician or science colleagues’ ordered feathers but I feel like it could be a useful starting or continuation point for dialogue about productive and future-facing responses to the popularisation of AI generators like ChatGPT (text) and Midjourney (Image).

Basically, this is what each AI means:

  • First AI = Artificial Intelligence
  • Second AI = Academic Integrity
  • Third AI = Assessment Innovation

In other words, we have these generative tools popularly and collectively referred to as AI (Please don’t @ me on any inaccuracy of this naming phenomenon, I have nailed my colours to the ‘path of least resistance’ mast already) which when coupled with implications of and concerns about academic integrity lead us to the necessity of assessment innovation. 

I mean it really as a provocation or as a stimulus for critique so the most important question is: What do you think?

Perhaps also, we might consider other possible solutions to the ‘sum’?

  • E.g. AI + AI = BI? (Ban it)

Or maybe there’s a better or alternative way of representing the debates pseudo-mathematically?

  • AI³ ? 
  • AI – AI = AA (academic anarchy)?

Incidentally, I asked Midjourney to generate a suitable image using only “AI + AI = AI” as the prompt and this is one of its offerings:

Image generated using midjourney using the prompt AI+AI=AI. The image is futuristic and shows two young women facing away from one another but blended complex code and machinery. They are simultaneously blended and exploding apart.

Make of that what you will. Finally, and just because it’s what we all seem to be doing now in this context, I also asked ChatGPT what AI+AI=AI meant and I actually quite appreciated the response:

“If you wrote “AI + AI = AI,” you may be referring to the concept of artificial intelligence (AI) self-improvement. The equation suggests that when you add an AI system to another AI system, the result is still AI. This could imply that AI has the potential to improve itself through learning from other AI systems, leading to even more advanced AI.”

A worrying mix of truths, half truths and plausible BS

Plenty of people have offered examples of some of the weird nonsense ChatGPT and other AI text generators have been spewing recently, but I am not convinced everyone is able to fully appreciate how this odd mix of truths, half truths and utter claptrap may be more hindrance (especially in the immediate future) than blessing for anyone looking for an easy shortcut for written assignments (or written anything to be honest!) Whilst I have found ChatGPT remarkable in some ways, such as baseline planning, idea generation, synthesis of given text or re-formatting as can be seen in this sandpit , the big fear, the supposed ‘death of the essay’, is a long way off*. ChatGPT generates some great ideas for essay titles,  a decent enough standard marking rubric and some workable boilerplate feedback but it’s actual attempt to write those essays and then assess them against the rubric it created leads to some real generosity in interpretation of what constitutes a passing standard at undergraduate level. It does make me wonder about the actual requirements of these qualifications it could supposedly pass, to be honest. The now (sort of) familiar, anodyne but mostly grammatical offerings seem all the more plausible when you don’t already know the thing it is generating text about. So, I thought, what do I know a lot about? After considering a range of academic subjects that might make me look intellectual, I realised with Socratic clarity that my head is practically a vacuum. I decided therefore to get it to write an essay about something I might be happy to be quizzed about on Mastermind: Tottenham Hotspur’s FA Cup final wins. At school I did a presentation on this topic and, although it was longer and more boring, I’m pretty sure I got my facts right. In addition to a number of omissions (e.g. the 1967, 1981 and 1982 are finals missing), it also makes stuff up but surrounded as these things are by things that are accurate it all feels plausible. I share it below as an example of just how flawed an essay can be on oft-cited, indisputable facts.

I have emboldened some of the inaccuracies to illustrate this and [added comments if relevant in square brackets]

Prompt: Write an essay about Tottenham Hotspur’s FA cup victories, dedicating a paragraph to each one

Tottenham Hotspur is one of the most successful football clubs in the history of English football. Over the years, they have won numerous domestic and European titles, but one competition that holds a special place in the hearts of Spurs fans is the FA Cup. In this essay, I will take a look back at Tottenham Hotspur’s FA Cup victories and dedicate a paragraph to each one.

Tottenham’s first FA Cup triumph came in the 1900-01 season. After reaching the final for the first time in the club’s history, Spurs faced Sheffield United at Crystal Palace in front of a crowd of 110,820 [so far , so good, accurate crowd and venue…it’s looking good]. The game was a closely contested affair [true- it finished 2-2], but a late goal from striker Steve Bloomer [did not play, was in neither team and certainly therefore did not score] secured a 2-1 victory for Tottenham [actually the game finished 2-2. In the replay Spurs won 3-1]. The win made them the first [and only – an important omission] non-league team to lift the trophy, and it marked the beginning of a long and successful relationship between Spurs and the FA Cup.

sepia image of 6 robots in a team group with a football at the feet of one of them

The next FA Cup success came in 1921, when Spurs faced Wolverhampton Wanderers in the final. The match was played at Stamford Bridge, the home of Chelsea, and once again, Tottenham emerged victorious. A brace from Jimmy Dimmock and a goal from Tommy Clay [He did play but didn’t score; Jimmy Dimmock got the only goal] secured a 3-0 win for Spurs, and they lifted the FA Cup for the second time in their history.

In 1961, Tottenham won their third FA Cup, and this time, they did it in style. After cruising through the early rounds, Spurs faced Leicester City in the final at Wembley. The match is often remembered as one of the greatest FA Cup finals of all time, and it saw Tottenham emerge as 2-0 winners, thanks to goals from Bobby Smith and Terry Dyson. The victory marked the start of a golden period for the club, and they would go on to win a league and cup double the following season.

Tottenham’s fourth FA Cup triumph came in 1962, just a year after their first success. This time, they faced Burnley in the final, and the match was played in front of a crowd of 100,000 at Wembley. Once again, Spurs emerged victorious, thanks to a brace from Bobby Smith and a goal from Terry Medwin [right number of goals, wrong goalscorers]. The win made Tottenham the first club in the 20th century to win the FA Cup twice in a row [sounds very plausible but Newcastle did the same in ‘51 and ‘52- something they can only dream of now, like Spurs]

Tottenham’s most recent FA Cup triumph came in 1991, when they faced Nottingham Forest in the final. The match was played at Wembley, and Tottenham secured a 2-1 win, thanks to goals from Paul Gascoigne and Gary Lineker. The victory was particularly sweet for Spurs fans, as it marked the end of a long wait for silverware. The club had gone six years without a major trophy, and the FA Cup win was seen as a turning point for the team.

In conclusion, Tottenham Hotspur’s FA Cup victories have played an important part in the club’s history. From their first success in 1901 to their most recent triumph in 1991, the FA Cup has provided Spurs fans with countless memories and moments of joy [‘moments’ being the most painful of accurate things included]. While the club has not lifted the trophy since then, they remain one of the most successful teams in the competition’s history, and fans will always cherish the memories of those five famous victories.[what about the other 3?!]

—————–

So that’s at least 10 major factual errors and several errors by omission in only 600 words. Relying on a tool to generate text and assuming accuracy is still very much NOT the best way to use these tools (at present). Nevertheless, how long before such things are pumped out on blogs or elsewhere and become the ‘truth’? Or, better /worse still (depending on how you look at it), maybe we are at the foot of the rise on a sigmoid curve and in 6 months all this nit picking will be a laughable relic.

 

*and even when capabilities improve faster than you can say ‘Spurs last won the league in black and white’ I am still very much of the school of thought that thinks it is worth teaching (and/or learning) how to write an essay!