Special Session: Large Language Models - Are these the next pocket calculators?

Attention Presenters - please review the Speaker Information Page available here

This session, which is organized jointly by the ISCB Publication Committee and the ISCB Science in Society Committee, focusses on the topic of large language models (LLMs) and the consequences of their use in science and teaching. In this session, we are not focussing on the applications of LLMs in computational biology research. A list of talks at the conference addressing the use of LLMs in computational biology research will be appended to this abstract, as the conference program develops.

LLMs are a type of statistical learning method that can process and generate human-like language from models pre-trained on large text corpora. The availability of LLMs like chatGPT has ushered in a new phase of AI pervasion in society and of humans interacting with computers, with many open questions and problems. Here we will focus on issues pertaining to the use of LLMs in the scientific community. The session will begin with an introduction to the architecture of LLMs and the capabilities and limitations resulting from this architecture, as well as general issues with their use. After two presentations on the use of LLMs in scientific publishing and in teaching, respectively, the session will conclude with a discussion of ethical issues regarding the use of the new technology.

Chairs: Thomas Lengauer & Ragothaman Yennamalli

Notes added post-session:

1) The talk by David Leslie was based on his paper "Does the sun rise for ChatGPT? Scientific discovery in the age of generative AI" in the journal AI & Ethics which you can access freely at https://link.springer.com/article/10.1007/s43681-023-00315-3

2) If you have comments on the session or question to the presenters, please fill the online form at https://forms.gle/k2hvkcaJPhCFMLUi8

Schedule subject to change
All times listed are in CEST
Tuesday, July 25th
10:30-11:00
Invited Presentation: Large Language Models: Architectures, Training Strategies, and Applications
Room: Pasteur Lounge
Format: Live from venue

  • Christian Dallago, NVIDIA


Presentation Overview: Show

Large Language Models (LLMs) are a type of neural network architecture that can be trained on massive amounts of text data to learn the patterns and relationships between words and phrases. They evolved from Recurrent Neural Networks (RNNs) to massively scalable Transformers, found in systems like ChatGPT. One of the key drivers of the success of Transformer architectures is the use of self-attention mechanisms to capture long-range dependencies between distant words. However, training LLMs is nontrivial, requiring massive amounts of data and large compute clusters. However, once trained, LLMs have many practical applications, including text classification, sentiment analysis, machine translation, question answering, and more. In the medical setting, LLMs can be used for clinical decision support systems, drug discovery, and patient monitoring. In this talk, we will briefly touch on the fundamentals of LLM architectures, training and applications.

11:00-11:30
Invited Presentation: Plausible nonsense: An Editors worst nightmare
Room: Pasteur Lounge
Format: Live from venue

  • Alex Bateman


Presentation Overview: Show

Large language models are one of the most exciting AI developments of recent times, but like many powerful tools they can be used in positive and negative ways. LLMs offer powerful language editing at no/low cost as well as highly accurate translation of technical text. Access to these tools can help level the playing field for scientific authors from around the globe. However, there has also been a growing trend of fake papers that are the output of so-called papermills. Current LLMs are able to generate extremely plausible scientific text based on minimal prompts. While completely AI generated papers are unlikely to make their way past a rigourous peer review, there remains a grey area where papers are partially written by AI. One of the recent findings is that the latest generation of AI are prone to hallucinating incorrect facts in a very confident tone. Checking facts in AI generated texts is extremely time consuming and as an Editor I fear that authors who are pressed for time will not be as careful as they should be.
To provide authors with some guidances on the use of LLMs the ISCB in association with Editors from Bioinformatics and Bioinformatics Advances have drafted an LLM policy (https://www.iscb.org/iscb-policy-statements/iscb-policy-for-acceptable-use-of-large-language-models). This policy outlines the permitted uses of LLMs in the publication process and highlights what is not permitted. This policy is likely to evolve as the tools improve and peoples level of acceptance changes.

11:30-12:00
Invited Presentation: LLMs for teaching – game changers
Room: Pasteur Lounge
Format: Live from venue

  • Patricia Palagi


Presentation Overview: Show

AI is on everybody's lips these days. Large language models (LLMs) like ChatGPT are transforming how we interact with technology and use data. They seriously affect how we teach and learn - LLMs are game changers in this domain. In this talk, we will delve into the impact of LLMs on education, teaching and learning. We will start with ethical considerations such as bias, honesty and transparency and with a few words about some of the regulatory measures trying to mitigate these issues put into place so far by Higher Education Institutions. Additionally, we will explore some of the many challenges teachers face, including evaluating student progression and adapting learning outcomes to ensure effectiveness and achievability within this new context. Lastly, we will discuss the potential benefits of LLMs in education and training, such as personalized learning experiences for students and increased focus for teachers on critical educational aspects.

12:00-12:30
Invited Presentation: Scientific Discovery in the Age of Large Language Models
Room: Pasteur Lounge
Format: Live from venue

  • David Leslie


Presentation Overview: Show

The past few years have seen the rapid emergence of LLM applications that promise to accelerate scientific productivity and enhance the efficiency of research practice. These systems can carry out brute force text mining, “knowledge discovery”, and information synthesis and streamline scientific writing processes through the generation of literature reviews, article summaries, and academic papers. More recently, researchers have stitched together multiple transformer-based language models to create so-called “Intelligent Agents” which can perform “automated experiments” by searching the internet, combing through existing documentation, and planning and executing laboratory activities. In this talk, I explore some of the limitations of this new set of AI-supported affordances and reflect on their ethical implications for science as a community of practice. I argue that, amidst growing swells of magical thinking among scientists about the take-off of “artificial general intelligence” or the emergence of autonomous, Nobel Prize winning “AI scientists,” researchers need to take a conceptually sound, circumspect, and sober approach to understanding the limitations of these technologies. This involves understanding LLMs, in a deflationary way, as software-based sequence predictors. These systems produce outputs by drawing on the underlying statistical distribution of previously generated text rather than by accessing the evolving space of reasons, theories, and justifications in which warm-blooded scientific discovery takes place. Understanding this limitation, I claim, can empower scientists both to better recognise their own exceptional capacities for collaborative world-making, theorisation, interpretation, and truth and to better understand that contexts of scientific discovery are principal sites for human empowerment and for the expression of democratic agency and creativity.