This panel will discuss how scientific and academic journals should respond to the growing concerns related to large language models (LLMs) and authorship. For example, are concerns regarding the use of LLMs for research writing overblown? Are there any appropriate ways for using LLMs in academic scholarship? Should journal networks update policies so that researchers can specify that text generated by ChatGPT (or any other AI tools) was not used in their work?
Dr. Gordon Hull (Professor of Philosophy & Director of Center for Professional and Applied Ethics, UNC Charlotte)
Gordon Hull's work is in ethics and political philosophy. His current research focuses primarily on the intersection of law, regulatory policy and political theory, with particular emphases on intellectual property, privacy and AI. He also works in the history of philosophy - the early modern period in particular - and on 20th Century European theory. He is the author of The Biopolitics of Intellectual Property (2019) and Hobbes and the Making of Early Modern Thought (2008).
Michael Flierl (Associate Professor & Student Learning Librarian, OSU Libraries)
Michael Flierl is an Associate Professor and Student Learning Librarian in The Ohio State University Libraries. His research interests include information literacy, student-centered teaching and learning environments, and AI. Michael has recently explored the dual-use nature of advanced foundational AI systems—presenting on this subject to the University Senate, OSU College of Medicine’s Education Leadership Team, and international conferences.