Main Article Content

Abstract

This paper investigates and compares the use of three Large Language Models (LLMs), i.e., ChatGPT4, Google Gemini, and Claude 3.5, as decision support systems to plan the syllabus of a university course. The experiment was conducted in the context of the bachelor’s degree in Music Informatics of the University of Milan. The course under the exam focuses on the MIDI protocol, a very technical subject in the field of sound and music computing. The responses provided by LLMs have been evaluated by the author, a domain expert who has been teaching the subject for more than 10 years. From here emerges the provocative question in the title: can an AI-based chatbot prove to be more effective than an experienced teacher in defining educational objectives, materials, and the lesson plan for a university course? The results of the experimentation show that all three LLMs enable the formulation of a fairly comprehensive syllabus, allowing for the structuring of a university course. Their responses present interesting points of convergence in terms of both structure and content, while also highlighting some specificities. At present, biases and limitations still exist that make chatbots excellent co-pilots but do not replace the role of the teacher. Finally, a generalization is proposed to outline potential benefits and risks in the applicability of LLMs to the planning of educational activities.

Keywords

Large Language Models Syllabus University Course Educational Activities

Article Details

How to Cite
Ludovico, L. A. (2025). Is ChatGPT better than me? Analyzing the applicability of Large Language Models to the syllabus of a university course. Journal of E-Learning and Knowledge Society, 21(1), 201-210. https://doi.org/10.20368/1971-8829/1136193

References

  1. Brown, P. F., Della Pietra, V. J., Desouza, P. V., Lai, J. C., & Mercer, R. L. (1992). Class-based n-gram models of natural language. Computational linguistics, 18(4), 467-480.
  2. Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al. (2020). Language models are few-shot learners. Advances in Neural Information Processing Systems, 33, 1877–1901.
  3. Chang, Y., Wang, X., Wang, J., Wu, Y., Yang, L., Zhu, K., Chen, H., Yi, X., Wang, C., Wang, Y., Ye, W., Zhang, Y., Chang, Y., Yu, P. S., Yang, Q., & Xie, X. (2024). A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology, 15(3), 1-45.
  4. Christiano, P. F., Leike, J., Brown, T., Martic, M., Legg, S., & Amodei, D. (2017). Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30.
  5. Perković, G., Drobnjak, A., & Botički, I. (2024). Hallucinations in LLMs: Understanding and addressing challenges. 2024 47th MIPRO ICT and Electronics Convention (MIPRO), 2084-2088. IEEE.
  6. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł, Polosukhin, I. (2017). Attention is all you need. Advances in Neural Information Processing Systems, 30.
  7. Wei, J., Bosma, M., Zhao, V. Y., Guu, K., Yu, A. W., Lester, B., Du, N., Dai, A. M., & Le, Q. V. (2022). Finetuned language models are zero-shot learners. International Conference on Learning Representations (ICLR 2022). https://iclr.cc/virtual/2022/oral/6255