Session 13

Large Language Models for linguistic annotation

Author

Masaki EGUCHI, Ph.D.

Modified

August 7, 2025

One-liner

We will explore how powerful and limited Large Language Models (LLMs) can be at the same time.

🎯 Learning Objectives

By the end of this session, students will be able to:

  • Describe how LLMs are trained generally and what LLMs do to produce language.
  • Explain the benefits and drawbacks of using LLMs for linguistic annotation.
  • Demonstrate/discuss potential impacts of prompts on the LLMs performance on linguistic annotation.
  • Design an experiment to investigate LLMs output accuracy on a given annotation task.

πŸ”‘ Key Concepts

  • Large Language Models (LLMs) and Language Generation
  • Prompt engineering
  • Fine-tuning

πŸ“š Required Readings

  • Mizumoto, A., Shintani, N., Sasaki, M., & Teng, M. F. (2024). Testing the viability of ChatGPT as a companion in L2 writing accuracy assessment. Research Methods in Applied Linguistics, 3(2), 100116. https://doi.org/10.1016/j.rmal.2024.100116

  • (Skim) Kim, M., & Lu, X. (2024). Exploring the potential of using ChatGPT for rhetorical move-step analysis: The impact of prompt refinement, few-shot learning, and fine-tuning. Journal of English for Academic Purposes, 71, 101422. https://doi.org/10.1016/j.jeap.2024.101422

Materials

Slides for the session

Other resources

Reflection

You can now: