Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Nils Feldhus
  • Aliki Anagnostopoulou
  • Qianli Wang
  • Milad Alshomary
  • Henning Wachsmuth
  • Daniel Sonntag
  • Sebastian Möller

Research Organisations

External Research Organisations

  • German Research Centre for Artificial Intelligence (DFKI)
  • Columbia University
View graph of relations

Details

Original languageEnglish
Title of host publicationProceedings of the 2024 International Conference on Information Technology for Social Good
Place of PublicationNew York, NY, USA
Pages225–230
Publication statusPublished - 4 Sept 2024

Abstract

For dialogues in which teachers explain difficult concepts to students, didactics research often debates which teaching strategies lead to the best learning outcome. In this paper, we test if LLMs can reliably annotate such explanation dialogues, s.t. they could assist in lesson planning and tutoring systems. We first create a new annotation scheme of teaching acts aligned with contemporary teaching models and re-annotate a dataset of conversational explanations about communicating scientific understanding in teacher-student settings on five levels of the explainee’s expertise: ReWIRED contains three layers of acts (Teaching, Explanation, Dialogue) with increased granularity (span-level). We then evaluate language models on the labeling of such acts and find that the broad range and structure of the proposed labels is hard to model for LLMs such as GPT-3.5/-4 via prompting, but a fine-tuned BERT can perform both act classification and span labeling well. Finally, we operationalize a series of quality metrics for instructional explanations in the form of a test suite, finding that they match the five expertise levels well.1

Keywords

    Dialogue, Discourse Analysis, Evaluation, Explanations

Cite this

Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues. / Feldhus, Nils; Anagnostopoulou, Aliki; Wang, Qianli et al.
Proceedings of the 2024 International Conference on Information Technology for Social Good. New York, NY, USA, 2024. p. 225–230.

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Feldhus, N, Anagnostopoulou, A, Wang, Q, Alshomary, M, Wachsmuth, H, Sonntag, D & Möller, S 2024, Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues. in Proceedings of the 2024 International Conference on Information Technology for Social Good. New York, NY, USA, pp. 225–230. https://doi.org/10.1145/3677525.3678665
Feldhus, N., Anagnostopoulou, A., Wang, Q., Alshomary, M., Wachsmuth, H., Sonntag, D., & Möller, S. (2024). Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues. In Proceedings of the 2024 International Conference on Information Technology for Social Good (pp. 225–230). https://doi.org/10.1145/3677525.3678665
Feldhus N, Anagnostopoulou A, Wang Q, Alshomary M, Wachsmuth H, Sonntag D et al. Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues. In Proceedings of the 2024 International Conference on Information Technology for Social Good. New York, NY, USA. 2024. p. 225–230 doi: 10.1145/3677525.3678665
Feldhus, Nils ; Anagnostopoulou, Aliki ; Wang, Qianli et al. / Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues. Proceedings of the 2024 International Conference on Information Technology for Social Good. New York, NY, USA, 2024. pp. 225–230
Download
@inproceedings{bc8295271aaf4c9288b08ab778d89afc,
title = "Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues",
abstract = "For dialogues in which teachers explain difficult concepts to students, didactics research often debates which teaching strategies lead to the best learning outcome. In this paper, we test if LLMs can reliably annotate such explanation dialogues, s.t. they could assist in lesson planning and tutoring systems. We first create a new annotation scheme of teaching acts aligned with contemporary teaching models and re-annotate a dataset of conversational explanations about communicating scientific understanding in teacher-student settings on five levels of the explainee{\textquoteright}s expertise: ReWIRED contains three layers of acts (Teaching, Explanation, Dialogue) with increased granularity (span-level). We then evaluate language models on the labeling of such acts and find that the broad range and structure of the proposed labels is hard to model for LLMs such as GPT-3.5/-4 via prompting, but a fine-tuned BERT can perform both act classification and span labeling well. Finally, we operationalize a series of quality metrics for instructional explanations in the form of a test suite, finding that they match the five expertise levels well.1",
keywords = "Dialogue, Discourse Analysis, Evaluation, Explanations",
author = "Nils Feldhus and Aliki Anagnostopoulou and Qianli Wang and Milad Alshomary and Henning Wachsmuth and Daniel Sonntag and Sebastian M{\"o}ller",
year = "2024",
month = sep,
day = "4",
doi = "10.1145/3677525.3678665",
language = "English",
isbn = "9798400710940",
pages = "225–230",
booktitle = "Proceedings of the 2024 International Conference on Information Technology for Social Good",

}

Download

TY - GEN

T1 - Towards Modeling and Evaluating Instructional Explanations in Teacher-Student Dialogues

AU - Feldhus, Nils

AU - Anagnostopoulou, Aliki

AU - Wang, Qianli

AU - Alshomary, Milad

AU - Wachsmuth, Henning

AU - Sonntag, Daniel

AU - Möller, Sebastian

PY - 2024/9/4

Y1 - 2024/9/4

N2 - For dialogues in which teachers explain difficult concepts to students, didactics research often debates which teaching strategies lead to the best learning outcome. In this paper, we test if LLMs can reliably annotate such explanation dialogues, s.t. they could assist in lesson planning and tutoring systems. We first create a new annotation scheme of teaching acts aligned with contemporary teaching models and re-annotate a dataset of conversational explanations about communicating scientific understanding in teacher-student settings on five levels of the explainee’s expertise: ReWIRED contains three layers of acts (Teaching, Explanation, Dialogue) with increased granularity (span-level). We then evaluate language models on the labeling of such acts and find that the broad range and structure of the proposed labels is hard to model for LLMs such as GPT-3.5/-4 via prompting, but a fine-tuned BERT can perform both act classification and span labeling well. Finally, we operationalize a series of quality metrics for instructional explanations in the form of a test suite, finding that they match the five expertise levels well.1

AB - For dialogues in which teachers explain difficult concepts to students, didactics research often debates which teaching strategies lead to the best learning outcome. In this paper, we test if LLMs can reliably annotate such explanation dialogues, s.t. they could assist in lesson planning and tutoring systems. We first create a new annotation scheme of teaching acts aligned with contemporary teaching models and re-annotate a dataset of conversational explanations about communicating scientific understanding in teacher-student settings on five levels of the explainee’s expertise: ReWIRED contains three layers of acts (Teaching, Explanation, Dialogue) with increased granularity (span-level). We then evaluate language models on the labeling of such acts and find that the broad range and structure of the proposed labels is hard to model for LLMs such as GPT-3.5/-4 via prompting, but a fine-tuned BERT can perform both act classification and span labeling well. Finally, we operationalize a series of quality metrics for instructional explanations in the form of a test suite, finding that they match the five expertise levels well.1

KW - Dialogue

KW - Discourse Analysis

KW - Evaluation

KW - Explanations

U2 - 10.1145/3677525.3678665

DO - 10.1145/3677525.3678665

M3 - Conference contribution

SN - 9798400710940

SP - 225

EP - 230

BT - Proceedings of the 2024 International Conference on Information Technology for Social Good

CY - New York, NY, USA

ER -

By the same author(s)