Difference between revisions of "Other NLP Applications of Algorithms in Education"

From Penn Center for Learning Analytics Wiki
Jump to navigation Jump to search
m
 
(6 intermediate revisions by 3 users not shown)
Line 1: Line 1:
Naismith et al. (2018) [[http://d-scholarship.pitt.edu/40665/1/EDM2018_paper_37.pdf pdf]]
Naismith et al. (2018) [http://d-scholarship.pitt.edu/40665/1/EDM2018_paper_37.pdf pdf]


* a model that measures L2 learners’ lexical sophistication with the frequency list based on the native speaker corpora
* a model that measures L2 learners’ lexical sophistication with the frequency list based on the native speaker corpora
Line 7: Line 7:




Samei et al. (2015) [[https://files.eric.ed.gov/fulltext/ED560879.pdf pdf]]
Samei et al. (2015) [https://files.eric.ed.gov/fulltext/ED560879.pdf pdf]


* Models predicting classroom discourse properties (e.g. authenticity and uptake)
* Models predicting classroom discourse properties (e.g. authenticity and uptake)
* Model trained on urban students (authenticity: 0.62, uptake: 0.60) performed with similar accuracy when tested on non-urban students (authenticity: 0.62, uptake: 0.62)
* Model trained on urban students (authenticity: 0.62, uptake: 0.60) performed with similar accuracy when tested on non-urban students (authenticity: 0.62, uptake: 0.62)
* Model trained on non-urban (authenticity: 0.61, uptake: 0.59) performed with similar accuracy when tested on urban students (authenticity: 0.60, uptake: 0.63)
* Model trained on non-urban (authenticity: 0.61, uptake: 0.59) performed with similar accuracy when tested on urban students (authenticity: 0.60, uptake: 0.63)
Sha et al. (2021) [https://angusglchen.github.io/files/AIED2021_Lele_Assessing.pdf pdf]
* Models predicting a MOOC discussion forum post is content-relevant or content-irrelevant
* MOOCs taught in English
* Some algorithms achieved ABROCA under 0.01 for female students versus male students, but other algorithms (Naive Bayes) had ABROCA as high as 0.06
* ABROCA varied from 0.03 to 0.08 for non-native speakers of English versus native speakers
* Balancing the size of each group in the training set reduced ABROCA values
Sha et al. (2022) [https://ieeexplore.ieee.org/abstract/document/9849852]
* Predicting forum post relevance to course in Moodle data (neural network)
* A range of over-sampling methods tested
* Regardless of over-sampling method used, forum post relevance performance was moderately better for females.
Zhang et al.(2023) [https://learninganalytics.upenn.edu/ryanbaker/ISLS23_annotation%20detector_short_submit.pdf pdf]
* Models developed to detect attributes of student feedback for other students’ mathematics solutions, reflecting the presence of three constructs:1) commenting on the process, 2) commenting on the answer, and 3) relating to self.
* Models have approximately equal performance for males and females and for African American, Hispanic/Latinx, and White students.

Latest revision as of 20:10, 28 June 2023

Naismith et al. (2018) pdf

  • a model that measures L2 learners’ lexical sophistication with the frequency list based on the native speaker corpora
  • Arabic-speaking learners are rated systematically lower across all levels of English proficiency than speakers of Chinese, Japanese, Korean, and Spanish.
  • Level 5 Arabic-speaking learners are unfairly evaluated to have similar level of lexical sophistication as Level 4 learners from China, Japan, Korean and Spain .
  • When used on ETS corpus, “high”-labeled essays by Japanese-speaking learners are rated significantly lower in lexical sophistication than Arabic, Japanese, Korean and Spanish peers.


Samei et al. (2015) pdf

  • Models predicting classroom discourse properties (e.g. authenticity and uptake)
  • Model trained on urban students (authenticity: 0.62, uptake: 0.60) performed with similar accuracy when tested on non-urban students (authenticity: 0.62, uptake: 0.62)
  • Model trained on non-urban (authenticity: 0.61, uptake: 0.59) performed with similar accuracy when tested on urban students (authenticity: 0.60, uptake: 0.63)


Sha et al. (2021) pdf

  • Models predicting a MOOC discussion forum post is content-relevant or content-irrelevant
  • MOOCs taught in English
  • Some algorithms achieved ABROCA under 0.01 for female students versus male students, but other algorithms (Naive Bayes) had ABROCA as high as 0.06
  • ABROCA varied from 0.03 to 0.08 for non-native speakers of English versus native speakers
  • Balancing the size of each group in the training set reduced ABROCA values


Sha et al. (2022) [1]

  • Predicting forum post relevance to course in Moodle data (neural network)
  • A range of over-sampling methods tested
  • Regardless of over-sampling method used, forum post relevance performance was moderately better for females.


Zhang et al.(2023) pdf

  • Models developed to detect attributes of student feedback for other students’ mathematics solutions, reflecting the presence of three constructs:1) commenting on the process, 2) commenting on the answer, and 3) relating to self.
  • Models have approximately equal performance for males and females and for African American, Hispanic/Latinx, and White students.