Difference between revisions of "Socioeconomic Status"

From Penn Center for Learning Analytics Wiki
Jump to navigation Jump to search
(Added Litman et al. (2021))
Line 32: Line 32:
* Automated essay scoring models inferring text evidence usage
* Automated essay scoring models inferring text evidence usage
* All algorithms studied have less than 1% of error explained by whether student receives free/reduced price lunch
* All algorithms studied have less than 1% of error explained by whether student receives free/reduced price lunch
Queiroga et al. (2022) [https://doi.org/10.3390/info13090401 pdf]
* Models predicting secondary school students at risk of failure or dropping out
* Models achieved high performances with an AUC higher than 0.90 and F1 higher than 0.88
* Equal performance for both genders students who participated in the educational system and completed their studies
* Students engaged in social welfare programs results in fewer problems in education
* The F1 score for social welfare program is 0.80, while for no social welfare program is 0, indicating bias toward social welfare program students
* The two most important features to predict students at risk in secondary school in Uruguay early are First-year primary school zones (rural or urban) and sixth-year assessment-based grouping

Revision as of 16:41, 1 June 2023

Yudelson et al. (2014) pdf

  • Models discovering generalizable sub-populations of students across different schools to predict students' learning with Carnegie Learning’s Cognitive Tutor (CLCT)
  • Models trained on schools with a high proportion of low-SES student performed worse than those trained with medium or low proportion
  • Models trained on schools with low, medium proportion of SES students performed similarly well for schools with high proportions of low-SES students


Yu et al. (2020) pdf

  • Models predicting undergraduate course grades and average GPA
  • Students from low-income households were inaccurately predicted to perform worse for both short-term (final course grade) and long-term (GPA)
  • Fairness of model improved if it included only clickstream and survey data


Yu et al. (2021) pdf

  • Models predicting college dropout for students in residential and fully online program
  • Whether the socio-demographic information was included or not, the model showed worse accuracy and true negative rates for residential students with greater financial needs
  • The model showed better recall for students with greater financial needs, especially for those studying in person


Kung & Yu (2020) pdf

  • Predicting course grades and later GPA at public U.S. university
  • Equal performance for low-income and upper-income students in course grade prediction for several algorithms and metrics
  • Worse performance on independence for low-income students than high-income students in later GPA prediction for four of five algorithms; one algorithm had worse separation and two algorithms had worse sufficiency


Litman et al. (2021) html

  • Automated essay scoring models inferring text evidence usage
  • All algorithms studied have less than 1% of error explained by whether student receives free/reduced price lunch


Queiroga et al. (2022) pdf

  • Models predicting secondary school students at risk of failure or dropping out
  • Models achieved high performances with an AUC higher than 0.90 and F1 higher than 0.88
  • Equal performance for both genders students who participated in the educational system and completed their studies
  • Students engaged in social welfare programs results in fewer problems in education
  • The F1 score for social welfare program is 0.80, while for no social welfare program is 0, indicating bias toward social welfare program students
  • The two most important features to predict students at risk in secondary school in Uruguay early are First-year primary school zones (rural or urban) and sixth-year assessment-based grouping