High school student GPA prediction by various linear regression models

Research Article
Open access

High school student GPA prediction by various linear regression models

Weijia Zhu 1*
  • 1 Mathematics Science and Statistics Program, University of Toronto, ON, L5L 1C6, Canada    
  • *corresponding author weijia.zhu@mail.utoronto.ca
Published on 27 September 2024 | https://doi.org/10.54254/2753-8818/52/2024CH0136
TNS Vol.52
ISSN (Print): 2753-8826
ISSN (Online): 2753-8818
ISBN (Print): 978-1-83558-621-1
ISBN (Online): 978-1-83558-622-8

Abstract

Academic performance (GPA) is a significant index for high school students in North America. The research aims to develop and validate the best predictive regression model to evaluate the GPA of high school students. In addition, the prediction numeric data of GPA can correspond to specific classification data of GPA (Grade Level) to calculate and compare the models’ accuracy for predicting Grade Level. The research subjects are high school students from different backgrounds in North America, including their background, study habits, participation in extracurricular activities, etc. The experiment explores the impact of different factors on student GPA and finds that the number of absences from lectures is a key factor in predicting student GPA. Multiple linear regression analysis is used as the main model in the experiment, which may be improved by the stepwise regression methods. The generalization ability of the model is evaluated through cross-validation (CV) methods. Also, the boosting or random forest model is used to be the comparing model for predicting GPA. The experimental result shows that the multiple linear regression model has high accuracy (84%) and reliability (R^2 value is 0.95) in predicting student GPA. The conclusion of the research emphasizes the importance of predicting student GPA in high school education and the potential for guiding educational practice through data analysis. Future work will consider introducing more subjects and variables, such as different subject learning abilities, mental health, and social support, to further improve the predictive accuracy of the model.

Keywords:

GPA, linear regression models, stepwise regression, prediction

Zhu,W. (2024). High school student GPA prediction by various linear regression models. Theoretical and Natural Science,52,153-162.
Export citation

1. Introduction

For high school students in North America, there are plenty of memorable experiences and memories in the school. Some students make the most significant decisions in high school, such as becoming an occupational basketball player or singer. However, most high school students may choose to continue studying in the universities or colleges they applied to successfully. In this case, academic GPA is significant for high school students in North America because it is an important index for university admissions decisions [1]. GPA in high school can be calculated by all the course scores which are relative to homework and test scores. In this case, most high school students may hope to have a higher GPA to apply to the university. Also, students who have low GPA hope to change their performance to improve their GPA. However, most students with low GPA do not know how to increase their GPA. In this case, most students hope to understand how to get a higher GPA, and which features of students are most important for improving their GPA.

Sawyer found that GPA in high school is useful for universities to predict high school students’ academic success in the future [1]. Kobrin et al. found that the GPA in high school with a correlation of 0.36 is more significant than admission test scores (0.35) in predicting students’ first-year GPA in the university [2]. University of Miami researchers found that a point increase in high school students’ GPA may affect the increasing 12% - 14% of income in the future [3]. In addition, high school GPA is affected by plenty of factors. Philippe et al. proposed that participation in both civic and non-civic organized activities may positively affect students to increase their GPA in high school [4]. Rahafar et al. found that chronotype, gender, conscientiousness, and test may affect high school student GPA, where gender differences had the most impact on students’ GPA by using the structural equation model [5]. Warren showed that students who study in small learning communities could reduce their number of absences and improve their test scores, thereby, increasing their GPA in high school [6]. Therefore, the research plan is to figure out which feature is the most important for affecting high school student’s GPA in North America.

In addition, predicting student’s GPA by their different features without their course scores and homework performance is the main purpose of the research. In this case, the machine learning methods of the research are significant for predicting student’s GPA. Paolo et al. used a logistic multiple regression model to predict the GPA of students and found that cross-validation is important for the generalizability and overall utility of prediction models [7]. Hassan et al. used multiple regression models including linear, ridge, lasso, and LassoCV to predict the GPA of students and found that the LassoCV method is the best model with higher \( {R^{2}} \) and lower RMSE in his dataset selected based on the Community of Inquiry framework [8]. Nasiri et al. utilized regression analysis for predicting GPA and the C5.0 algorithm (a type of decision tree) to predict academic dismissal (Classification Data) [9]. Cai et al. found that gradient boosting regression and random forest models improved by the k-fold cross-validation algorithm are better than other models for predicting net ecosystem carbon exchange (NEE) [10].

Based on their ideas of models for predicting data, the experiment decides to use multiple linear regression, random forest, and boosting models improved by the k-fold cross-validation method to predict high school student’s GPA, thereby, making the comparison between these model methods to choose the best prediction model.

2. Methodology

2.1. Data source

The dataset collection is from Kharoua in Kaggle about features (Student ID, Age, Gender, Ethnicity, Parental Education, Week Study Time, Absences, Tutoring, Parental Support, Extracurricular, Sports, Music, Volunteering, GPA, and Grade Class) of 2392 high school students in North America.

2.2. Dataset preprocessing

For the details of the dataset, data visualization by histograms is necessary for every feature. In this case, a factor (Student ID), which is the ID for all the subjects, having no distribution in the histogram, being not relevant to the value of the GPA of the students (See Figure 1). Therefore, this factor (Student ID) is dropped before predicting the GPA. The other 12 factors' histograms may show that they may affect the GPA (See Figure 2, Figure 3). In addition, Grade Class (Grade Level) is a classification data of GPA, so Grade Level cannot be used to predict GPA. In this case, Grade Level is dropped in the experiment for predicting GPA. Besides, there are data errors and loss that do not exist since the dataset in Kaggle is reprocessed by Kharoua. The only problem with the dataset is that students’ GPA in this dataset are relatively low since Kharoua might collect the data from most students with bad academic performance. The boxplot is necessary for finding outliers for numeric factors but all numeric data in this dataset do not have outliers through the boxplot (See Figure 4). Also, the experiment normalizes 12 factors’ data (except Student ID, GPA, and Grade Class) in the dataset to eliminate data differences between factors.

/word/media/image1.png

Figure 1. Histogram of Student ID.

/word/media/image2.png

Figure 2. Histogram of age, gender, ethnicity, and parental education.

/word/media/image3.png

Figure 3. Histogram of extracurricular, sports, music, and volunteering.

/word/media/image4.png

Figure 4. Box plot of age, study time weekly, absences.

2.3. Variable selection

The experiment has already dropped 3 factors to have the 12 dataset factors as the variables including Age, Gender, Ethnicity, Parental Education, Week Study Time, Absences, Tutoring, Parental Support, Extracurricular, Sports, Music, and Volunteering. In this case, the experiments make variables have their logograms (See Table 1).

Table 1. Logogram of the 12 factors and target factor.

Factors

Logogram

Meaning

Age

Gender

\( {x_{1}} \)

\( {x_{2}} \)

The age of students (15 – 17 years old)

Male (0), Female (1)

Ethnicity

\( {x_{3}} \)

Caucasian (0), African American (1), Asian (2), and Other (3)

Parental Education

\( {x_{4}} \)

None (0), high school (1), some college (2), Bachelor’s (3), Higher (4).

Week Study Time

Absences

Tutoring

Parental Support

\( {x_{5}} \)

\( {x_{6}} \)

\( {x_{7}} \)

\( {x_{8}} \)

Students’ weekly study time in hours (0 to 20 hours)

The number of absences during the school year (0 to 30 times)

No (0), Yes (1)

None (0), Low (1), Moderate (2), High (3), and Very High (4)

Extracurricular

\( {x_{9}} \)

Participation: No (0), Yes (1)

Sports

Music

\( {x_{10}} \)

\( {x_{11}} \)

Participation: No (0), Yes (1)

Participation: No (0), Yes (1)

Volunteering

GPA

\( {x_{12}} \)

\( y \)

Participation: No (0), Yes (1)

Target factor: the grade value of students (range: 0 – 4)

2.4. Method introduction

The experiment utilizes a correlation matrix to determine the feature importance in the dataset. Before creating models to predict GPA, the dataset is divided into two datasets which are the model train dataset (80%) and the test dataset (20%). The experiment trains multiple linear regression models as the base model for predicting the GPA and uses the cross-validation and stepwise regression methods to improve the regression model. For the 12 factors in the model, the multilinearity check is necessary for having a good regression model. In addition, the experiment may use random forest and boosting to create models to compare with the multiple linear regression models to determine the best model for predicting students’ GPA. Finally, the experiment calculates the value of mean squared error (MSE) and \( {R^{2}} \) from every model for comparison between models. In addition, using the confusion matrix by predicting model data checks every model's accuracy.

3. Results and discussion

3.1. Descriptive analysis

The experiment describes the information of 12 factors and target variable from the original dataset, which includes name, mean, standard deviation (SD), median, trimmed mean (Trimmed), median absolute deviation (Mad), minimum (Min), maximum (Max), range, and standard error (SE) (See Table 2). In addition, the experiment uses the 12 factors' histograms to analyze their distribution.

Table 2. Factor information.

Name

Mean

SD

Median

Trimmed

Mad

Min

Max

SE

Age

16.47

1.12

16

16.46

1.48

15

18

0.02

Gender

0.51

0.5

1

0.51

0

0

1

0.01

Ethnicity

0.88

1.03

0

0.73

0

0

3

0.02

Parental Education

1.75

1

2

1.75

1.48

0

4

0.02

Study Time Weekly

9.77

5.65

9.71

9.73

6.97

0

19.98

0.12

Absences

14.54

8.47

15

14.57

10.38

0

29

0.17

Tutoring

0.3

0.46

0

0.25

0

0

1

0.01

Parental Support

2.12

1.12

2

2.14

1.48

0

4

0.02

Extracurricular

0.38

0.49

0

0.35

0

0

1

0.01

Sports

0.3

0.46

0

0.25

0

0

1

0.01

Music

0.2

0.4

0

0.12

0

0

1

0.01

Volunteering

0.16

0.36

0

0.07

0

0

1

0.02

GPA

1.91

0.92

1.89

1.9

1.07

0

4

0.03

3.2. Correlation analysis

For correlation between the factors and GPA, the correlation matrix is a good method to check which factor is the most important for predicting GPA. In this dataset, the figure shows that absences with a correlation (- 0.92) have the most impact on GPA, which is negatively correlated with students’ GPA (See Figure 5).

/word/media/image5.png

Figure 5. Correlation plot.

3.3. Model analysis

3.3.1. Multiple linear regression model 1. The experiment uses the training dataset to create a multiple linear regression model (Model 1) between 12 factors and GPA first (See Table 3).

Table 3. Summary of Model 1

Coefficients:

Estimate

Std. Error

t-value

Pr ( \( \gt |t| \) )

(intercept)

2.5214

0.0193

130.303

\( \lt 2×{10^{-16}} \)

Age

-0.0181

0.0120

-1.503

0.133

Gender

0.0089

0.0090

0.992

0.321

Ethnicity

0.0055

0.0131

0.417

0.677

Parental Education

0.0054

0.0181

0.301

0.764

Study Time Weekly

0.5841

0.0156

36.719

\( \lt 2×{10^{-16}} \)

Absences

-2.892

0.0156

-185.981

\( \lt 2×{10^{-16}} \)

Tutoring

0.2468

0.0098

25.073

\( \lt 2×{10^{-16}} \)

Parental Support

0.6002

0.016

37.494

\( \lt 2×{10^{-16}} \)

Extracurricular

0.1957

0.0093

21.034

\( \lt 2×{10^{-16}} \)

Sports

0.1937

0.0098

19.704

\( \lt 2×{10^{-16}} \)

Music

0.1430

0.0112

12.780

\( \lt 2×{10^{-16}} \)

Volunteering

-0.0088

0.0124

-0.706

0.480

3.3.2. Global validation linear model assumption. In this case, the experiment uses the global validation of the linear model assumption function to check whether the experiment needs to improve the linear regression model (See Table 4). In Table 4, the assumption of global stat and skewness is not satisfied, then it shows that some factors do not have a linear relationship with students’ GPA.

Table 4. GVLMA of Model 1

Value

p-value

Decision

Global stat

12.2420

0.0156

Assumptions not satisfied!

Skewness

7.9398

0.0048

Assumptions not satisfied!

Kurtosis

0.6388

0.4241

Assumptions acceptable.

Link function

0.4011

0.5265

Assumptions acceptable.

Heteroscedasticity

3.2623

0.0709

Assumptions acceptable.

3.3.3. Multicollinearity check of Model 1. Checking multicollinearity by using the Durbin-Watson Test (DW test), variance inflation factor (VIF), and kappa coefficient is necessary for finding some factors with autocorrelation. Therefore, there is no multicollinearity in this model because DW test value =1.9664, VIF < 2, and value of kappa < 5 (See Table 5).

Table 5. Value of VIF

Factors

VIF value

Factors

VIF value

Age

1.010

Tutoring

1.007

Gender

1.003

Parental Support

1.006

Ethnicity

1.004

Extracurricular

1.005

Parental Education

1.005

Sports

1.008

Study Time Weekly

1.004

Music

1.006

Absences

1.007

Volunteering

1.008

3.3.4. Stepwise regression of model 2. The stepwise regression method is a good idea to reduce several useless factors to create a new linear regression model (Model 2) (See Table 6). In Table 6, the stepwise regression model reduces four factors (Parental Education, Ethnicity, Volunteering, and Gender) to have a new linear regression model.

Table 6. Summary of Model 2

Coefficients:

Estimate

Std. Error

t-value

Pr ( \( \gt |t| \) )

(intercept)

2.5279

0.0167

151.447

\( \lt 2×{10^{-16}} \)

Age

-0.0180

0.0120

-1.501

0.134

Study Time Weekly

0.5844

0.0159

36.775

\( \lt 2×{10^{-16}} \)

Absences

-2.8912

0.0155

-186.403

\( \lt 2×{10^{-16}} \)

Tutoring

0.2470

0.0098

25.169

\( \lt 2×{10^{-16}} \)

Parental Support

0.6004

0.0160

37.562

\( \lt 2×{10^{-16}} \)

Extracurricular

0.1956

0.0093

21.050

\( \lt 2×{10^{-16}} \)

Sports

0.1938

0.0098

19.718

\( \lt 2×{10^{-16}} \)

Music

0.1429

0.0112

12.799

\( \lt 2×{10^{-16}} \)

3.3.5. Simple linear regression model 3. Figure 2 and Figure 3 show that Absences are the most impact factor for GPA, so it is meaningful for the experiment to create a simple linear regression model (Model 3) between Absences and GPA (See Table 7).

Table 7. Summary of Model 3

Coefficients:

Estimate

Std. Error

t-value

Pr ( \( \gt |t| \) )

(intercept)

3.3520

0.0164

204.8

\( \lt 2×{10^{-16}} \)

Absences

-2.8801

0.0283

-101.9

\( \lt 2×{10^{-16}} \)

3.3.6. Stepwise linear regression model 4 by 10-fold cross-validation. Table 3 shows that the p-value of five factors (Age, Gender, Ethnicity, Parental Education, and Volunteering) is larger than 0.05, which means that they are not significant for the model. In this case, the experiment utilizes the stepwise regression method by the 10-fold cross-validation to train a new multiple linear regression model (Model 4) (See Table 8).

Table 8. Summary of Model 4

Coefficients:

Estimate

Std. Error

t-value

Pr ( \( \gt |t| \) )

(intercept)

2.5189

0.0156

161.64

\( \lt 2×{10^{-16}} \)

Study Time Weekly

0.5848

0.0159

36.78

\( \lt 2×{10^{-16}} \)

Absences

-2.8911

0.0155

-186.34

\( \lt 2×{10^{-16}} \)

Tutoring

0.2472

0.0098

25.17

\( \lt 2×{10^{-16}} \)

Parental Support

0.5994

0.0160

37.52

\( \lt 2×{10^{-16}} \)

Extracurricular

0.1961

0.0093

21.10

\( \lt 2×{10^{-16}} \)

Sports

0.1946

0.0098

19.83

\( \lt 2×{10^{-16}} \)

Music

0.1429

0.0112

12.80

\( \lt 2×{10^{-16}} \)

3.3.7. Random forest model 5 and 6. Except for the multiple linear regression model, the experiment chooses to use the random forest method to train a new model (Model 5) with all 12 factors for predicting GPA. In addition, Model 4 shows that there are several factors (Age, Gender, Ethnicity, Parental Education, and Volunteering) are dropped, so the experiment trains a relative model (Model 6) with 7 factors left by using the random forest.

3.3.8. Boosting model 7 and 8. Boosting is a good regression method for predicting GPA in this dataset, so a new boosting model (Model 7) is trained with 12 factors. In addition, the experiment trains a relative boosting model (Model 8) with 7 factors left, since the experiment dropped 5 factors in Model 4.

3.4. Comparison analysis

The experiment predicts the value of GPA by different models, thereby, calculating MSE and \( {R^{2}} \) by using the prediction data and the test dataset (See Table 9). The experiment uses the prediction of GPA value and test dataset round to the nearest single digit to calculate the accuracy of every model through the confusion matrix (See Table 9).

Table 9. Comparison of Models

Model Methods

Model

MSE

\( {R^{2}} \)

Accuracy

Multiple Linear Regression Model

Model 1

0.0386

95.6%

83.82%

Stepwise Linear Regression Model

Model 2

0.0388

95.58%

83.82%

Simple Linear Regression Model

Model 3

0.1334

84.79%

69.12%

Stepwise Linear Regression Model

Model 4

0.0388

95.58%

84.03%

Random Forest Model

Model 5

0.0518

94.09%

81.72%

Random Forest Model

Model 6

0.0494

94.36%

81.93%

Boosting Model

Model 7

0.1437

83.62%

67.23%

Boosting Model

Model 8

0.0496

94.35%

82.98%

Compared with other models, Model 4 with only 7 factors has 84.03% accuracy which is the most accurate for predicting Grade Level (See Table 9). Also, Model 1 has the lowest MSE (0.0386) and highest \( {R^{2}} \) (95.6%), slightly better than Model 4 in predicting GPA. However, Model 1 is trained by using all 12 factors, which is more complex than Model 4. In this case, the experiment chooses to use Model 4 as the best model in the research (See Function of Model 4 below). Function of Model 4 (See Table 8) (Round to three decimal places):

\( y=0.585{x_{5}}-2.891{x_{6}}+0.247{x_{7}}+0.599{x_{8}}+0.196{x_{9}}+0.195{x_{10}}+0.143{x_{11}}+2.519\ \ \ (1) \)

By using the same dataset in Kaggle, Knapp showed that the support vector machine (SVM) is the best model to predict the Grade Level (classification data of GPA), which has 74.5% accuracy. Compared with his experiment results, Model 4 with 84.03% accuracy in the experiment has increased almost 10% accuracy for predicting Grade Level.

3.5. Resident test of model 4

The experiment chooses Model 4 to be the best model in this research, so the experiment needs to use the test plot to check the model assumptions including normal independent identically distribution errors, constant error variance, absence of influential cases, linear relationship between predictors and the outcome variable, and collinearity (See Figure 6). Figure 6 shows that Model 4 successfully passes these five tests. Therefore, Model 4 is a good model that does not need to be optimized deeply.

/word/media/image6.png

Figure 6. Residual test of Model 4

4. Conclusion

The research works on predicting high school students’ GPA by the experiment above. The experiment uses three machine learning methods to train 8 regression models to find the best model for predicting GPA and Grade Level. In this case, Model 4 with 7 factors is the most simple and accurate regression model in the experiment (See the function of Model 4 in the Result). Therefore, Model 4 is useful and important for school administrators to estimate the GPA of students. Also, the experiment can let students know which should be improved to increase their GPA.

For the experiment, several places can be improved in the future. Firstly, subjects in the dataset are not large enough (only 2000+), which may decrease the accuracy of models for predicting data. In addition, plenty of features from students, which is not in this dataset, may significantly affect the GPA in high school, such as subject learning abilities, mental health, and social support. For the machine learning methods, the experiment only uses three different methods, especially for using only one method from the random forest or boosting method. In this case, the research can use more regression methods to train more models to have the best model for predicting the GPA in the future.


References

[1]. Sawyer R 2013 Beyond Correlations: Usefulness of High School GPA and Test Scores in Making College Admissions Decisions. Applied Measurement in Education, 26(2), 89-112.

[2]. Kobrin J L, et al. 2008 Validity of the SAT® for Predicting First-Year College Grade Point Average. Research Report, College Board.

[3]. Miami U O 2014 New Research Shows How Students' High School GPA Could Affect Their Income. Lab Manager.

[4]. Philippe F L, et al. 2023 Organized civic and non-civic activities as predictors of academic GPA in high school students. Applied Developmental Science, 27(2), 189-204.

[5]. Rahafar A, Maghsudloo M, Farhangnia S, Vollmer C and Randler C 2015 The role of chronotype, gender, test anxiety, and conscientiousness in academic achievement of high school students. Chronobiology International, 33(1), 1-9.

[6]. Warren J E 2016 Small Learning Communities and High School Academic Success. ProQuest Dissertations & Theses.

[7]. Paolo A M, Bonaminio G A, Durham D and Stites S W 2004 Comparison and Cross-validation of Simple and Multiple Logistic Regression Models to Predict USMLE Step 1 Performance. Teaching and Learning in Medicine, 16(1), 69-73.

[8]. Hassan Y M I, Elkorany A and Wassif K 202. Utilizing Social Clustering-Based Regression Model for Predicting Student’s GPA. IEEE Access, 10, 48948-48963.

[9]. Nasiri M, Minaei B and Vafaei F 2012 Predicting GPA and academic dismissal in LMS using educational data mining: A case mining. 6th National and 3rd International Conference of E-Learning and E-Teaching, 53-58.

[10]. Cai J, Xu K, Zhu Y, Hu F and Li L 2020 Prediction and analysis of net ecosystem carbon exchange based on gradient boosting regression and random forest. Applied Energy, 262, 114566.


Cite this article

Zhu,W. (2024). High school student GPA prediction by various linear regression models. Theoretical and Natural Science,52,153-162.

Data availability

The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.

Disclaimer/Publisher's Note

The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

About volume

Volume title: Proceedings of CONF-MPCS 2024 Workshop: Quantum Machine Learning: Bridging Quantum Physics and Computational Simulations

ISBN:978-1-83558-621-1(Print) / 978-1-83558-622-8(Online)
Editor:Anil Fernando, Marwan Omar
Conference website: https://2024.confmpcs.org/
Conference date: 9 August 2024
Series: Theoretical and Natural Science
Volume number: Vol.52
ISSN:2753-8818(Print) / 2753-8826(Online)

© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license. Authors who publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See Open access policy for details).

References

[1]. Sawyer R 2013 Beyond Correlations: Usefulness of High School GPA and Test Scores in Making College Admissions Decisions. Applied Measurement in Education, 26(2), 89-112.

[2]. Kobrin J L, et al. 2008 Validity of the SAT® for Predicting First-Year College Grade Point Average. Research Report, College Board.

[3]. Miami U O 2014 New Research Shows How Students' High School GPA Could Affect Their Income. Lab Manager.

[4]. Philippe F L, et al. 2023 Organized civic and non-civic activities as predictors of academic GPA in high school students. Applied Developmental Science, 27(2), 189-204.

[5]. Rahafar A, Maghsudloo M, Farhangnia S, Vollmer C and Randler C 2015 The role of chronotype, gender, test anxiety, and conscientiousness in academic achievement of high school students. Chronobiology International, 33(1), 1-9.

[6]. Warren J E 2016 Small Learning Communities and High School Academic Success. ProQuest Dissertations & Theses.

[7]. Paolo A M, Bonaminio G A, Durham D and Stites S W 2004 Comparison and Cross-validation of Simple and Multiple Logistic Regression Models to Predict USMLE Step 1 Performance. Teaching and Learning in Medicine, 16(1), 69-73.

[8]. Hassan Y M I, Elkorany A and Wassif K 202. Utilizing Social Clustering-Based Regression Model for Predicting Student’s GPA. IEEE Access, 10, 48948-48963.

[9]. Nasiri M, Minaei B and Vafaei F 2012 Predicting GPA and academic dismissal in LMS using educational data mining: A case mining. 6th National and 3rd International Conference of E-Learning and E-Teaching, 53-58.

[10]. Cai J, Xu K, Zhu Y, Hu F and Li L 2020 Prediction and analysis of net ecosystem carbon exchange based on gradient boosting regression and random forest. Applied Energy, 262, 114566.