An Integrated Framework for TFT-LCD Quality Prediction Based on L1 Regularized Feature Selection and SSA-Optimized MLP-XGBoost Ensemble

Research Article
Open access

An Integrated Framework for TFT-LCD Quality Prediction Based on L1 Regularized Feature Selection and SSA-Optimized MLP-XGBoost Ensemble

Lan Jiang 1*
  • 1 Institute of Economics and Management, Tsinghua University, Beijing, China    
  • *corresponding author slm0718m@163.com
Published on 24 January 2025 | https://doi.org/10.54254/2755-2721/2025.20587
ACE Vol.116
ISSN (Print): 2755-2721
ISSN (Online): 2755-273X
ISBN (Print): 978-1-83558-791-1
ISBN (Online): 978-1-83558-792-8

Abstract

With the rapid development of display technology, accurate quality prediction of TFT-LCD is crucial because its performance directly affects the user experience. Traditional quality prediction methods rely on manual inspection or rule-based methods, facing the challenges of low efficiency, strong subjectivity, and lack of robustness to changes in production environment. To address these limitations, this paper proposes a novel regression model based on L1 regularized feature selection, Salp Swarm algorithm (SSA) optimization, and MLP-XGBoost integration. L1 regularization is used to automatically select key features from high-dimensional industrial data, reduce redundancy and enhance model generalization. Subsequently, SSA optimizes the hyperparameters of MLP and XGBoost, enabling effective exploration of the parameter space and thus improving model performance. The integrated MLP-XGBoost model combines the nonlinear feature extraction capability of MLP with the regression accuracy of XGBoost. Experimental results show that the proposed model outperforms traditional models in various indicators. Compared with the unoptimized MLP-XGBoost, the model achieves a significant reduction in prediction error, verifying its effectiveness in solving TFT-LCD quality prediction.

Keywords:

TFT-LCD Quality Prediction, L1 Regularization, Salp Swarm Algorithm, MLP-XGBoost Ensemble

Jiang,L. (2025). An Integrated Framework for TFT-LCD Quality Prediction Based on L1 Regularized Feature Selection and SSA-Optimized MLP-XGBoost Ensemble. Applied and Computational Engineering,116,154-160.
Export citation

1. Introduction

With the rapid development of display technology, thin-film transistor liquid crystal display (TFT-LCD) has become a mainstream display technology and has been widely used in smart phones, TVs, laptops and other devices[1]. The quality of its display effect directly affects the user experience of terminal devices, and the quality of TFT-LCD is affected by a variety of production processes and material characteristics[2]. Traditional quality prediction methods usually rely on manual experience or rule-based detection methods. Manual detection is inefficient and subjective, and is easily affected by the experience and status of the inspector; traditional rule-based models lack robustness to changes in the production environment and are difficult to adapt to complex and changing industrial scenarios. With the accumulation of industrial data and the improvement of computing power, intelligent prediction methods based on machine learning have gradually become the main means to solve the problem of TFT-LCD quality prediction. However, TFT-LCD production data usually has high-dimensional features, multivariate nonlinearity and strong correlation. It is often difficult to achieve good results by directly applying traditional machine learning models[3].

Feature selection is an important step in solving high-dimensional data problems. By selecting the most critical features for the target variable from a large number of redundant or irrelevant features, the data dimension can be effectively reduced and the performance and generalization ability of the model can be improved. Among them, the feature selection method based on L1 regularization can directly compress the weights of unimportant features to zero due to its sparsity constraint ability, thereby realizing automatic feature selection. In terms of model design, ensemble learning and deep learning methods have been widely used in quality prediction tasks in recent years[4]. Gradient boosting decision tree[5] (such as XGBoost) has become one of the mainstream models in the industry due to its modeling ability and efficiency for feature importance; multi-layer perceptron (MLP)[6] in deep learning can capture the complex nonlinear relationship of data. Combining the advantages of these two types of models, the MLP-XGBoost ensemble model was proposed to simultaneously extract deep features and model complex decision boundaries, providing a new technical approach to solving complex industrial prediction problems. However, the hyperparameter tuning of the ensemble model faces great challenges in large-scale industrial data. Traditional grid search or random search methods are inefficient and difficult to meet actual needs. In recent years, optimization algorithms based on swarm intelligence, such as genetic algorithm (GA) and particle swarm optimization (PSO), have shown great potential in model optimization. Among them, the Salp Swarm Algorithm (SSA) [7] has gradually become an effective tool for hyperparameter optimization due to its strong global search capability and fast convergence characteristics.

2. Method

In order to better predict the quality of TFT-LCD, this paper proposes a model framework based on L1 regularized feature selection, SSA optimization and MLP-XGBoost fusion, which can be divided into three key modules: feature selection module, hyperparameter optimization module and regression prediction module. The algorithm flow is shown in Figure 1.

/word/media/image1.png

Figure 1: Algorithm flow

2.1. Embedded method based on L1 regularization

In TFT-LCD quality prediction, a large amount of data is usually involved, and these features may have significant redundancy or correlation. High-dimensional feature data will not only increase the computational complexity of model training, but may also reduce the generalization ability of the model, leading to overfitting problems. Therefore, feature selection is a key step in building an efficient regression model. L1 regularization is an embedded feature selection method that removes irrelevant features through sparse constraints while retaining the features that have the greatest impact on the target variable. Taking linear regression as an example, its objective function can be expressed as follows after adding the L1 norm constraint:

\( {min_{w}}(\frac{1}{2n}{\sum _{i=1}^{n}{(y_{i}}-{X_{i}}w)^{2}}+λ||w|{|_{1}})\ \ \ (1) \)

Where: n is the number of samples; \( {y_{i}} \) is the target variable of the i-th sample (such as the brightness value of TFT-LCD); \( {X_{i}}∈{R^{m}} \) is the feature vector of the i-th sample (containing m features); \( w∈{R^{m}} \) is the model parameter vector; \( λ \gt 0 \) is the regularization strength, which is used to control sparsity. By adjusting the regularization strength \( λ \) , the sparsity and fitting ability of the model can be balanced. The feature selection process is as follows:

(1) Data standardization: In order to ensure the consistency of the scale of different features, the feature matrix X is first standardized so that the mean of each column feature is zero and the variance is one.

(2) Model training: Construct an L1 regularized linear regression model, take the target variable y as the predicted value, and minimize the objective function with L1 regularization.

(3) Weight sparsification: After training, observe the model weight vector w. For the feature \( {x_{j}} \) of \( {w_{j}}=0 \) , it is determined that it has no significant effect on the target variable and the feature is removed.

(4) Result generation: Output the features of weight \( {w_{j}}≠0 \) to generate the feature subset \( {X_{selected}} \) after dimensionality reduction.

After the above steps, the feature selection module can effectively reduce the data dimension, retain the features with the highest correlation with the target variable, and provide a concise and efficient data foundation for subsequent modeling.

2.2. Global optimization strategy based on SSA

We selects MLP (multi-layer perceptron) and XGBoost (extreme gradient boosting) to construct a regression model for TFT-LCD quality prediction, mainly based on the complementarity of the two in feature processing and regression performance. MLP has strong nonlinear modeling capabilities through its multi-layer network structure and activation function. It can extract deep features from complex high-dimensional data and map redundant features into a more compact and discriminative representation. On the other hand, XGBoost is based on gradient boosting decision trees, is good at processing complex nonlinear decision boundaries, and has high robustness to noise and outliers.

The hyperparameters of MLP and XGBoost have an important impact on model performance. However, traditional grid search and random search methods are inefficient in searching in high-dimensional parameter spaces, and finding the global optimal solution is difficult. This paper uses SSA (Salp Swarm Algorithm) to optimize the model's hyperparameters. SSA is an optimization algorithm based on swarm intelligence, which completes the global search by simulating the movement behavior of willow leaf fish. The algorithm realizes the dynamic optimization of the objective function through the collaboration of individuals in the population.

(1) Population initialization: Let the population size be N, and each position represents a set of hyperparameter vectors \( {x_{i}}=[{p_{1}},{p_{2}},…,{p_{d}}] \) , where d is the hyperparameter dimension. The initial position matrix is:

\( X=[\begin{matrix}{x_{1,1}} & {x_{1,2}} & ⋯ & {x_{1,d}} \\ {x_{2,1}} & {x_{2,2}} & ⋯ & {x_{2,d}} \\ ⋮ & ⋮ & ⋱ & ⋮ \\ {x_{N,1}} & {x_{N,2}} & ⋯ & {x_{N,d}} \\ \end{matrix}]\ \ \ (2) \)

(2) Individual update rule:

The individual with the best objective function value is selected as the leader. The leader position update formula is:

\( x_{leader,j}^{(t+1)}=x_{leader,j}^{(t)}+{c_{1}}{r_{1}}({x_{target,j}}-x_{leader,j}^{(t)})\ \ \ (3) \)

Where \( {c_{1}} \) is the scaling factor, \( {r_{1}}∪(0,1) \) is a random number. The positions of other individuals are dynamically adjusted based on the leader:

\( x_{i,j}^{(t+1)}=\frac{x_{i-1,j}^{(t)}+x_{i,j}^{(t)}}{2},i≥2\ \ \ (4) \)

(3) Constraint processing: If the individual exceeds the hyperparameter range, it will be reinitialized.

(4) Stop condition: Optimization is terminated when the maximum number of iterations is reached or the objective function value converges.

2.3. MLP-XGBoost integrated modeling

The MLP-XGBoost integrated model combines the nonlinear feature extraction capability of deep learning and the regression prediction capability of XGBoost. Its overall structure is divided into two stages:

(1) MLP feature extraction: MLP consists of an input layer, a hidden layer, and an output layer, and generates high-dimensional feature representations through multiple layers of nonlinear transformations. The hidden layer output formula is:

\( {h^{(l)}}=f({W^{(l)}}{h^{(l-1)}}+{b^{(l)}})\ \ \ (5) \)

Where \( {W^{(l)}} \) and \( {b^{(l)}} \) are the weight matrix and bias vector, respectively, and \( f(∙) \) is the activation function.

(2) XGBoost regression prediction: The features generated by MLP are input into XGBoost, and the target variable is predicted using the gradient boosting decision tree. The objective function of XGBoost is:

\( L=\sum _{i=1}^{n} l({\hat{y}_{i}},{y_{i}})+\sum _{k=1}^{K} Ω({f_{k}})\ \ \ (6) \)

Where \( l(∙) \) is the loss function, \( Ω({f_{k}}) \) is the regularization term of the tree.

The integration process can be summarized as: using MLP to perform deep feature extraction on feature-selected data; inputting the extracted features into XGBoost to model the complex relationship between features and target variables; and finally outputting the predicted value \( \hat{y} \) .

3. Experimental analysis

3.1. Dataset and Experimental Settings

This study experimentally verifies the performance of the proposed MLP-XGBoost integrated model based on L1 feature selection and SSA optimization in the TFT-LCD quality regression prediction task. The experiment uses TFT-LCD production data collected from industrial production lines. The data set contains thousands of sample records, each of which corresponds to the process parameters and quality indicators of a TFT-LCD product, totaling 5000 data, 50 original features, and the target variable is TFT-LCD brightness (unit: cd/m²). The data set is randomly divided into training set, validation set and test set in a ratio of 7:2:1 for model training, hyperparameter tuning and performance testing. In order to ensure the fairness of the experiment, all features are standardized so that the mean of each feature is 0 and the variance is 1. In addition, the brightness value of the target variable is normalized so that its distribution range is scaled to [0,1]. In order to comprehensively evaluate the regression performance of the model, this paper uses the following three common indicators:

Mean Squared Error (MSE):

\( MSE=\frac{1}{n}\sum _{i=1}^{n}{(\hat{{y_{i}}}-{y_{i}})^{2}}\ \ \ (7) \)

Root Mean Squared Error, (RMSE):

\( RMSE=\sqrt[]{MSE}\ \ \ (8) \)

Coefficient of determination(R2):

\( {R^{2}}=1-\frac{\sum _{i=1}^{n}{(\hat{{y_{i}}}-{y_{i}})^{2}}}{\sum _{i=1}^{n}{({y_{i}}-\bar{y})^{2}}}\ \ \ (9) \)

3.2. Comparison model

In order to verify the effectiveness of the proposed method, the following classic models and their variants are selected for comparison. (1)Linear Regression (LR): A simple linear regression method as a baseline model. (2)Support Vector Regression (SVR): A kernel function-based regression method for dealing with nonlinear problems. (3)Random Forest Regression (RF): An integrated model based on decision trees that is good at dealing with high-dimensional nonlinear data. (4)XGBoost: An efficient gradient boosting decision tree model widely used in industrial prediction tasks. (5)MLP: A multi-layer perceptron model in deep learning, used to extract complex nonlinear features of data. (6)MLP-XGBoost (unoptimized): An integrated model combining MLP and XGBoost, but without SSA hyperparameter optimization. The following are the performance results of different models on the test set, as shown in Table1.

Table 1: Performance comparison of different models

Model

MSE

RMSE

R2

LR

0.0821

0.2865

0.7412

SVR

0.0603

0.2455

0.8124

RF

0.0457

0.2139

0.8631

XGBoost

0.0382

0.1955

0.8873

MLP

0.0425

0.2061

0.8712

MLP-XGBoost(unoptimized)

0.0354

0.1880

0.8976

Ours

0.0293

0.1712

0.9221

Traditional linear regression and SVR methods have limitations in nonlinear feature modeling and poor prediction performance; Random forest and XGBoost models have better performance than traditional methods due to their good modeling ability for nonlinear features; MLP can capture complex nonlinear relationships, but due to the limitations of hyperparameter settings, its performance is not as good as that of the integrated model; The unoptimized MLP-XGBoost integrated model is worse than either MLP or XGBoost alone, indicating that the combination of the two has a synergistic gain effect; The model in this paper performs best in all evaluation indicators, verifying the contribution of L1 feature selection and SSA optimization to performance improvement.

3.3. SSA hyperparameter optimization effect

In order to verify the effect of SSA on hyperparameter optimization, the performance of the model before and after SSA optimization is compared, and the results are as follows, as shown in Table 2.

Table 2: Performance of hyperparameter optimization

Optimization status

MSE

RMSE

R2

Unoptimized

0.0354

0.1880

0.8976

SSA optimization

0.0293

0.1712

0.9221

The results show that SSA optimization significantly reduces the prediction error of the model and improves the model's fitting ability. Through experimental analysis, the MLP-XGBoost integrated model based on L1 feature selection and SSA optimization proposed in this paper performs well in TFT-LCD quality regression prediction. The experimental results verify the effectiveness of L1 feature selection for dimensionality reduction and feature extraction, the significant contribution of SSA to hyperparameter optimization, and the synergy of MLP and XGBoost in complex data modeling.

4. Conclusion

This study proposes a new framework for TFT-LCD quality regression prediction by integrating L1 feature selection, SSA-based hyperparameter optimization, and MLP-XGBoost ensemble modeling. Experimental results on real-world TFT-LCD production data show that the proposed model outperforms traditional machine learning methods and unoptimized models, achieving superior accuracy and reducing prediction errors.


References

[1]. Su, Y. C., Hung, M. H., Cheng, F. T., & Chen, Y. T. (2006). A processing quality prognostics scheme for plasma sputtering in TFT-LCD manufacturing. IEEE Transactions on semiconductor manufacturing, 19(2), 183-194.

[2]. Su, Y. C., Cheng, F. T., Huang, G. W., Hung, M. H., & Yang, T. (2004, November). A quality prognostics scheme for semiconductor and TFT-LCD manufacturing processes. In 30th Annual Conference of IEEE Industrial Electronics Society, 2004. IECON 2004 (Vol. 2, pp. 1972-1977). IEEE.

[3]. Jen, C. H., Fan, S. K. S., & Lin, Y. Y. (2022). Data-driven virtual metrology and retraining systems for color filter processes of TFT-LCD manufacturing. IEEE Transactions on Instrumentation and Measurement, 71, 1-12.

[4]. Shih, M. S., Chen, J. C., Chen, T. L., & Hsu, C. L. (2024). Two-phase cost-sensitive-learning-based framework on customer-side quality inspection for TFT-LCD industry. Journal of Intelligent Manufacturing, 1-17.

[5]. Chen, T., & Guestrin, C. (2016, August). Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining (pp. 785-794).

[6]. Taud, H., & Mas, J. F. (2018). Multilayer perceptron (MLP). Geomatic approaches for modeling land change scenarios, 451-455.

[7]. Hereford, J. M. (2010, July). Analysis of a new swarm search algorithm based on trophallaxis. In IEEE Congress on Evolutionary Computation (pp. 1-8). IEEE.


Cite this article

Jiang,L. (2025). An Integrated Framework for TFT-LCD Quality Prediction Based on L1 Regularized Feature Selection and SSA-Optimized MLP-XGBoost Ensemble. Applied and Computational Engineering,116,154-160.

Data availability

The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.

Disclaimer/Publisher's Note

The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

About volume

Volume title: Proceedings of the 5th International Conference on Signal Processing and Machine Learning

ISBN:978-1-83558-791-1(Print) / 978-1-83558-792-8(Online)
Editor:Stavros Shiaeles
Conference website: https://2025.confspml.org/
Conference date: 12 January 2025
Series: Applied and Computational Engineering
Volume number: Vol.116
ISSN:2755-2721(Print) / 2755-273X(Online)

© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license. Authors who publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See Open access policy for details).

References

[1]. Su, Y. C., Hung, M. H., Cheng, F. T., & Chen, Y. T. (2006). A processing quality prognostics scheme for plasma sputtering in TFT-LCD manufacturing. IEEE Transactions on semiconductor manufacturing, 19(2), 183-194.

[2]. Su, Y. C., Cheng, F. T., Huang, G. W., Hung, M. H., & Yang, T. (2004, November). A quality prognostics scheme for semiconductor and TFT-LCD manufacturing processes. In 30th Annual Conference of IEEE Industrial Electronics Society, 2004. IECON 2004 (Vol. 2, pp. 1972-1977). IEEE.

[3]. Jen, C. H., Fan, S. K. S., & Lin, Y. Y. (2022). Data-driven virtual metrology and retraining systems for color filter processes of TFT-LCD manufacturing. IEEE Transactions on Instrumentation and Measurement, 71, 1-12.

[4]. Shih, M. S., Chen, J. C., Chen, T. L., & Hsu, C. L. (2024). Two-phase cost-sensitive-learning-based framework on customer-side quality inspection for TFT-LCD industry. Journal of Intelligent Manufacturing, 1-17.

[5]. Chen, T., & Guestrin, C. (2016, August). Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining (pp. 785-794).

[6]. Taud, H., & Mas, J. F. (2018). Multilayer perceptron (MLP). Geomatic approaches for modeling land change scenarios, 451-455.

[7]. Hereford, J. M. (2010, July). Analysis of a new swarm search algorithm based on trophallaxis. In IEEE Congress on Evolutionary Computation (pp. 1-8). IEEE.