Three different approaches for robust fuzzy clusterwise regression are reviewed. They are all based on the simultaneous application of trimming and constraints. The first one follows from the joint modeling of the response and explanatory variables through a normal component fitted in each cluster. The second one assumes normally distributed error terms conditional on the explanatory variables while the third approach is an extension of the Cluster Weighted Model. A fixed proportion of “most outlying” observations are trimmed. The use of appropriate constraints turns these problem into mathematically well-defined ones and, additionally, serves to avoid the detection of non-interesting or “spurious” linear clusters. The third proposal is specially appealing because it is able to protect us against outliers in the explanatory variables which may act as “bad leverage” points. Feasible and practical algorithms are outlined. Their performances, in terms of robustness, are illustrated in some simple simulated examples.
García-Escudero, L., Gordaliza, A., Greselin, F., Mayo-Iscar, A. (2018). Robust Approaches for Fuzzy Clusterwise Regression Based on Trimming and Constraints. In E. Gil, E. Gil, J. Gil, M.A. Gil (a cura di), Studies in Systems, Decision and Control (pp. 155-164). Springer International Publishing [10.1007/978-3-319-73848-2_15].
Robust Approaches for Fuzzy Clusterwise Regression Based on Trimming and Constraints
Greselin, F
;
2018
Abstract
Three different approaches for robust fuzzy clusterwise regression are reviewed. They are all based on the simultaneous application of trimming and constraints. The first one follows from the joint modeling of the response and explanatory variables through a normal component fitted in each cluster. The second one assumes normally distributed error terms conditional on the explanatory variables while the third approach is an extension of the Cluster Weighted Model. A fixed proportion of “most outlying” observations are trimmed. The use of appropriate constraints turns these problem into mathematically well-defined ones and, additionally, serves to avoid the detection of non-interesting or “spurious” linear clusters. The third proposal is specially appealing because it is able to protect us against outliers in the explanatory variables which may act as “bad leverage” points. Feasible and practical algorithms are outlined. Their performances, in terms of robustness, are illustrated in some simple simulated examples.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.