Redirigiendo al acceso original de articulo en 16 segundos...
Inicio  /  Algorithms  /  Vol: 13 Par: 6 (2020)  /  Artículo
ARTÍCULO
TITULO

Sparse Logistic Regression: Comparison of Regularization and Bayesian Implementations

Mattia Zanon    
Giuliano Zambonin    
Gian Antonio Susto and Seán McLoone    

Resumen

In knowledge-based systems, besides obtaining good output prediction accuracy, it is crucial to understand the subset of input variables that have most influence on the output, with the goal of gaining deeper insight into the underlying process. These requirements call for logistic model estimation techniques that provide a sparse solution, i.e., where coefficients associated with non-important variables are set to zero. In this work we compare the performance of two methods: the first one is based on the well known Least Absolute Shrinkage and Selection Operator (LASSO) which involves regularization with an l1 l 1 norm; the second one is the Relevance Vector Machine (RVM) which is based on a Bayesian implementation of the linear logistic model. The two methods are extensively compared in this paper, on real and simulated datasets. Results show that, in general, the two approaches are comparable in terms of prediction performance. RVM outperforms the LASSO both in term of structure recovery (estimation of the correct non-zero model coefficients) and prediction accuracy when the dimensionality of the data tends to increase. However, LASSO shows comparable performance to RVM when the dimensionality of the data is much higher than number of samples that is ??>>?? p > > n .

 Artículos similares