| Literature DB >> 24983000 |
Iván Gómez1, Sergio A Cannas2, Omar Osenda2, José M Jerez1, Leonardo Franco1.
Abstract
We introduce in this work an extension for the generalization complexity measure to continuous input data. The measure, originally defined in Boolean space, quantifies the complexity of data in relationship to the prediction accuracy that can be expected when using a supervised classifier like a neural network, SVM, and so forth. We first extend the original measure for its use with continuous functions to later on, using an approach based on the use of the set of Walsh functions, consider the case of having a finite number of data points (inputs/outputs pairs), that is, usually the practical case. Using a set of trigonometric functions a model that gives a relationship between the size of the hidden layer of a neural network and the complexity is constructed. Finally, we demonstrate the application of the introduced complexity measure, by using the generated model, to the problem of estimating an adequate neural network architecture for real-world data sets.Entities:
Mesh:
Year: 2014 PMID: 24983000 PMCID: PMC4005069 DOI: 10.1155/2014/815156
Source DB: PubMed Journal: ScientificWorldJournal ISSN: 1537-744X
Figure 1A comparison of the continuous and discrete versions of the first-order term generalization complexities for the D = 1 and D = 2 set of functions from (15) using N = 100. The discrete GC C 1 is computed over a grid with spacing h and so the continuous input complexity C 1′ is plotted multiplied by h.
Figure 2Comparison of the second terms of the complexities in their continuous and discrete versions, h𝒞 2′ and 𝒞 2, for the two-dimensional set of trigonometric functions from (15) as a function of n/n max.
Figure 3Approximation of two Walsh functions (W 1(x) and W 2(x)) using hyperbolic tangent functions combined with constant ones (G 1(x) and G 2(x)).
Figure 4The model constructed for N = 4 input dimensions and its application to estimate an adequate size neural network for three test benchmark functions. The continuous line represents the model estimated from a set of trigonometric functions of variable complexity and the blue dashed line indicates the size estimated by the model (using the Y-axis values), while the red dashed line is the best size obtained from exhaustive numerical simulations.
Results of the application of the model constructed by approximating the CGC of 10 benchmark data sets from the UCI repository.
| ID | Data set |
CGC |
|
|
|---|---|---|---|---|
|
| Balance Scale2,3,4,5 | 0.001 | 6.08 | 4 |
|
| Ecoli2,4,6,7 | 0.03 | 7.1 | 10 |
|
| Blood1,2,3,4 | 0.06 | 8.47 | 4 |
|
| TicTacToe1,2,5,8 | 0.1 | 9.84 | 4 |
|
| Liver Disorders1,2,5,6 | 0.11 | 10.8 | 4 |
|
| Mammografic2,3,4,5 | 0.18 | 14.5 | 13 |
|
| Hayes-Roth2,3,4,5 | 0.22 | 16.9 | 4 |
|
| Spectf2,3,5,7 | 0.26 | 17.8 | 23 |
|
| Vertebral Column1,2,3,4 | 0.36 | 26.6 | 24 |
|
| Haberman1,2,3,4 | 0.43 | 31.8 | 26 |
The table shows the identifier of the function, the name of the data set with superscripts indicating the 4 input used variables, the estimated CGC, the estimated size of an adequate neural network according to the model (N ), and the best architecture found from intensive simulations (N ).