Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/60816
Citations
Scopus Web of Science® Altmetric
?
?
Type: Journal article
Title: Data splitting for artificial neural networks using SOM-based stratified sampling
Author: May, R.
Maier, H.
Dandy, G.
Citation: Neural Networks, 2010; 23(2):283-294
Publisher: Pergamon-Elsevier Science Ltd
Issue Date: 2010
ISSN: 0893-6080
1879-2782
Statement of
Responsibility: 
R. J. May, H. R. Maier and G. C. Dandy
Abstract: Data splitting is an important consideration during artificial neural network (ANN) development where hold-out cross-validation is commonly employed to ensure generalization. Even for a moderate sample size, the sampling methodology used for data splitting can have a significant effect on the quality of the subsets used for training, testing and validating an ANN. Poor data splitting can result in inaccurate and highly variable model performance; however, the choice of sampling methodology is rarely given due consideration by ANN modellers. Increased confidence in the sampling is of paramount importance, since the hold-out sampling is generally performed only once during ANN development. This paper considers the variability in the quality of subsets that are obtained using different data splitting approaches. A novel approach to stratified sampling, based on Neyman sampling of the self-organizing map (SOM), is developed, with several guidelines identified for setting the SOM size and sample allocation in order to minimize the bias and variance in the datasets. Using an example ANN function approximation task, the SOM-based approach is evaluated in comparison to random sampling, DUPLEX, systematic stratified sampling, and trial-and-error sampling to minimize the statistical differences between data sets. Of these approaches, DUPLEX is found to provide benchmark performance with good model performance, with no variability. The results show that the SOM-based approach also reliably generates high-quality samples and can therefore be used with greater confidence than other approaches, especially in the case of non-uniform datasets, with the benefit of scalability to perform data splitting on large datasets.
Keywords: Multivariate Analysis
Cluster Analysis
Models, Statistical
Reproducibility of Results
Learning
Algorithms
Databases, Factual
Databases as Topic
Neural Networks, Computer
Rights: © 2009 Elsevier
DOI: 10.1016/j.neunet.2009.11.009
Published version: http://dx.doi.org/10.1016/j.neunet.2009.11.009
Appears in Collections:Aurora harvest
Civil and Environmental Engineering publications
Environment Institute publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.