Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/60816
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorMay, R.-
dc.contributor.authorMaier, H.-
dc.contributor.authorDandy, G.-
dc.date.issued2010-
dc.identifier.citationNeural Networks, 2010; 23(2):283-294-
dc.identifier.issn0893-6080-
dc.identifier.issn1879-2782-
dc.identifier.urihttp://hdl.handle.net/2440/60816-
dc.description.abstractData splitting is an important consideration during artificial neural network (ANN) development where hold-out cross-validation is commonly employed to ensure generalization. Even for a moderate sample size, the sampling methodology used for data splitting can have a significant effect on the quality of the subsets used for training, testing and validating an ANN. Poor data splitting can result in inaccurate and highly variable model performance; however, the choice of sampling methodology is rarely given due consideration by ANN modellers. Increased confidence in the sampling is of paramount importance, since the hold-out sampling is generally performed only once during ANN development. This paper considers the variability in the quality of subsets that are obtained using different data splitting approaches. A novel approach to stratified sampling, based on Neyman sampling of the self-organizing map (SOM), is developed, with several guidelines identified for setting the SOM size and sample allocation in order to minimize the bias and variance in the datasets. Using an example ANN function approximation task, the SOM-based approach is evaluated in comparison to random sampling, DUPLEX, systematic stratified sampling, and trial-and-error sampling to minimize the statistical differences between data sets. Of these approaches, DUPLEX is found to provide benchmark performance with good model performance, with no variability. The results show that the SOM-based approach also reliably generates high-quality samples and can therefore be used with greater confidence than other approaches, especially in the case of non-uniform datasets, with the benefit of scalability to perform data splitting on large datasets.-
dc.description.statementofresponsibilityR. J. May, H. R. Maier and G. C. Dandy-
dc.language.isoen-
dc.publisherPergamon-Elsevier Science Ltd-
dc.rights© 2009 Elsevier-
dc.source.urihttp://dx.doi.org/10.1016/j.neunet.2009.11.009-
dc.subjectMultivariate Analysis-
dc.subjectCluster Analysis-
dc.subjectModels, Statistical-
dc.subjectReproducibility of Results-
dc.subjectLearning-
dc.subjectAlgorithms-
dc.subjectDatabases, Factual-
dc.subjectDatabases as Topic-
dc.subjectNeural Networks, Computer-
dc.titleData splitting for artificial neural networks using SOM-based stratified sampling-
dc.typeJournal article-
dc.identifier.doi10.1016/j.neunet.2009.11.009-
pubs.publication-statusPublished-
dc.identifier.orcidMaier, H. [0000-0002-0277-6887]-
dc.identifier.orcidDandy, G. [0000-0001-5846-7365]-
Appears in Collections:Aurora harvest
Civil and Environmental Engineering publications
Environment Institute publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.