1
|
Zhang H, Yang C, Qiao J. Emotional Neural Network Based on Improved CLPSO Algorithm For Time Series Prediction. Neural Process Lett 2021. [DOI: 10.1007/s11063-021-10672-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
2
|
Morala P, Cifuentes JA, Lillo RE, Ucar I. Towards a mathematical framework to inform neural network modelling via polynomial regression. Neural Netw 2021; 142:57-72. [PMID: 33984736 DOI: 10.1016/j.neunet.2021.04.036] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2021] [Revised: 04/20/2021] [Accepted: 04/26/2021] [Indexed: 11/18/2022]
Abstract
Even when neural networks are widely used in a large number of applications, they are still considered as black boxes and present some difficulties for dimensioning or evaluating their prediction error. This has led to an increasing interest in the overlapping area between neural networks and more traditional statistical methods, which can help overcome those problems. In this article, a mathematical framework relating neural networks and polynomial regression is explored by building an explicit expression for the coefficients of a polynomial regression from the weights of a given neural network, using a Taylor expansion approach. This is achieved for single hidden layer neural networks in regression problems. The validity of the proposed method depends on different factors like the distribution of the synaptic potentials or the chosen activation function. The performance of this method is empirically tested via simulation of synthetic data generated from polynomials to train neural networks with different structures and hyperparameters, showing that almost identical predictions can be obtained when certain conditions are met. Lastly, when learning from polynomial generated data, the proposed method produces polynomials that approximate correctly the data locally.
Collapse
Affiliation(s)
- Pablo Morala
- uc3m-Santander Big Data Institute, Universidad Carlos III de Madrid. Getafe (Madrid), Spain.
| | | | - Rosa E Lillo
- uc3m-Santander Big Data Institute, Universidad Carlos III de Madrid. Getafe (Madrid), Spain; Department of Statistics, Universidad Carlos III de Madrid. Getafe (Madrid), Spain
| | - Iñaki Ucar
- uc3m-Santander Big Data Institute, Universidad Carlos III de Madrid. Getafe (Madrid), Spain
| |
Collapse
|
3
|
Heravi AR, Abed Hodtani G. A New Correntropy-Based Conjugate Gradient Backpropagation Algorithm for Improving Training in Neural Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:6252-6263. [PMID: 29993752 DOI: 10.1109/tnnls.2018.2827778] [Citation(s) in RCA: 15] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Mean square error (MSE) is the most prominent criterion in training neural networks and has been employed in numerous learning problems. In this paper, we suggest a group of novel robust information theoretic backpropagation (BP) methods, as correntropy-based conjugate gradient BP (CCG-BP). CCG-BP algorithms converge faster than the common correntropy-based BP algorithms and have better performance than the common CG-BP algorithms based on MSE, especially in nonGaussian environments and in cases with impulsive noise or heavy-tailed distributions noise. In addition, a convergence analysis of this new type of method is particularly considered. Numerical results for several samples of function approximation, synthetic function estimation, and chaotic time series prediction illustrate that our new BP method is more robust than the MSE-based method in the sense of impulsive noise, especially when SNR is low.
Collapse
|
4
|
Nayyeri M, Sadoghi Yazdi H, Maskooki A, Rouhani M. Universal Approximation by Using the Correntropy Objective Function. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:4515-4521. [PMID: 29035228 DOI: 10.1109/tnnls.2017.2753725] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Several objective functions have been proposed in the literature to adjust the input parameters of a node in constructive networks. Furthermore, many researchers have focused on the universal approximation capability of the network based on the existing objective functions. In this brief, we use a correntropy measure based on the sigmoid kernel in the objective function to adjust the input parameters of a newly added node in a cascade network. The proposed network is shown to be capable of approximating any continuous nonlinear mapping with probability one in a compact input sample space. Thus, the convergence is guaranteed. The performance of our method was compared with that of eight different objective functions, as well as with an existing one hidden layer feedforward network on several real regression data sets with and without impulsive noise. The experimental results indicate the benefits of using a correntropy measure in reducing the root mean square error and increasing the robustness to noise.
Collapse
|
5
|
Reducing the complexity of an adaptive radial basis function network with a histogram algorithm. Neural Comput Appl 2017. [DOI: 10.1007/s00521-016-2350-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
6
|
Predicting Top-of-Atmosphere Thermal Radiance Using MERRA-2 Atmospheric Data with Deep Learning. REMOTE SENSING 2017. [DOI: 10.3390/rs9111133] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
7
|
A winner-take-all approach to emotional neural networks with universal approximation property. Inf Sci (N Y) 2016. [DOI: 10.1016/j.ins.2016.01.055] [Citation(s) in RCA: 37] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
8
|
Khosravi A, Nahavandi S, Srinivasan D, Khosravi R. Constructing Optimal Prediction Intervals by Using Neural Networks and Bootstrap Method. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2015; 26:1810-1815. [PMID: 25216487 DOI: 10.1109/tnnls.2014.2354418] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
This brief proposes an efficient technique for the construction of optimized prediction intervals (PIs) by using the bootstrap technique. The method employs an innovative PI-based cost function in the training of neural networks (NNs) used for estimation of the target variance in the bootstrap method. An optimization algorithm is developed for minimization of the cost function and adjustment of NN parameters. The performance of the optimized bootstrap method is examined for seven synthetic and real-world case studies. It is shown that application of the proposed method improves the quality of constructed PIs by more than 28% over the existing technique, leading to narrower PIs with a coverage probability greater than the nominal confidence level.
Collapse
|
9
|
Thomas P, Suhner MC. A New Multilayer Perceptron Pruning Algorithm for Classification and Regression Applications. Neural Process Lett 2014. [DOI: 10.1007/s11063-014-9366-5] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
10
|
|
11
|
Quan H, Srinivasan D, Khosravi A. Short-term load and wind power forecasting using neural network-based prediction intervals. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2014; 25:303-315. [PMID: 24807030 DOI: 10.1109/tnnls.2013.2276053] [Citation(s) in RCA: 83] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/03/2023]
Abstract
Electrical power systems are evolving from today's centralized bulk systems to more decentralized systems. Penetrations of renewable energies, such as wind and solar power, significantly increase the level of uncertainty in power systems. Accurate load forecasting becomes more complex, yet more important for management of power systems. Traditional methods for generating point forecasts of load demands cannot properly handle uncertainties in system operations. To quantify potential uncertainties associated with forecasts, this paper implements a neural network (NN)-based method for the construction of prediction intervals (PIs). A newly introduced method, called lower upper bound estimation (LUBE), is applied and extended to develop PIs using NN models. A new problem formulation is proposed, which translates the primary multiobjective problem into a constrained single-objective problem. Compared with the cost function, this new formulation is closer to the primary problem and has fewer parameters. Particle swarm optimization (PSO) integrated with the mutation operator is used to solve the problem. Electrical demands from Singapore and New South Wales (Australia), as well as wind power generation from Capital Wind Farm, are used to validate the PSO-based LUBE method. Comparative results show that the proposed method can construct higher quality PIs for load and wind power generation forecasts in a short time.
Collapse
|
12
|
Ren H, Li J, Yuan ZA, Hu JY, Yu Y, Lu YH. The development of a combined mathematical model to forecast the incidence of hepatitis E in Shanghai, China. BMC Infect Dis 2013; 13:421. [PMID: 24010871 PMCID: PMC3847129 DOI: 10.1186/1471-2334-13-421] [Citation(s) in RCA: 38] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2013] [Accepted: 09/04/2013] [Indexed: 01/17/2023] Open
Abstract
Background Sporadic hepatitis E has become an important public health concern in China. Accurate forecasting of the incidence of hepatitis E is needed to better plan future medical needs. Few mathematical models can be used because hepatitis E morbidity data has both linear and nonlinear patterns. We developed a combined mathematical model using an autoregressive integrated moving average model (ARIMA) and a back propagation neural network (BPNN) to forecast the incidence of hepatitis E. Methods The morbidity data of hepatitis E in Shanghai from 2000 to 2012 were retrieved from the China Information System for Disease Control and Prevention. The ARIMA-BPNN combined model was trained with 144 months of morbidity data from January 2000 to December 2011, validated with 12 months of data January 2012 to December 2012, and then employed to forecast hepatitis E incidence January 2013 to December 2013 in Shanghai. Residual analysis, Root Mean Square Error (RMSE), normalized Bayesian Information Criterion (BIC), and stationary R square methods were used to compare the goodness-of-fit among ARIMA models. The Bayesian regularization back-propagation algorithm was used to train the network. The mean error rate (MER) was used to assess the validity of the combined model. Results A total of 7,489 hepatitis E cases was reported in Shanghai from 2000 to 2012. Goodness-of-fit (stationary R2=0.531, BIC= −4.768, Ljung-Box Q statistics=15.59, P=0.482) and parameter estimates were used to determine the best-fitting model as ARIMA (0,1,1)×(0,1,1)12. Predicted morbidity values in 2012 from best-fitting ARIMA model and actual morbidity data from 2000 to 2011 were used to further construct the combined model. The MER of the ARIMA model and the ARIMA-BPNN combined model were 0.250 and 0.176, respectively. The forecasted incidence of hepatitis E in 2013 was 0.095 to 0.372 per 100,000 population. There was a seasonal variation with a peak during January-March and a nadir during August-October. Conclusions Time series analysis suggested a seasonal pattern of hepatitis E morbidity in Shanghai, China. An ARIMA-BPNN combined model was used to fit the linear and nonlinear patterns of time series data, and accurately forecast hepatitis E infections.
Collapse
Affiliation(s)
- Hong Ren
- The Key Laboratory of Public Health Safety of Minister of Education - Department of Epidemiology, Fudan University School of Public Health, Building 8 Room 441, 138 Yi Xue Yuan Road, Shanghai 200032, China.
| | | | | | | | | | | |
Collapse
|
13
|
Pirdashti M, Curteanu S, Kamangar MH, Hassim MH, Khatami MA. Artificial neural networks: applications in chemical engineering. REV CHEM ENG 2013. [DOI: 10.1515/revce-2013-0013] [Citation(s) in RCA: 57] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
14
|
ARAN OYA, YILDIZ OLCAYTANER, ALPAYDIN ETHEM. AN INCREMENTAL FRAMEWORK BASED ON CROSS-VALIDATION FOR ESTIMATING THE ARCHITECTURE OF A MULTILAYER PERCEPTRON. INT J PATTERN RECOGN 2011. [DOI: 10.1142/s0218001409007132] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
Abstract
We define the problem of optimizing the architecture of a multilayer perceptron (MLP) as a state space search and propose the MOST (Multiple Operators using Statistical Tests) framework that incrementally modifies the structure and checks for improvement using cross-validation. We consider five variants that implement forward/backward search, using single/multiple operators and searching depth-first/breadth-first. On 44 classification and 30 regression datasets, we exhaustively search for the optimal and evaluate the goodness based on: (1) Order, the accuracy with respect to the optimal and (2) Rank, the computational complexity. We check for the effect of two resampling methods (5 × 2, ten-fold cv), four statistical tests (5 × 2 cv t, ten-fold cv t, Wilcoxon, sign) and two corrections for multiple comparisons (Bonferroni, Holm). We also compare with Dynamic Node Creation (DNC) and Cascade Correlation (CC). Our results show that: (1) On most datasets, networks with few hidden units are optimal, (2) forward searching finds simpler architectures, (3) variants using single node additions (deletions) generally stop early and get stuck in simple (complex) networks, (4) choosing the best of multiple operators finds networks closer to the optimal, (5) MOST variants generally find simpler networks having lower or comparable error rates than DNC and CC.
Collapse
Affiliation(s)
- OYA ARAN
- Department of Computer Engineering, Boğaziçi University, TR-34342, Istanbul, Turkey
| | - OLCAY TANER YILDIZ
- Department of Computer Engineering, Boğaziçi University, TR-34342, Istanbul, Turkey
| | - ETHEM ALPAYDIN
- Department of Computer Engineering, Boğaziçi University, TR-34342, Istanbul, Turkey
| |
Collapse
|
15
|
Oliveira ALI, Medeiros EA, Rocha TABV, Bezerra MER, Veras RC. ON THE INFLUENCE OF PARAMETER θ- ON PERFORMANCE OF RBF NEURAL NETWORKS TRAINED WITH THE DYNAMIC DECAY ADJUSTMENT ALGORITHM. Int J Neural Syst 2011; 16:271-81. [PMID: 16972315 DOI: 10.1142/s0129065706000676] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2005] [Accepted: 06/22/2006] [Indexed: 01/17/2023]
Abstract
The dynamic decay adjustment (DDA) algorithm is a fast constructive algorithm for training RBF neural networks (RBFNs) and probabilistic neural networks (PNNs). The algorithm has two parameters, namely, θ+ and θ-. The papers which introduced DDA argued that those parameters would not heavily influence classification performance and therefore they recommended using always the default values of these parameters. In contrast, this paper shows that smaller values of parameter θ- can, for a considerable number of datasets, result in strong improvement in generalization performance. The experiments described here were carried out using twenty benchmark classification datasets from both Proben1 and the UCI repositories. The results show that for eleven of the datasets, the parameter θ- strongly influenced classification performance. The influence of θ- was also noticeable, although much less, on six of the datasets considered. This paper also compares the performance of RBF-DDA with θ- selection with both AdaBoost and Support Vector Machines (SVMs).
Collapse
Affiliation(s)
- Adriano L I Oliveira
- Department of Computing Systems, Polytechnic School of Engineering, Pernambuco State University, Rua Benfica, 455, Madalena, Recife - PE, Brazil.
| | | | | | | | | |
Collapse
|
16
|
Setiono R, Baesens B, Mues C. Rule extraction from minimal neural networks for credit card screening. Int J Neural Syst 2011; 21:265-76. [PMID: 21809474 DOI: 10.1142/s0129065711002821] [Citation(s) in RCA: 28] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
Abstract
While feedforward neural networks have been widely accepted as effective tools for solving classification problems, the issue of finding the best network architecture remains unresolved, particularly so in real-world problem settings. We address this issue in the context of credit card screening, where it is important to not only find a neural network with good predictive performance but also one that facilitates a clear explanation of how it produces its predictions. We show that minimal neural networks with as few as one hidden unit provide good predictive accuracy, while having the added advantage of making it easier to generate concise and comprehensible classification rules for the user. To further reduce model size, a novel approach is suggested in which network connections from the input units to this hidden unit are removed by a very straightaway pruning procedure. In terms of predictive accuracy, both the minimized neural networks and the rule sets generated from them are shown to compare favorably with other neural network based classifiers. The rules generated from the minimized neural networks are concise and thus easier to validate in a real-life setting.
Collapse
Affiliation(s)
- Rudy Setiono
- School of Computing, National University of Singapore, 13 Computing Drive, Singapore 117417, Republic of Singapore.
| | | | | |
Collapse
|
17
|
Mohamed MH. Rules extraction from constructively trained neural networks based on genetic algorithms. Neurocomputing 2011. [DOI: 10.1016/j.neucom.2011.04.009] [Citation(s) in RCA: 35] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
18
|
Khosravi A, Nahavandi S, Creighton D, Atiya AF. Comprehensive review of neural network-based prediction intervals and new advances. ACTA ACUST UNITED AC 2011; 22:1341-56. [PMID: 21803683 DOI: 10.1109/tnn.2011.2162110] [Citation(s) in RCA: 95] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
This paper evaluates the four leading techniques proposed in the literature for construction of prediction intervals (PIs) for neural network point forecasts. The delta, Bayesian, bootstrap, and mean-variance estimation (MVE) methods are reviewed and their performance for generating high-quality PIs is compared. PI-based measures are proposed and applied for the objective and quantitative assessment of each method's performance. A selection of 12 synthetic and real-world case studies is used to examine each method's performance for PI construction. The comparison is performed on the basis of the quality of generated PIs, the repeatability of the results, the computational requirements and the PIs variability with regard to the data uncertainty. The obtained results in this paper indicate that: 1) the delta and Bayesian methods are the best in terms of quality and repeatability, and 2) the MVE and bootstrap methods are the best in terms of low computational load and the width variability of PIs. This paper also introduces the concept of combinations of PIs, and proposes a new method for generating combined PIs using the traditional PIs. Genetic algorithm is applied for adjusting the combiner parameters through minimization of a PI-based cost function subject to two sets of restrictions. It is shown that the quality of PIs produced by the combiners is dramatically better than the quality of PIs obtained from each individual method.
Collapse
Affiliation(s)
- Abbas Khosravi
- Centre for Intelligent Systems Research, Deakin University, Geelong, Victoria 3117, Australia.
| | | | | | | |
Collapse
|
19
|
Khosravi A, Nahavandi S, Creighton D, Atiya AF. Lower Upper Bound Estimation Method for Construction of Neural Network-Based Prediction Intervals. ACTA ACUST UNITED AC 2011; 22:337-46. [DOI: 10.1109/tnn.2010.2096824] [Citation(s) in RCA: 384] [Impact Index Per Article: 27.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
20
|
Piuleac CG, Poulios I, Leon F, Curteanu S, Kouras A. Modeling Methodology Based on Stacked Neural Networks Applied to the Photocatalytic Degradation of Triclopyr. SEP SCI TECHNOL 2010. [DOI: 10.1080/01496395.2010.487736] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
21
|
A multi-objective memetic and hybrid methodology for optimizing the parameters and performance of artificial neural networks. Neurocomputing 2010. [DOI: 10.1016/j.neucom.2009.11.007] [Citation(s) in RCA: 56] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
22
|
Anselmino M, Malmberg K, Rydén L, Ohrvik J. A gluco-metabolic risk index with cardiovascular risk stratification potential in patients with coronary artery disease. Diab Vasc Dis Res 2009; 6:62-70. [PMID: 20368195 DOI: 10.1177/1479164109336052] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/17/2023] Open
Abstract
UNLABELLED The primary objective of this study was to classify patients with CAD as regards their gluco-metabolic state by easily available clinical variables. A secondary objective was to explore if it was possible to identify CAD patients at a high cardiovascular risk due to metabolic perturbations. The 1,867 patients with CAD were gluco-metabolically classified by an OGTT. Among these, 990 patients had complete data regarding all components of the metabolic syndrome, BMI, HbA1c and medical history. Only FPG and HDL-c adjusting for age significantly impacted OGTT classification. Based on these variables, a neural network reached a cross-validated misclassification rate of 37.8% compared with OGTT. By this criterion, 1,283 patients with complete one-year follow-up concerning all-cause mortality, myocardial infarction and stroke (CVE) were divided into low- and high-risk groups within which CVE were, respectively, 5.1 and 9.4% (p=0.016).Adjusting for confounding variables the relative risk for a CVE based on the neural network was 2.06 (95% CI: 1.18-3.58) compared with 1.37 (95% CI: 0.79-2.36) for OGTT. CONCLUSIONS The neural network, based on FPG, HDL-c and age, showed useful risk stratification capacities; it may, therefore, be of help when stratifying further risk of CVE in CAD patients.
Collapse
Affiliation(s)
- Matteo Anselmino
- Department of Medicine, Karolinska Institute, Solna, Stockholm, Sweden
| | | | | | | |
Collapse
|
23
|
An Evolutionary Approach for Tuning Artificial Neural Network Parameters. LECTURE NOTES IN COMPUTER SCIENCE 2009. [DOI: 10.1007/978-3-540-87656-4_20] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/02/2023]
|
24
|
do Carmo Nicoletti M, Bertini JR, Elizondo D, Franco L, Jerez JM. Constructive Neural Network Algorithms for Feedforward Architectures Suitable for Classification Tasks. CONSTRUCTIVE NEURAL NETWORKS 2009. [DOI: 10.1007/978-3-642-04512-7_1] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
25
|
Ang J, Tan K, Al-Mamun A. Training neural networks for classification using growth probability-based evolution. Neurocomputing 2008. [DOI: 10.1016/j.neucom.2007.10.011] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
26
|
Romero E, Alquézar R. Heuristics for the selection of weights in sequential feed-forward neural networks: An experimental study. Neurocomputing 2007. [DOI: 10.1016/j.neucom.2006.05.022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
27
|
Xu J, Ho DW. A new training and pruning algorithm based on node dependence and Jacobian rank deficiency. Neurocomputing 2006. [DOI: 10.1016/j.neucom.2005.11.005] [Citation(s) in RCA: 17] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
28
|
Romero E, Alquézar R. A sequential algorithm for feed-forward neural networks with optimal coefficients and interacting frequencies. Neurocomputing 2006. [DOI: 10.1016/j.neucom.2005.07.006] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
29
|
Ahmed FE. Artificial neural networks for diagnosis and survival prediction in colon cancer. Mol Cancer 2005; 4:29. [PMID: 16083507 PMCID: PMC1208946 DOI: 10.1186/1476-4598-4-29] [Citation(s) in RCA: 100] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/19/2005] [Accepted: 08/06/2005] [Indexed: 12/11/2022] Open
Abstract
ANNs are nonlinear regression computational devices that have been used for over 45 years in classification and survival prediction in several biomedical systems, including colon cancer. Described in this article is the theory behind the three-layer free forward artificial neural networks with backpropagation error, which is widely used in biomedical fields, and a methodological approach to its application for cancer research, as exemplified by colon cancer. Review of the literature shows that applications of these networks have improved the accuracy of colon cancer classification and survival prediction when compared to other statistical or clinicopathological methods. Accuracy, however, must be exercised when designing, using and publishing biomedical results employing machine-learning devices such as ANNs in worldwide literature in order to enhance confidence in the quality and reliability of reported data.
Collapse
Affiliation(s)
- Farid E Ahmed
- Department of Radiation Oncology, Leo W Jenkins Cancer Center, The Brody School of Medicine, East Carolina University, Greenville, NC 27858, USA.
| |
Collapse
|
30
|
Ma L, Khorasani K. Constructive Feedforward Neural Networks Using Hermite Polynomial Activation Functions. ACTA ACUST UNITED AC 2005; 16:821-33. [PMID: 16121724 DOI: 10.1109/tnn.2005.851786] [Citation(s) in RCA: 94] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
In this paper, a constructive one-hidden-layer network is introduced where each hidden unit employs a polynomial function for its activation function that is different from other units. Specifically, both a structure level as well as a function level adaptation methodologies are utilized in constructing the network. The functional level adaptation scheme ensures that the "growing" or constructive network has different activation functions for each neuron such that the network may be able to capture the underlying input-output map more effectively. The activation functions considered consist of orthonormal Hermite polynomials. It is shown through extensive simulations that the proposed network yields improved performance when compared to networks having identical sigmoidal activation functions.
Collapse
Affiliation(s)
- Liying Ma
- Department of Electrical and Computer Engineering, Concordia University, Montreal, QC, H3G 1M8 Canada
| | | |
Collapse
|
31
|
Guan P, Huang DS, Zhou BS. Forecasting model for the incidence of hepatitis A based on artificial neural network. World J Gastroenterol 2004; 10:3579-82. [PMID: 15534910 PMCID: PMC4611996 DOI: 10.3748/wjg.v10.i24.3579] [Citation(s) in RCA: 34] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/15/2022] Open
Abstract
AIM: To study the application of artificial neural network (ANN) in forecasting the incidence of hepatitis A, which had an autoregression phenomenon.
METHODS: The data of the incidence of hepatitis A in Liaoning Province from 1981 to 2001 were obtained from Liaoning Disease Control and Prevention Center. We used the autoregressive integrated moving average (ARIMA) model of time series analysis to determine whether there was any autoregression phenomenon in the data. Then the data of the incidence were switched into [0,1] intervals as the network theoretical output. The data from 1981 to 1997 were used as the training and verifying sets and the data from 1998 to 2001 were made up into the test set. STATISTICA neural network (ST NN) was used to construct, train and simulate the artificial neural network.
RESULTS: Twenty-four networks were tested and seven were retained. The best network we found had excellent performance, its regression ratio was 0.73, and its correlation was 0.69. There were 2 input variables in the network, one was AR(1), and the other was time. The number of units in hidden layer was 3. In ARIMA time series analysis results, the best model was first order autoregression without difference and smoothness. The total sum square error of the ANN model was 9090.21, the sum square error of the training set and testing set was 8377.52 and 712.69, respectively, they were all less than that of ARIMA model. The corresponding value of ARIMA was 12291.79, 8944.95 and 3346.84, respectively. The correlation coefficient of nonlinear regression (RNL) of ANN was 0.71, while the RNL of ARIMA linear autoregression model was 0.66.
CONCLUSION: ANN is superior to conventional methods in forecasting the incidence of hepatitis A which has an autoregression phenomenon.
Collapse
Affiliation(s)
- Peng Guan
- Department of Epidemiology, School of Public Health, China Medical University, Shenyang 110001, Liaoning Province, China
| | | | | |
Collapse
|