1
|
Ludwig O, Nunes U, Araujo R. Eigenvalue decay: A new method for neural network regularization. Neurocomputing 2014. [DOI: 10.1016/j.neucom.2013.08.005] [Citation(s) in RCA: 50] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
2
|
Garcia-Pedrajas N, Hervas-Martinez C, Munoz-Perez J. COVNET: a cooperative coevolutionary model for evolving artificial neural networks. ACTA ACUST UNITED AC 2012; 14:575-96. [PMID: 18238040 DOI: 10.1109/tnn.2003.810618] [Citation(s) in RCA: 69] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
This paper presents COVNET, a new cooperative coevolutionary model for evolving artificial neural networks. This model is based on the idea of coevolving subnetworks that must cooperate to form a solution for a specific problem, instead of evolving complete networks. The combination of this subnetworks is part of a coevolutionary process. The best combinations of subnetworks must be evolved together with the coevolution of the subnetworks. Several subpopulations of subnetworks coevolve cooperatively and genetically isolated. The individual of every subpopulation are combined to form whole networks. This is a different approach from most current models of evolutionary neural networks which try to develop whole networks. COVNET places as few restrictions as possible over the network structure, allowing the model to reach a wide variety of architectures during the evolution and to be easily extensible to other kind of neural networks. The performance of the model in solving three real problems of classification is compared with a modular network, the adaptive mixture of experts and with the results presented in the bibliography. COVNET has shown better generalization and produced smaller networks than the adaptive mixture of experts and has also achieved results, at least, comparable with the results in the bibliography.
Collapse
|
3
|
Performance comparison of neural architectures for on-line flux estimation in sensor-less vector-controlled IM drives. Neural Comput Appl 2012. [DOI: 10.1007/s00521-012-1107-y] [Citation(s) in RCA: 11] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
4
|
SETIONO RUDY, AZCARRAGA ARNULFO. GENERATING CONCISE SETS OF LINEAR REGRESSION RULES FROM ARTIFICIAL NEURAL NETWORKS. INT J ARTIF INTELL T 2011. [DOI: 10.1142/s0218213002000848] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Neural networks with a single hidden layer are known to be universal function approximators. However, due to the complexity of the network topology and the nonlinear transfer function used in computing the hidden unit activations, the predictions of a trained network are difficult to comprehend. On the other hand, predictions from a multiple linear regression equation are easy to understand but are not accurate when the underlying relationship between the input variables and the output variable is nonlinear. We have thus developed a method for multivariate function approximation which combines neural network learning, clustering and multiple regression. This method generates a set of multiple linear regression equations using neural networks, where the number of regression equations is determined by clustering the weighted input variables. The predictions for samples of the same cluster are computed by the same regression equation. Experimental results on a number of real-world data demonstrate that this new method generates relatively few regression equations from the training data samples. Yet, drawing from the universal function approximation capacity of neural networks, the predictive accuracy is high. The prediction errors are comparable to or lower than those achieved by existing function approximation methods.
Collapse
Affiliation(s)
- RUDY SETIONO
- School of Computing, National University of Singapore, 3 Science Drive 2, Singapore 117543, Singapore
| | - ARNULFO AZCARRAGA
- School of Computing, National University of Singapore, 3 Science Drive 2, Singapore 117543, Singapore
| |
Collapse
|
5
|
LING SH, LEUNG FHF, WONG LK, LAM HK. COMPUTATIONAL INTELLIGENCE TECHNIQUES FOR HOME ELECTRIC LOAD FORECASTING AND BALANCING. INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS 2011. [DOI: 10.1142/s1469026805001659] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
The paper presents an electric load balancing system for domestic use. An electric load forecasting system, which is realized by a genetic algorithm-based modified neural network, is employed. On forecasting the home power consumption profile, the load balancing system can adjust the amount of energy stored in battery accordingly, preventing it from reaching certain practical limits. A steady consumption from the AC mains can then be obtained which will benefit both the users and the utility company. An example will be given to illustrate the merits of the forecaster, and its performance on achieving the load balancing.
Collapse
Affiliation(s)
- S. H. LING
- Centre for Multimedia Signal Processing, Department of Electronic and Information Engineering, The Hong Kong Polytechnic University, Hung Hom, Kowloon, Hong Kong, China
| | - F. H. F. LEUNG
- Centre for Multimedia Signal Processing, Department of Electronic and Information Engineering, The Hong Kong Polytechnic University, Hung Hom, Kowloon, Hong Kong, China
| | - L. K. WONG
- Centre for Multimedia Signal Processing, Department of Electronic and Information Engineering, The Hong Kong Polytechnic University, Hung Hom, Kowloon, Hong Kong, China
| | - H. K. LAM
- Centre for Multimedia Signal Processing, Department of Electronic and Information Engineering, The Hong Kong Polytechnic University, Hung Hom, Kowloon, Hong Kong, China
| |
Collapse
|
6
|
Ludwig O, Nunes U. Novel maximum-margin training algorithms for supervised neural networks. ACTA ACUST UNITED AC 2010; 21:972-84. [PMID: 20409990 DOI: 10.1109/tnn.2010.2046423] [Citation(s) in RCA: 63] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
This paper proposes three novel training methods, two of them based on the backpropagation approach and a third one based on information theory for multilayer perceptron (MLP) binary classifiers. Both backpropagation methods are based on the maximal-margin (MM) principle. The first one, based on the gradient descent with adaptive learning rate algorithm (GDX) and named maximum-margin GDX (MMGDX), directly increases the margin of the MLP output-layer hyperplane. The proposed method jointly optimizes both MLP layers in a single process, backpropagating the gradient of an MM-based objective function, through the output and hidden layers, in order to create a hidden-layer space that enables a higher margin for the output-layer hyperplane, avoiding the testing of many arbitrary kernels, as occurs in case of support vector machine (SVM) training. The proposed MM-based objective function aims to stretch out the margin to its limit. An objective function based on Lp-norm is also proposed in order to take into account the idea of support vectors, however, overcoming the complexity involved in solving a constrained optimization problem, usually in SVM training. In fact, all the training methods proposed in this paper have time and space complexities O(N) while usual SVM training methods have time complexity O(N (3)) and space complexity O(N (2)) , where N is the training-data-set size. The second approach, named minimization of interclass interference (MICI), has an objective function inspired on the Fisher discriminant analysis. Such algorithm aims to create an MLP hidden output where the patterns have a desirable statistical distribution. In both training methods, the maximum area under ROC curve (AUC) is applied as stop criterion. The third approach offers a robust training framework able to take the best of each proposed training method. The main idea is to compose a neural model by using neurons extracted from three other neural networks, each one previously trained by MICI, MMGDX, and Levenberg-Marquard (LM), respectively. The resulting neural network was named assembled neural network (ASNN). Benchmark data sets of real-world problems have been used in experiments that enable a comparison with other state-of-the-art classifiers. The results provide evidence of the effectiveness of our methods regarding accuracy, AUC, and balanced error rate.
Collapse
Affiliation(s)
- Oswaldo Ludwig
- ISR-Institute of Systems and Robotics, Department of Electrical and Computer Engineering, University of Coimbra Polo II, 3030-290 Coimbra, Portugal.
| | | |
Collapse
|
7
|
Islam MM, Yao X, Shahriar Nirjon SMS, Islam MA, Murase K. Bagging and boosting negatively correlated neural networks. ACTA ACUST UNITED AC 2008; 38:771-84. [PMID: 18558541 DOI: 10.1109/tsmcb.2008.922055] [Citation(s) in RCA: 68] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
In this paper, we propose two cooperative ensemble learning algorithms, i.e., NegBagg and NegBoost, for designing neural network (NN) ensembles. The proposed algorithms incrementally train different individual NNs in an ensemble using the negative correlation learning algorithm. Bagging and boosting algorithms are used in NegBagg and NegBoost, respectively, to create different training sets for different NNs in the ensemble. The idea behind using negative correlation learning in conjunction with the bagging/boosting algorithm is to facilitate interaction and cooperation among NNs during their training. Both NegBagg and NegBoost use a constructive approach to automatically determine the number of hidden neurons for NNs. NegBoost also uses the constructive approach to automatically determine the number of NNs for the ensemble. The two algorithms have been tested on a number of benchmark problems in machine learning and NNs, including Australian credit card assessment, breast cancer, diabetes, glass, heart disease, letter recognition, satellite, soybean, and waveform problems. The experimental results show that NegBagg and NegBoost require a small number of training epochs to produce compact NN ensembles with good generalization.
Collapse
Affiliation(s)
- Md Monirul Islam
- Bangladesh University of Engineering and Technology (BUET), Dhaka 1000, Bangladesh
| | | | | | | | | |
Collapse
|
8
|
Romero E, Alquézar R. Heuristics for the selection of weights in sequential feed-forward neural networks: An experimental study. Neurocomputing 2007. [DOI: 10.1016/j.neucom.2006.05.022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
9
|
Romero E, Alquézar R. A sequential algorithm for feed-forward neural networks with optimal coefficients and interacting frequencies. Neurocomputing 2006. [DOI: 10.1016/j.neucom.2005.07.006] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/17/2023]
|
10
|
García-Pedrajas N, Ortiz-Boyer D, Hervás-Martínez C. An alternative approach for neural network evolution with a genetic algorithm: Crossover by combinatorial optimization. Neural Netw 2006; 19:514-28. [PMID: 16343847 DOI: 10.1016/j.neunet.2005.08.014] [Citation(s) in RCA: 51] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2004] [Accepted: 08/11/2005] [Indexed: 11/24/2022]
Abstract
In this work we present a new approach to crossover operator in the genetic evolution of neural networks. The most widely used evolutionary computation paradigm for neural network evolution is evolutionary programming. This paradigm is usually preferred due to the problems caused by the application of crossover to neural network evolution. However, crossover is the most innovative operator within the field of evolutionary computation. One of the most notorious problems with the application of crossover to neural networks is known as the permutation problem. This problem occurs due to the fact that the same network can be represented in a genetic coding by many different codifications. Our approach modifies the standard crossover operator taking into account the special features of the individuals to be mated. We present a new model for mating individuals that considers the structure of the hidden layer and redefines the crossover operator. As each hidden node represents a non-linear projection of the input variables, we approach the crossover as a problem on combinatorial optimization. We can formulate the problem as the extraction of a subset of near-optimal projections to create the hidden layer of the new network. This new approach is compared to a classical crossover in 25 real-world problems with an excellent performance. Moreover, the networks obtained are much smaller than those obtained with classical crossover operator.
Collapse
Affiliation(s)
- Nicolás García-Pedrajas
- Department of Computing and Numerical Analysis, University of Córdoba, 14071 Córdoba, Spain.
| | | | | |
Collapse
|
11
|
Tsai JT, Chou JH, Liu TK. Tuning the structure and parameters of a neural network by using hybrid Taguchi-genetic algorithm. ACTA ACUST UNITED AC 2006; 17:69-80. [PMID: 16526477 DOI: 10.1109/tnn.2005.860885] [Citation(s) in RCA: 80] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
In this paper, a hybrid Taguchi-genetic algorithm (HTGA) is applied to solve the problem of tuning both network structure and parameters of a feedforward neural network. The HTGA approach is a method of combining the traditional genetic algorithm (TGA), which has a powerful global exploration capability, with the Taguchi method, which can exploit the optimum offspring. The Taguchi method is inserted between crossover and mutation operations of a TGA. Then, the systematic reasoning ability of the Taguchi method is incorporated in the crossover operations to select the better genes to achieve crossover, and consequently enhance the genetic algorithms. Therefore, the HTGA approach can be more robust, statistically sound, and quickly convergent. First, the authors evaluate the performance of the presented HTGA approach by studying some global numerical optimization problems. Then, the presented HTGA approach is effectively applied to solve three examples on forecasting the sunspot numbers, tuning the associative memory, and solving the XOR problem. The numbers of hidden nodes and the links of the feedforward neural network are chosen by increasing them from small numbers until the learning performance is good enough. As a result, a partially connected feedforward neural network can be obtained after tuning. This implies that the cost of implementation of the neural network can be reduced. In these studied problems of tuning both network structure and parameters of a feedforward neural network, there are many parameters and numerous local optima so that these studied problems are challenging enough for evaluating the performances of any proposed GA-based approaches. The computational experiments show that the presented HTGA approach can obtain better results than the existing method reported recently in the literature.
Collapse
Affiliation(s)
- Jinn-Tsong Tsai
- Department of Medical Information Management, Kaohsiung Medical University, Kaohsiung 807, Taiwan, ROC
| | | | | |
Collapse
|
12
|
Garcı́a-Pedrajas N, Ortiz-Boyer D, Hervás-Martı́nez C. Cooperative coevolution of generalized multi-layer perceptrons. Neurocomputing 2004. [DOI: 10.1016/j.neucom.2003.09.004] [Citation(s) in RCA: 9] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
13
|
Abstract
In this paper, a functions localized network with branch gates (FLN-bg) is studied, which consists of a basic network and a branch gate network. The branch gate network is used to determine which intermediate nodes of the basic network should be connected to the output node with a gate coefficient ranging from 0 to 1. This determination will adjust the outputs of the intermediate nodes of the basic network depending on the values of the inputs of the network in order to realize a functions localized network. FLN-bg is applied to function approximation problems and a two-spiral problem. The simulation results show that FLN-bg exhibits better performance than conventional neural networks with comparable complexity.
Collapse
Affiliation(s)
- Qingyu Xiong
- Automation College, Chongqing University, Chongqing, People's Republic of China.
| | | | | | | |
Collapse
|
14
|
Islam M, Xin Yao, Murase K. A constructive algorithm for training cooperative neural network ensembles. ACTA ACUST UNITED AC 2003; 14:820-34. [DOI: 10.1109/tnn.2003.813832] [Citation(s) in RCA: 215] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
15
|
Thangavel P, Kathirvalavakumar T. Simultaneous perturbation for single hidden layer networks — cascade learning. Neurocomputing 2003. [DOI: 10.1016/s0925-2312(01)00704-4] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
16
|
Leung F, Lam H, Ling S, Tam P. Tuning of the structure and parameters of a neural network using an improved genetic algorithm. ACTA ACUST UNITED AC 2003; 14:79-88. [DOI: 10.1109/tnn.2002.804317] [Citation(s) in RCA: 510] [Impact Index Per Article: 23.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
17
|
Deterministic nonmonotone strategies for effective training of multilayer perceptrons. ACTA ACUST UNITED AC 2002; 13:1268-84. [DOI: 10.1109/tnn.2002.804225] [Citation(s) in RCA: 25] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
18
|
Setiono R, Wee Kheng Leow, Zurada J. Extraction of rules from artificial neural networks for nonlinear regression. ACTA ACUST UNITED AC 2002; 13:564-77. [DOI: 10.1109/tnn.2002.1000125] [Citation(s) in RCA: 137] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
|
19
|
|
20
|
Abstract
This article presents an algorithm that constructs feedforward neural networks with a single hidden layer for pattern classification. The algorithm starts with a small number of hidden units in the network and adds more hidden units as needed to improve the network's predictive accuracy. To determine when to stop adding new hidden units, the algorithm makes use of a subset of the available training samples for cross validation. New hidden units are added to the network only if they improve the classification accuracy of the network on the training samples and on the cross-validation samples. Extensive experimental results show that the algorithm is effective in obtaining networks with predictive accuracy rates that are better than those obtained by state-of-the-art decision tree methods.
Collapse
Affiliation(s)
- R Setiono
- School of Computing, National University of Singapore, Singapore 117543.
| |
Collapse
|
21
|
Hervás C, Toledo R, Silva M. Use of pruned computational neural networks for processing the response of oscillating chemical reactions with a view to analyzing nonlinear multicomponent mixtures. JOURNAL OF CHEMICAL INFORMATION AND COMPUTER SCIENCES 2001; 41:1083-92. [PMID: 11500128 DOI: 10.1021/ci010012j] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Abstract
The suitability of pruned computational neural networks (CNNs) for resolving nonlinear multicomponent systems involving synergistic effects by use of oscillating chemical reaction-based methods implemented using the analyte pulse perturbation technique is demonstrated. The CNN input data used for this purpose are estimates provided by the Levenberg-Marquardt method in the form of a three-parameter Gaussian curve associated with the singular profile obtained when the oscillating system is perturbed by an analyte mixture. The performance of the proposed method was assessed by applying it to the resolution of mixtures of pyrogallol and gallic acid based on their perturbating effect on a classical oscillating chemical system, viz. the Belousov-Zhabotinskyi reaction. A straightforward network topology (3:3:2, with 18 connections after pruning) allowed the resolution of mixtures of the two analytes in concentration ratios from 1:7 to 6:2 with a standard error of prediction for the testing set of 4.01 and 8.98% for pyrogallol and gallic acid, respectively. The reduced dimensions of the selected CNN architecture allowed a mathematical transformation of the input vector into the output one that can be easily implemented via software. Finally, the suitability of response surface analysis as an alternative to CNNs was also tested. The results were poor (relative errors were high), which confirms that properly selected pruned CNNs are effective tools for solving the analytical problem addressed in this work.
Collapse
Affiliation(s)
- C Hervás
- Department of Computer Science, University of Córdoba, Spain.
| | | | | |
Collapse
|
22
|
Abstract
The Constraint Based Decomposition (CBD) is a constructive neural network technique that builds a three or four layer network, has guaranteed convergence and can deal with binary, n-ary, class labeled and real-value problems. CBD is shown to be able to solve complicated problems in a simple, fast and reliable manner. The technique is further enhanced by two modifications (locking detection and redundancy elimination) which address the training speed and the efficiency of the internal representation built by the network. The redundancy elimination aims at building more compact architectures while the locking detection aims at improving the training speed. The computational cost of the redundancy elimination is negligible and this enhancement can be used for any problem. However, the computational cost of the locking detection is exponential in the number of dimensions and should only be used in low dimensional spaces. The experimental results show the performance of the algorithm presented in a series of classical benchmark problems including the 2-spiral problem and the Iris, Wine, Glass, Lenses, Ionosphere, Lung cancer, Pima Indians, Bupa, TicTacToe, Balance and Zoo data sets from the UCI machine learning repository. CBD's generalization accuracy is compared with that of C4.5, C4.5 with rules, incremental decision trees, oblique classifiers, linear machine decision trees, CN2, learning vector quantization (LVQ), backpropagation, nearest neighbor, Q* and radial basis functions (RBFs). CBD provides the second best average accuracy on the problems tested as well as the best reliability (the lowest standard deviation).
Collapse
Affiliation(s)
- S Draghici
- Department of Computer Science, Wayne State University, Detroit, MI 48202, USA.
| |
Collapse
|