Research Article Open Access

Adaptive Resonance Theory Training Parameters: Pretty Good Sets

Taisir Mohammad Eldos and Abdulaziz Suleiman Almazyad


Problem statement: ART1 artificial neural networks offer good tools for test clustering, where no expert is needed if the system is well trained. However, having no output reference for the input patterns makes it hard to judge the quality of the training. Moreover, the performance depends to a great extent on a set of training parameters. Designers follow some recommendations or depend on their expertise in finding good sets with no performance guarantees. Many methods were proposed; from greedy methods offering quick and acceptable solutions to evolutionary algorithms offering suboptimal sets of parameters. While the evolutionary algorithms are a good choice for quality, the computational cost is large even for an offline process; after all, computing resources are not for free. Approach: We introduced a method for selecting a set of parameters that yields a comparable performance and robust operation, with relatively low cost compared to the evolutionary methods. This method located a suitable set through repetitive portioning of the range, by considering the best subset for the next iteration. Results: Tests have shown that performance comparable with the computationally intensive evolutionary methods could be achieved in much less time. Conclusion: The repetitive portioning method for finding a good set of training parameters is very cost effective and yields good performance.

Journal of Computer Science
Volume 6 No. 12, 2010, 1443-1449


Submitted On: 9 September 2010 Published On: 6 November 2010

How to Cite: Eldos, T. M. & Almazyad, A. S. (2010). Adaptive Resonance Theory Training Parameters: Pretty Good Sets. Journal of Computer Science, 6(12), 1443-1449.

  • 1 Citations



  • Adaptive resonance theory
  • pretty good set
  • artificial neural network
  • optimization