Abstract
Established techniques from information retrieval (IR) and machine learning (ML) have shown varying degrees of success in the automatic classification of real-world text. The capabilities of an extended version of the Simple recurrent network (SRN) for classifying news titles from the Reuters-21578 Corpus are explored. The architecture is composed of two hidden layers where each layer has an associated context layer that takes copies of previous activation states and integrates them with current activations. This results in improved performance, stability and generalisation by the adjustment of the percentage of previous activation strengths kept "in memory" by what is defined as the hysteresis parameter. The study demonstrates that this partial feedback of activations must be carefully fine-tuned to maintain optimal performance. Correctly adjusting the hysteresis values for very long and noisy text sequences is critical as classification performance degrades catastrophic ally when values are not optimally set.
Original language | English |
---|---|
Title of host publication | 2007 International Joint Conference on Neural Networks |
Publisher | IEEE |
Pages | 2936 - 2941 |
Number of pages | 6 |
ISBN (Print) | 978-1-4244-1379-9 |
DOIs | |
Publication status | Published - 2007 |
Externally published | Yes |
Event | 2007 International Joint Conference on Neural Networks - Orlando, United States Duration: 12 Aug 2007 → 17 Aug 2007 |
Conference
Conference | 2007 International Joint Conference on Neural Networks |
---|---|
Country/Territory | United States |
City | Orlando |
Period | 12/08/07 → 17/08/07 |