Artificial neural networks can be trained to perform excellently in many application areas. While they can learn from raw data to solve sophisticated recognition and analysis problems, the acquired knowledge remains hidden within the network architecture and is not readily accessible for analysis or further use: Trained networks are black boxes. Recent research efforts therefore investigate the possibility to extract symbolic knowledge from trained networks, in order to analyze, validate, and reuse the structural insights gained implicitly during the training process. In this paper, we will study how knowledge in form of propositional logic programs can be obtained in such a way that the programs are as simple as possible - where simple is being understood in some clearly defined and meaningful way.
& Hitzler, P.
(2005). Extracting Reduced Logic Programs from Artificial Neural Networks. Proceedings of the IJCAI-05 Workshop on Neural-Symbolic Learning and Reasoning.