Semantic neural network
This article possibly contains original research. (September 2007) |
Semantic neural network (SNN) is based on John von Neumann's neural network [von Neumann, 1966] and Nikolai Amosov M-Network.[1][2] There are limitations to a link topology for the von Neumann’s network but SNN accept a case without these limitations. Only logical values can be processed, but SNN accept that fuzzy values can be processed too. All neurons into the von Neumann network are synchronized by tacts. For further use of self-synchronizing circuit technique SNN accepts neurons can be self-running or synchronized.
In contrast to the von Neumann network there are no limitations for topology of neurons for semantic networks. It leads to the impossibility of relative addressing of neurons as it was done by von Neumann. In this case an absolute readdressing should be used. Every neuron should have a unique identifier that would provide a direct access to another neuron. Of course, neurons interacting by axons-dendrites should have each other's identifiers. An absolute readdressing can be modulated by using neuron specificity as it was realized for biological neural networks.
There’s no description for self-reflectiveness and self-modification abilities into the initial description of semantic networks [Dudar Z.V., Shuklin D.E., 2000]. But in [Shuklin D.E. 2004] a conclusion had been drawn about the necessity of introspection and self-modification abilities in the system. For maintenance of these abilities a concept of pointer to neuron is provided. Pointers represent virtual connections between neurons. In this model, bodies and signals transferring through the neurons connections represent a physical body, and virtual connections between neurons are representing an astral body. It is proposed to create models of artificial neuron networks on the basis of virtual machine supporting the opportunity for paranormal effects.
SNN is generally used for natural language processing.
Related models
[edit]- Computational creativity[3]
- Semantic hashing[4]
- Semantic Pointer Architecture[5]
- Sparse distributed memory
References
[edit]- ^ Amosov, N. M., A. M. Kasatkin, and L. M. Kasatkina. "Active semantic networks in robots with independent control." Proceedings of the 4th international joint conference on Artificial intelligence-Volume 1. Morgan Kaufmann Publishers Inc., 1975.
- ^ Amosov, N. M., E. M. Kussul, and A. M. Kasatkin. "29. NEURONLIKE NETWORKS, ATTENTION, ARTIFICIAL INTELLIGENCE." Neurocomputers and Attention: Connectionism and neurocomputers 2 (1991): 433.
- ^ Marupaka, Nagendra, and Ali A. Minai. "Connectivity and creativity in semantic neural networks." Neural Networks (IJCNN), The 2011 International Joint Conference on. IEEE, 2011.
- ^ Salakhutdinov, Ruslan, and Geoffrey Hinton. "Semantic hashing." RBM 500.3 (2007): 500.
- ^ Eliasmith, Chris, et al. "A large-scale model of the functioning brain." science 338.6111 (2012): 1202-1205.
- Neumann, J., 1966. Theory of self-reproducing automata, edited and completed by Arthur W. Burks. - University of Illinois press, Urbana and London
- Dudar Z.V., Shuklin D.E., 2000. Implementation of neurons for semantic neural nets that’s understanding texts in natural language. In Radio-electronika i informatika KhTURE, 2000. No 4. Р. 89-96.
- Shuklin D.E., 2004. The further development of semantic neural network models. In Artificial Intelligence, Donetsk, "Nauka i obrazovanie" Institute of Artificial Intelligence, Ukraine, 2004, No 3. P. 598-606
- Shuklin D.E. The Structure of a Semantic Neural Network Extracting the Meaning from a Text, In Cybernetics and Systems Analysis, Volume 37, Number 2, 4 March 2001, pp. 182–186(5) [1]
- Shuklin D.E. The Structure of a Semantic Neural Network Realizing Morphological and Syntactic Analysis of a Text, In Cybernetics and Systems Analysis, Volume 37, Number 5, September 2001, pp. 770–776(7)
- Shuklin D.E. Realization of a Binary Clocked Linear Tree and Its Use for Processing Texts in Natural Languages, In Cybernetics and Systems Analysis, Volume 38, Number 4, July 2002, pp. 503–508(6)