TY - GEN
T1 - Hyper neuron - One neuron with infinite states
AU - Bazrafkan, Shabab
AU - Lemley, Joseph
AU - Corcoran, Peter
N1 - Publisher Copyright:
© Springer International Publishing AG 2017.
PY - 2017
Y1 - 2017
N2 - Neural networks learn by creating a decision boundary. The shape and smoothness of the decision boundary is ultimately determined by the activation function and the architecture. As the number of neurons in an artificial neural network increase to infinity, the decision function becomes smooth. A network with an infiniete number of neurons is impossible to implement with finite resources, but the behavior of such a network can be modeled to an arbitrary degree of precision by using standard numerical techniques. We named the resulting model Hyper Neuron. A flexible characteristic function controlling the rate of variations in the weights of these neurons is used. The Hyper Neuron does not require any assumptions about the parameter distribution. It utilizes a numerical methodology that contrasts with previous work (such as infinite neural networks) which relies on assumptions about the distribution. In the classical model of a neuron, each neuron has a single state and output which is determined by an input, the weights, and the bias. Consider a neuron with more than one distinct output for the same input. A layer made from an infinite number of these neurons can be modeled as a single neuron with an infinite number of states and an infinite weight field. This kind of neuron is called a “Hyper Neuron” indicated by symbolQ. Now consider the independent variable x where the Hyper Neuron is defined over it i.e., the functionQ(x) is defined over the space x ∈ ℝN To model the data in such a way that it represents the target distribution, we use weighted inputs and non-linearity functions, where the weights are not vectors but instead are multidimensional functions (Formula Presented) which define the weight field between two Hyper Neurons when the input is given by another Hyper Neuron. This function is simplified as (Formula Presented) when the input is a feature space or a conventional layer. In these equations k is the previous layer, k + 1 is the layer with the Hyper Neuron, ik is the ith element in the previous layer and p represents parameters of the weight field function. Hyper Neurons follow naturally from numerical models involving an infinite number of conventional neurons in a single layer and the associated weight fields are described by characteristic functions. To validate this idea, experiments were performed that used sinusoidal functions for the weight fields, because they allowed rapid changes due to the inclusion of their frequency as a parameter that the network learned from the input data. A comparison between the proposed model and a conventional model containing up to 7 neurons was performed.
AB - Neural networks learn by creating a decision boundary. The shape and smoothness of the decision boundary is ultimately determined by the activation function and the architecture. As the number of neurons in an artificial neural network increase to infinity, the decision function becomes smooth. A network with an infiniete number of neurons is impossible to implement with finite resources, but the behavior of such a network can be modeled to an arbitrary degree of precision by using standard numerical techniques. We named the resulting model Hyper Neuron. A flexible characteristic function controlling the rate of variations in the weights of these neurons is used. The Hyper Neuron does not require any assumptions about the parameter distribution. It utilizes a numerical methodology that contrasts with previous work (such as infinite neural networks) which relies on assumptions about the distribution. In the classical model of a neuron, each neuron has a single state and output which is determined by an input, the weights, and the bias. Consider a neuron with more than one distinct output for the same input. A layer made from an infinite number of these neurons can be modeled as a single neuron with an infinite number of states and an infinite weight field. This kind of neuron is called a “Hyper Neuron” indicated by symbolQ. Now consider the independent variable x where the Hyper Neuron is defined over it i.e., the functionQ(x) is defined over the space x ∈ ℝN To model the data in such a way that it represents the target distribution, we use weighted inputs and non-linearity functions, where the weights are not vectors but instead are multidimensional functions (Formula Presented) which define the weight field between two Hyper Neurons when the input is given by another Hyper Neuron. This function is simplified as (Formula Presented) when the input is a feature space or a conventional layer. In these equations k is the previous layer, k + 1 is the layer with the Hyper Neuron, ik is the ith element in the previous layer and p represents parameters of the weight field function. Hyper Neurons follow naturally from numerical models involving an infinite number of conventional neurons in a single layer and the associated weight fields are described by characteristic functions. To validate this idea, experiments were performed that used sinusoidal functions for the weight fields, because they allowed rapid changes due to the inclusion of their frequency as a parameter that the network learned from the input data. A comparison between the proposed model and a conventional model containing up to 7 neurons was performed.
KW - Artificial neural networks
KW - Infinite neural networks
KW - Numerical modeling
UR - https://www.scopus.com/pages/publications/85034217046
M3 - Conference Publication
AN - SCOPUS:85034217046
SN - 9783319685991
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 432
EP - 433
BT - Artificial Neural Networks and Machine Learning – ICANN 2017 - 26th International Conference on Artificial Neural Networks, Proceedings
A2 - Verschure, Paul F.
A2 - Lintas, Alessandra
A2 - Villa, Alessandro E.
A2 - Rovetta, Stefano
PB - Springer-Verlag
T2 - 26th International Conference on Artificial Neural Networks, ICANN 2017
Y2 - 11 September 2017 through 14 September 2017
ER -