Introduction - If you have any usage issues, please Google them yourself
This a collection of four different S-function implementations of the recurrent fuzzy neural network (RFNN) described in detail in [1]. It is a four-layer, neuro-fuzzy network trained exclusively by error backpropagation at layers 2 and 4. The network employs 4 sets of adjustable parameters. In Layer 2: mean[i,j], sigma[i,j] and Theta[i,j] and in Layer 4: Weights w4[m,j]. The network uses considerably less adjustable parameters than ANFIS/CANFIS and therefore, its training is generally faster. This makes it ideal for on-line learning/operation. Also, its approximating/mapping power is increased due to the employment of dynamic elements within Layer 2. Scatter-type and Grid-type methods are selected for input space partitioning.