A Hardware Implementation of a Mixture of Experts

By Antony Savich, PhD, May 2008 - 2011
****

Abstract: Recent applications of Artificial Neural Networks (ANN) have increased the size, performance and difficulty requirements of single networks used to solve complex problems of today. Previous research focused on optimizing single ANN performance and hardware requirements for accelerated training cycles and use of cheaper implementation platforms (both on generic processors and more recently on reconfigurable hardware). After training these single networks were inflexible to change, and in general it was demonstrated difficult to use the complete parallelism of single ANN algorithms and fit a large flexible fully-parallel network design into the confines of a single chip solution. This research focuses on realizing solutions for complex problems using, instead of a single large ANN, a subdivision of the solution into a group of smaller networks in a Mixture of Experts configuration, with the flexibility of modularly adding new functionality after initial training process. The goal of this research is to demonstrate the ability of such a solution to have a performance advantage over current implementations with single network topologies by using smaller faster networks, and flexibility advantage over monolithically trained networks in adapting to new conditions by expanding the problem/solution set. This will be accomplished through the implementation of a fully functional single FPGA Mixture of Experts system and its subsequent performance analysis on various sample problems.