Training stuff
A common criticism of neural networks, particularly in robotics, is that they require a large diversity of training for real-world operation. This is not surprising, since any learning machine needs enough representative examples in order to capture the underlying structure that allows it to generalize to new cases. Dean A. Powerless, in his research presented in the paper “Knowledge-Based Training of Artificial Neural Networks for Autonomous Robot Driving,” uses a neural network to train a robotic vehicle to drive on multiple types of roads (single lane, multi-lane, dirt, etc.). A lot of his research is dedicated to (1) extrapolating multiple training scenarios from a single training experience, and (2) preserving past training diversity so that the system does not become over-trained (if, for example, it is presented with a series of right turns - it should not learn to always turn right). These problems are common in neural networks that must decide from a wide variety of responses, but they can be treated in several ways, for example by randomly shuffling training examples, by using a numerical optimization algorithm that does not take too large steps when changing network connections following an example, or by grouping examples into so-called mini-batches.
Theoretical issues
AK Dewdney"), a mathematical and computer science scientist at the University of Western Ontario and former Scientific American columnist, wrote in 1997, "Although neural networks do solve some toy problems, their computing power is so limited that I am surprised no one takes them seriously as a general problem-solving tool." No neural network has ever been shown to solve computationally difficult problems, such as the N-Queens problem, the traveling salesman problem, or the problem of factoring large integers.
Aside from their usefulness, a fundamental objection to artificial neural networks is that they fail to reflect how real neurons work. Backward propagation is at the heart of artificial neural networks and most not only is there no evidence of any mechanism for such natural neural networks,[54] it appears to contradict the fundamental principle of real neurons that information can only flow forward along the axon. How information is encoded by actual neurons is not yet known. What is known is that sensory neurons fire action potentials more frequently with sensor activation and muscle cells fire more strongly when their associated motor neurons receive action potentials more frequently.[55] Apart from the simplest case of just transmitting information from a neuron to a sensor motor neuron almost nothing is known of the underlying general principles of how information is handled by real neural networks.
The purpose of artificial neural networks is not necessarily to replicate real neural function but to be inspired by natural neural networks as an approach to inherently parallel computing that provides solutions to problems that until now have been intractable. Therefore, a central claim of artificial neural networks is that they embody some new and powerful general principle for information processing. Unfortunately, these general principles are poorly defined and are often claimed to be emergent from the neural network itself. This allows simple statistical association (the basic function of artificial neural networks), which is described as learning or recognition. As a result, artificial neural networks have, according to Dewdney, a "somewhat not at all quality, which imparts a peculiar aura of laziness and a distinct lack of curiosity about how good these computing systems are. No human hand (or mind) intervenes; solutions. They meet as if by magic, and no one, it seems, has learned anything."[56].
Hardware problems.
Implementing large, efficient neural network software requires considerable processing and storage resources. While the brain has adapted its hardware to the task of signal processing through a graph of neurons, simulating even a simplified form in the von Neumann architecture can force a neural network designer to use many millions of database rows for its connections, which can consume large amounts of RAM and hard drive space. Furthermore, the designer of neural network systems will often need to use an incredible amount of processing power and CPU time to simulate the transmission of signals through many of these connections and their associated neurons.
Jürgen Schmidhuber notes that the resurgence of neural networks in the 20th century, and their renewed success in image recognition tasks, is largely attributable to advances in hardware: from 1991 to 2015, computing power, especially as delivered by GPGPUs (on GPUs), has increased about a million times, making the standard backpropagation algorithm viable for training networks that are several layers deeper than before. (but he adds that this does not solve algorithmic problems such as the gradient vanishing problem "in a fundamental way"). Using the GPU instead of ordinary CPUs can bring training times for some networks down from months to mere days.
Computing power continues to grow more or less in accordance with Moore's Law, which can provide sufficient resources to carry out new tasks. Neuromorphic engineering addresses the hardware difficulty directly, by building non-von Neumann chips with circuits designed to implement neural networks from scratch. Google has also designed a chip optimized for neural network processing called the Tensor Processing Unit, or TPU.
Practical counterexamples to criticism.
Arguments against Dewdney's position are that neural networks have been successfully used to solve many complex and diverse tasks, ranging from autonomously flying airplanes to credit card fraud detection.
Technology writer Roger Bridgman has commented on DEWDNEY's statements about neural networks:
Neural networks, for example, are in the dock not only because they have been hyped to the high heavens, (what have you, right?), but also because you can create a successful network without understanding how it worked: the pile of numbers that captures its behavior would in all likelihood be "an opaque, unreadable table... of no value as a scientific resource."
Despite his emphatic statement that science is not technology, it seems Dewdney is here to deride neural networks as bad science when most of those devising them are trying to be good engineers. A readable table that a useful machine could read would still be well worth having.
While it is true that analyzing what has been learned by an artificial neural network is difficult, it is much easier to do than analyzing what has been learned by a biological neural network. On the other hand, researchers involved in exploring learning algorithms for neural networks are gradually discovering generic principles that allow a learning machine to be successful. For example, Bengio and LeCun (2007) wrote an article on local vs. local learning. Non-local as well as shallow versus deep architecture.
Hybrid approaches.
Some other criticism comes from proponents of hybrid models (combination of neural networks and symbolic approaches), who believe that the intermix of these two approaches can better capture the mechanisms of the human mind.