HOME

TheInfoList



OR:

In
computational neuroscience Computational neuroscience (also known as theoretical neuroscience or mathematical neuroscience) is a branch of  neuroscience which employs mathematics, computer science, theoretical analysis and abstractions of the brain to understand th ...
, SUPS (for Synaptic Updates Per Second) or formerly CUPS (Connections Updates Per Second) is a measure of a neuronal network performance, useful in fields of
neuroscience Neuroscience is the scientific study of the nervous system (the brain, spinal cord, and peripheral nervous system), its functions, and its disorders. It is a multidisciplinary science that combines physiology, anatomy, molecular biology, ...
,
cognitive science Cognitive science is the interdisciplinary, scientific study of the mind and its processes. It examines the nature, the tasks, and the functions of cognition (in a broad sense). Mental faculties of concern to cognitive scientists include percep ...
,
artificial intelligence Artificial intelligence (AI) is the capability of computer, computational systems to perform tasks typically associated with human intelligence, such as learning, reasoning, problem-solving, perception, and decision-making. It is a field of re ...
, and
computer science Computer science is the study of computation, information, and automation. Computer science spans Theoretical computer science, theoretical disciplines (such as algorithms, theory of computation, and information theory) to Applied science, ...
.


Computing

For a processor or computer designed to simulate a neural network SUPS is measured as the product of simulated neurons N and average connectivity c(synapses) per neuron per second: SUPS = c \times N Depending on the type of simulation it is usually equal to the total number of synapses simulated. In an "asynchronous" dynamic simulation if a neuron spikes at \upsilon Hz, the average rate of synaptic updates provoked by the activity of that neuron is \upsilon cN. In a synchronous simulation with step \Delta t the number of synaptic updates per second would be \frac. As \Delta t has to be chosen much smaller than the average interval between two successive afferent spikes, which implies \Delta t < \frac, giving an average of synaptic updates equal to \upsilon c N^2. Therefore, spike-driven synaptic dynamics leads to a linear scaling of computational complexity O(N) per neuron, compared with the O(N2) in the "synchronous" case.


Records

Developed in the 1980s Adaptive Solutions' CNAPS-1064 Digital Parallel Processor chip is a full neural network (NNW). It was designed as a
coprocessor A coprocessor is a computer processor used to supplement the functions of the primary processor (the CPU). Operations performed by the coprocessor may be floating-point arithmetic, graphics, signal processing, string processing, cryptography or ...
to a host and has 64 sub-processors arranged in a 1D array and operating in a
SIMD Single instruction, multiple data (SIMD) is a type of parallel computer, parallel processing in Flynn's taxonomy. SIMD describes computers with multiple processing elements that perform the same operation on multiple data points simultaneousl ...
mode. Each sub-processor can emulate one or more neurons and multiple chips can be grouped together. At 25 MHz it is capable of 1.28  GMAC.''Real-Time Computing: Implications for General Microprocessors''
Chip Weems, Steve Dropsho
After the presentation of the RN-100 (12 MHz) single neuron chip at Seattle 1991
Ricoh is a Japanese multinational imaging and electronics company. It was founded by the now-defunct commercial division of the Institute of Physical and Chemical Research (Riken) known as the ''Riken Concern'', on 6 February 1936 as . Ricoh's hea ...
developed the multi-neuron chip RN-200. It had 16 neurons and 16 synapses per neuron. The chip has on-chip learning ability using a proprietary backdrop algorithm. It came in a 257-pin PGA encapsulation and drew 3.0 W at a maximum. It was capable of 3 GCPS (1 GCPS at 32 MHz). In 1991–97,
Siemens Siemens AG ( ) is a German multinational technology conglomerate. It is focused on industrial automation, building automation, rail transport and health technology. Siemens is the largest engineering company in Europe, and holds the positi ...
developed the MA-16 chip, SYNAPSE-1 and SYNAPSE-3 Neurocomputer. The MA-16 was a fast matrix-matrix multiplier that can be combined to form systolic arrays. It could process 4 patterns of 16 elements each (16-bit), with 16 neuron values (16-bit) at a rate of 800 MMAC or 400 MCPS at 50 MHz. The SYNAPSE3-PC PCI card contained 2 MA-16 with a peak performance of 2560 MOPS (1.28 GMAC); 7160 MOPS (3.58 GMAC) when using three boards.''Neural Network Hardware''
Clark S. Lindsey, Bruce Denby, Thomas Lindblad, 1998
In 2013, the K computer was used to simulate a neural network of 1.73 billion neurons with a total of 10.4 trillion synapses (1% of the human brain). The simulation ran for 40 minutes to simulate 1 s of brain activity at a normal activity level (4.4 on average). The simulation required 1 Petabyte of storage.''Fujitsu supercomputer simulates 1 second of brain activity''
Tim Hornyak, CNET, August 5, 2013


See also

* FLOP * SPECint * SPECfp * Multiply–accumulate operation * Orders of magnitude (computing) *
SyNAPSE In the nervous system, a synapse is a structure that allows a neuron (or nerve cell) to pass an electrical or chemical signal to another neuron or a target effector cell. Synapses can be classified as either chemical or electrical, depending o ...


References

{{Reflist, 30em Benchmarks (computing) Units of frequency Artificial intelligence Computational neuroscience Neurotechnology