Which scaling rule applies to large artificial neural networks

Experience shows that cooperating and communicating computing systems, comprising segregated single processors, have severe performance limitations, which cannot be explained using von Neumann’s classic computing paradigm. In his classic “First Draft,” he warned that using a “too fast processor” vitiates his simple “procedure” (but not his computing model!); furthermore, that using the classic computing paradigm for imitating neuronal operations is unsound. Amdahl added that large machines, comprising many processors, have an inherent disadvantage. Given that artificial neural network’s (ANN’s) components are heavily communicating with each other, they are built from a large number of components designed/fabricated for use in conventional computing, furthermore they attempt to mimic biological operation using improper technological solutions, and their achievable payload computing performance is conceptually modest. The type of workload that artificial intelligence-based systems generate leads to an exceptionally low payload computational performance, and their design/technology limits their size to just above the “toy” level systems: The scaling of processor-based ANN systems is strongly nonlinear. Given the proliferation and growing size of ANN systems, we suggest ideas to estimate in advance the efficiency of the device or application. The wealth of ANN implementations and the proprietary technical data do not enable more. Through analyzing published measurements, we provide evidence that the role of data transfer time drastically influences both ANNs performance and feasibility. It is discussed how some major theoretical limiting factors, ANN’s layer structure and their methods of technical implementation of communication affect their efficiency. The paper starts from von Neumann’s original model, without neglecting the transfer time apart from processing time, and derives an appropriate interpretation and handling for Amdahl’s law. It shows that, in that interpretation, Amdahl’s law correctly describes ANNs.


Introduction
Given the proliferation of ANN-based devices, applications, and methods, furthermore that even supercomputers are re-targeted for AI applications, the efficacy of such systems is gaining growing importance.Von Neumann in his "First Draft" [35] provided an approximation, for (the timing relations of) vacuum tubes only.He warned (in his section 6.3), that because the data transfer time is neglected in his model, using a "too fast processor" vitiates the procedure; furthermore, that using his paradigm for imitating neuronal operations, is unsound, given that the conduction (transfer) time is longer than the synaptic (processing) time.This limitation means that in today's technology background, it is at least doubly unsound to apply his paradigm to describe scaling ANNs1 .However, von Neumann did not provide another procedure that can consider the case corresponding to the today's technology, and the case of neural computing, respectively [47].The question opens, how then networks of computing objects using neuronal operations, mimicking or inspired by biology, can be scaled.
From computational point of view, an ANN is an adaptive distributed processor-based architecture widely used to utilize its inputs and simulate the human-processing in terms of computation, response, and decision-making.To analyze their performance, we start from the "first principles" of computing, scrutinizing the terms and omissions.Amdahl's famous idea for distributed many-processor systems introduced "strong scaling".He predicted [40]: "scaling thus put larger machines [the brain inspired computers built from components designed for Single Processor Approach (SPA) computers] at an inherent disadvantage".It was guessed early [40], that the payload performance of processor assemblies does not scale linearly with the number of processors.However, the appearance of "massively parallel" systems improved the degree of parallelization so much that a new approximation: the "weak scaling" [18] appeared.Due to the peculiarities of its workload, the AI-related progress (including ANNs) shows up much worse scaling [27] than expected, and led to that "Core progress in AI has stalled in some fields" [21].Also, the Gordon Bell Prize jury noticed [8] that "Surprisingly, there have been no brain inspired massively parallel specialized computers [among the winners] ".
All the reasons listed above have their root essentially in neglecting the temporal behavior of computing components and methods.In section 2 we shortly review the considered scaling methods, and some of their consequences.In section 2.1, Amdahl's idea is shortly described (and partly: reinterpreted): his famous formula using our notations is introduced.In section 2.2, we scrutinize the primary purpose of massively parallel processing, Gustafson's idea.Section 3 discusses different factors affecting computing performance of processorbased (as well as some aspects of other electronic equipments) ANN systems.

Common scaling methods
The scaling methods used to model different implementations of parallelized sequential processing (aka "distributed computing") are approximations to the more general model presented in [46,48].As discussed in [45,51], parallelized sequential systems have their inherent performance limitation.Using that formalism and data from the TOP500 database [?], we could estimate performance limits for present supercomputers.It enabled us to comprehend why supercomputers have their inherent performance limit [45].For the accuracy of the scaling method, see the case of supercomputer Piz Daint in [45].We also validated [50] our "time-aware scaling" (as a mostly empirical experience) through applying it, among others, for qualifying load balancing compiler, cloud operation, on-chip communication.Given that experts, with the same background, also build ANN systems, from similar components, we can safely assume that the same scaling is valid for those systems, too.Calibrating our systems for some specific workload (due to the lack of validated data) is not always possible, but one can compare the behavior of systems and draw some general conclusions.

Amdahl's Law
Amdahl's Law (called also 'strong scaling') is usually formulated as where N is the number of parallelized code fragments, α is the ratio of parallelizable fraction to total (so (1 − α) is the "serial percentage"), S is a measurable speedup.That is, Amdahl's Law considers a fixed-size problem, and α portion of the task is distributed to fellow processors.When calculating the speedup, one calculates However, as expressed in [38]: "Even though Amdahl's Law is theoretically correct, the serial percentage is not practically obtainable."That is, concerning S, there is no doubt that it is derived as the ratio of measured execution times, for non-parallelized and parallelized cases, respectively.But, what is the exact interpretation of α, and how can it be used?
Amdahl listed performance affecting factors, such as "boundaries are likely to be irregular; interiors are inhomogeneous; computations required may be dependent on the states of the variables at each point; propagation rates of different physical effects may be quite different; the rate of convergence or convergence at all may be strongly dependent on sweeping through the array along different axes on succeeding passes, etc." Amdahl has foreseen issues with "sparse" calculations (or in general: the role of data transfer ) as well as that the physical size of computer and the interconnection of its computing units (especially in the case of distributed systems) also matters.
Amdahl used wording "the fraction of the computational load ", giving way to his followers to give meaning to that term.This (unfortunately formulated) phrase "has caused nearly three decades of confusion in the parallel processing community.This confusion disappears when processing times are used in the formulations" [38].On one side, it was guessed that Amdahl's Law is valid only for software (for the number of executed instructions), and on the other side other affecting factors, he mentioned but did not discuss in detail, were forgotten.
Expressing Amdahl's speedup is not simple: "For example, if the following percentage is to be derived from computational experiments, i.e., recording the total parallel elapsed time and the parallel-only elapsed time, then it can contain all overheads, such as communication, synchronization, input/output and memory access.The law offers no help to separate these factors.On the other hand, if we obtain the serial percentage by counting the number of total serial and parallel instructions in a program, then all other overheads are excluded.However, in this case, the predicted speedup may never agree with the experiments."[38] Moreover, the experimental one is always smaller than the theoretical one.
From computational experiments, one can express α from Eq. ( 1) in terms measurable experimentally as It is useful to express computing efficiency with those experimentally measurable data.Efficiency is an especially valuable parameter, given that constructors of many parallelized sequential systems (including TOP500 supercomputers) provide the efficiency (as R M ax /R P eak ) of their computing system, and of course, the number of processors N in their system.Via reversing Equ.(4), the value of α ef f can be expressed with measured data as As seen, the efficiency of a parallelized system is a two-parameter function (the corresponding parametrical surface is shown in Fig. 1), demonstratively Fig. 1: The 2-parameter efficiency surface (in function of parallelization efficiency measured by benchmark High Performance Linpack (HPL) and number of processing elements) as concluded from Amdahl's Law (see Eq. ( 4)), in first order approximation.Some sample efficiency values for some selected supercomputers are shown, measured with benchmarks HPL and High Performance Conjugate Gradients (HPCG), respectively.Also, the estimated efficacy of brain simulation using conventional computing is shown.
underpinning that "This decay in performance is not a fault of the architecture, but is dictated by the limited parallelism" [40].Furthermore, that its dependence can be correctly described by the properly interpreted Amdahl's Law, rather than being an unexplained "empirical efficiency".

Gustafson's Law
Partly because of the outstanding achievements of parallelization technology, partly because of issues around practical utilization of Amdahl's Law, a 'weak scaling' (also called Gustafson's Law [18]) was also introduced.Its assumption is that the computing resources grow proportionally with the task size, and the speedup (using our notations) is formulated as Similarly to Amdahl's Law, the efficiency can be derived for Gustafson's Law as (compare to Eq. ( 4)) From these equations immediately follows that speedup (aka parallelization gain) increases linearly with the number of processors, without limitation; a conclusion that was launched amid much fanfare.They imply, however, some more immediate findings, such as the efficiency slightly increases with the number of processors N (the more processors, the better efficacy) -the non-parallelizable portion of the task either shrinks as the number of processors grows, or despite that it is non-parallelizable, the portion (1 − α) is distributed between the N processors executing the extra machine instructions needed to organize the joint work need no time all non-payload computing contributions such as communication (including network transfer), synchronization, input/output and memory access take no time However, an error was made in deriving Eq. ( 6): the N − 1 processors are idle waiting (see the term with subscript 'idle' below) while the first one is executing the sequential-only portion.Because of this, the time that serves as the base for calculating2 the speed up in the case of using N processors is That is, before fixing the arithmetic error, strange conclusions follow, after fixing it, the conceptual mistake comes to light: 'weak scaling' assumes that single-processor efficiency can be transferred to parallelized sequential subsystems without loss.Weak scaling assumes that the efficacy of a system comprising N single-thread processors remains the same as that of a single-thread processor.This assumption strongly contradicts the experienced 'empirical efficiency' (several hundred-fold deviation from its predicted value) of parallelized systems, not mentioning the 'different efficiencies' [45], see also Fig. 1.
However, that "in practice, for several applications, the fraction of the serial part happens to be very, very small thus leading to near-linear speedups" [28], mislead the researchers.Gustafson concluded his "scaling" for several hundred processors only.The interplay of improving parallelization and general hardware (HW) development (including the non-determinism of modern HW [53]), covered for decades that this scaling was used far outside of its range of validity.
That is, Gustafson's Law is simply a misinterpretation of its argument α: a simple function form transforms Gustafson' Law to Amdahl's Law [38].After making that transformation, the two (apparently very different) laws become identical.However, as suspected by [38]: "Gustafson's formulation gives an illusion that as if N can increase indefinitely".Although collective experience showed, that it was not valid for the case of systems comprising an ever higher number of processors (an "empirical efficiency" appeared), and later researchers measured "two different efficiencies" [22] for the same supercomputer (under different workloads), the "weak scaling" was not suspected to be responsible for the issues.
"Weak scaling" omits all non-payload (but needed for the operation) activities, such as interconnection time, physical size (signal propagation time), accessing data in an amount exceeding cache size, synchronization of different kinds, that are undoubtedly present when working with ANNs.In other words, 'weak scaling" neglects the temporal behavior, a crucial science-based feature of computing [46].This illusion led to the moon-shot of targeting to build supercomputers with computing performance well above feasible (and reasonable) size [45] and leads to false conclusions in the case of clouds.Because of this, except some very few neuron systems, "weak scaling" cannot be safely used for ANNs, even as a rough approximation.

Time-aware scaling
The role of α was theoretically established [25], and the phenomenon itself, that the efficiency (in contrast with Eq. ( 7)) decreases as the number of processing units increases, is known since decades [40] (although it was not formulated in the functional form given by Eq. ( 4)).The 'gold rush' for building exascale computers made finally evident, that under the extreme conditions represented by the need of millions of processors, 'weak scaling' leads to false conclusions.It had to be admitted that it "can be seen in our current situation where the historical ten-year cadence between the attainment of megaflops, teraflops, and petaflops has not been the case for exaflops" [16].It looks like, however, that in feasibility studies of supercomputing using parallelized sequential systems, and an analysis, whether building computers of such size is feasible (and reasonable) remained (and remains) out of sight either in USA [43,37] or in EU [14] or in Japan [15] or in China [30].
Fig. 1 depicts the two-parameter efficiency surface stemming from the timeaware interpretation of Amdahl's law.On the parametric surface, described by Eq. ( 4), some measured efficiencies of present top supercomputers are also depicted, only to illustrate some general rules.The HPL 3 efficiencies are sitting on the surface, while the corresponding HPCG4 values are much below those values.The conclusion drawn here was that "the supercomputers have two different efficiencies" [22], because that experience cannot be explained in the frame of 'classic computing paradigm' and/or 'weak scaling'.
Supercomputers T aihulight, F ugaku and K computer stand out from the "millions core" middle group.Thanks to its 0.3M cores, K computer has the best efficiency for HP CG benchmark, while T aihulight with its 10M cores the worst one.The middle group follows the rules [45].For HP L benchmark: the more cores, the lower efficiency.It looks like the community experienced the effect of the two-dimensional efficiency.The top supercomputers run HPL benchmark with using all their cores, but some of them only use a fragment of their cores to measure performance with HPCG.This reduction happens because of the inflection point: as can be concluded from the figure, increasing their nominal performance by an order of magnitude, decreases their efficiency (and so: payload performance) by more than an order of magnitude.For HP CG benchmark: the "roofline" [54] of that communication intensity was already reached, all computers have about the same efficiency.
Beginning with June 2021, the datum "Measured cores" are not provided any more; covering this aspect.The presumable reason of this new trend is that in this way the measures HPCG/HPL ratio gets much higher, providing the illusion that the vast supercomputers became more suitable for real-life tasks.For more discussion on supercomputers, see [45] (and its continuous upgrades or arXiv).
3 Performance limit of processor-based AI systems

General considerations
As discussed in [45], payload performance P (N, α) of parallelized systems comprising N processors is described 5 as where P single is the single-thread performance of individual processors, and α is describing parallelization of the given system for the given workload (i.e., it depends on both of them).This simple formula explains why payload performance of a system is not a linear function of its nominal performance and why in the case of very good parallelization ((1 − α) 1) and low N , this non-linearity cannot be noticed.In contrast with the prediction of 'weak scaling', payload performance and nominal performance differ by a factor, growing with the number of cores.This conclusion is well-known, but forgotten: "This decay in performance is not a fault of the architecture, but is dictated by the limited parallelism" [40].
The key issue is, however, that one can hardly calculate the value of α for the present complex HW/software (SW) systems from their technical data, although some estimated values can be derived.For supercomputers, however, one can derive a theoretical "best possible", and already achieved "worst case" values [45].It gives us reasonable confidence that those values deviate only within a factor of two.We cannot expect similar results for ANNs.There are no generally accepted benchmark computations, and also there are no standard architectures 6 .Using a benchmark means a particular workload, and comparing the results of even a standardized ANN benchmark on different architectures is conceptually as little useful as comparing the results of benchmarks HPL and HPCG on the same architecture.
Recall also, that at a large number of processors, the internal latency of processor also matters.Following the failure of supercomputer Aurora'18, Intel admitted: "Knights Hill was canceled and instead be replaced by a "new platform and new microarchitecture specifically designed for exascale"" [55].We expect that shortly it shall be admitted that building large-scale AI systems is simply not possible based on the old architectural principles [27].The potential new architectures, however, require a new computing paradigm (considering both temporal behavior of computing systems [46], and the old truth that "more is different" [4]), that can give a proper reply to power consumption and performance issues of -among others, ANN -computing.

Communication-to-computation ratio
As we learned decades ago, "the inherent communication-to-computation ratio in a parallel application is one of the important determinants of its performance on any architecture" [40], suggesting that communication can be a dominant contribution to system's non-payload performance.In the case of neural simulation, a very intensive communication must take place, so the nonpayload to payload ratio has a significant impact on the performance of ANNtype computations.That ratio and the corresponding workload type are closely related: using a specific benchmark implies using a specific communicationto-computation ratio.In the case of supercomputing, the same workload is running on (nearly) the same type of architecture, which is not the case for ANNs.Communication is implemented through Input/Output (I/O) instructions; this is why "artificial intelligence, . . .it's the most disruptive workload from an I/O pattern perspective" 7

Computing benchmarks
There are two commonly used benchmarks in supercomputing.Historically, benchmark HPL was used to compare different configurations by executing a standard computing task.However, their performance depends very much on their workload type.Although HPL is excellent for racing purposes (produces high figures), its behavior on vast configurations strongly deviates from that of "real-life" tasks: the surface in Fig. 1 represents a kind of theoretical upper limit for distributed processing.
The HPL class tasks essentially need communication only at the very beginning and at the very end of the job.Real-life programs, however, usually work in a different way.Because of this reason, a couple of years ago, the community introduced benchmark HPCG: the collective experience shows that the payload performance is much more accurately approximated by HPCG than by HPL, because real-life tasks need much more communication than HPL.Importantly, since their interconnection quality improved considerably, supercomputers show different efficiencies when using various benchmark programs [22].Their efficiencies differ by a factor of ca.200-500 (a fact that remains unexplained in the frame of "weak scaling"), when measured by HPL and HPCG, respectively.
In the HPL class, the communication intensity is the lowest possible one: computing units receive their task (and parameters) at the beginning of computation, and they return their result at the very end.The core orchestrating their work must deal with the fellow cores only in these periods, so the communication intensity is proportional to the number of cores in the system.Notice the need to queue requests at the beginning and the end of the task.
In the HPCG class, iteration occurs: the fellow cores return the result of one iteration to the coordinator core, which makes sequential operations: receives and re-sends the parameters, then it needs to compute new parameters before sending them back to the fellow cores.The program repeats the process several times.As a consequence, the non-parallelizable fraction of benchmarking time grows proportionally to the number of iterations.The effect of that extra communication decreases the achievable performance roofline [54]: as shown in Fig. 2, the HPCG roofline is about 200 times smaller than the HPL one.As depicted, for benchmark HPL multiple rooflines can be located.The highest roofline can be attributed to a processor having slightly different computing principle [58].The second highest roofline (dominated by the calculation itself) is the commonly achievable performance gain8 ; their theoretical expectation matches the empirical value [45].The third roofline shows the effect of the improved connection technology.The fourth roofline (dominated mainly by the internal interconnection) shows how much performance can can be achieved without using expensive "racing" parallelization technologies.For benchmark HPCG, only one roofline exists: the calculation dominates.Neither The small black dots mark the performance data of supercomputers JU QU EEN and K as of 2014 June, for HPL and HPCG benchmarks, respectively.The big black dot denotes the payload performance of the system used by [3].The saturation effect can be observed for both HPL and HPCG benchmarks.
direct interconnection of processors nor advanced interconnection technology can increase the performance gain.As discussed below, for another types of calculations (non-standard benchmarks), the roofline can be even lower: [11] measured a value as low as 30 for a certain kind of applications.The "roofline for brain simulation" is guessed from one single measurement datum.That datum was measured using a only a small fragment (about 1%) of available cores (see section 3.9), so the real efficiency (see also section 2.3 on benchmarking supercomputers using only a fragment of their available cores) can be about two orders of magnitude lower, i.e., in the order of a dozen.
Unfortunately, an ANN workload, in general, cannot be defined, so one cannot run one single benchmark to guess the performance characteristics of a new application.The performance gain for ANNs can be guessed to be around Year Ranking Performance gain Fig. 3: History of supercomputing in terms of performance gain: performance values of the first 25 supercomputers, in function of year of their ranking.Data measured using benchmark HPL or above that of the brain simulation 9 ; depending on ANN's architecture.The neural communication, using a vast number of simultaneous communication, combined with the idea of using a single high speed bus, introduces an additional bottleneck: a very different type of workload, see section 3.8.Because of this different workload, the above benchmarks cannot be used to estimate execution time of ANN-type tasks.
As expressed by Eq. ( 8), the resulting performance of parallelized computing systems depends on both single-processor performance and performance gain.To separate these two factors, Fig. 3 displays the performance gain of supercomputers in the function of their year of construction and ranking in the given year.Two 'plateaus' can be localized before the year 2000 and after the year 2010 also, unfortunately; underpinning Amdahl's Law and refuting Gustafson's Law, and also confirming the prediction "Why we need Exascale and why we won't get there by 2020 " [39].The "hillside" reflects the enormous development of interconnection technology between the years 2000 and 2010 (for more details see [45]).For the reason of the "humps" around the beginning of the second plateau, see section 3.5.Unfortunately, different individual factors (such as interconnection quality, using accelerators and clustering, using on-chip memories, or using slightly different computing paradigm, etc.) cannot be separated in this way.However, some limited validity conclusions can be drawn.

Workload type
The role of the workload came to light after that interconnection technology was greatly improved, and as a consequence, the benchmarking computation (defining the type of workload ) became the dominating contributor, defining value of α (and as a consequence, payload performance), for a discussion see [45].The overly complex Fig. 4 illustrates the phenomenon, why and how the payload performance of a configuration depends on the application it runs.Notice that at low nominal performance values, the payload performance depends linearly on the nominal performance (the blue diagram line), and only slightly depends on the workload type.
The performance breakdown shown in Fig. 4 was experimentally measured by [40,11], [23](Fig.7) and [3](Fig.8), but it used not to be a subject of studies.The fact that speedup diagram line turns back at a critical number of processor, was noticed early [40].After exceeding a critical number of cores, housekeeping gradually becomes the dominating factor of the performance limitation, and leads to a decrease in the payload performance: "there comes a point when using more Processing Unit (PU)s . . .actually increases the execution time rather than reducing it".In that paper (at a different workload and architecture) the achievable parallelization gain was about 8, and it was achieved using 20-30 processors.Recently, there are few experimental investigations in this direction.One of the rare exceptions is [11].The careful systematic investigation of results of running bioinformatics applications pointed out that the speedup curve has a maximum, and breaks down for a higher number of processor cores: "The execution time and the speed-up on IPDATA reach the best values within about 90 processors.Furthermore, that . . ."the parallel version is up to 30 times faster than the serial one".For ANNs, it is just a few dozens [27] where "strong scaling" stalls.For different applications (workloads) these figures are of course different, but the conclusion persists.Fig. 4 compares three workloads (owing different communication intensity).In the top and middle figures, the communication intensities of the standard supercomputer benchmarks HP L and HP CG are displayed in the style of AI networks.The "input layer" and "output layer" are the same, and comprise the initiating node only, while the other "layers" are again the same: the rest of the cores.Subfigure 4.C depicts an AI network comprising n input nodes and k output nodes, furthermore the h hidden layers are comprising m nodes.The communication-to-computation intensity [40] is, of course, not proportional in the cases of subfigures, but the figure illustrates excellently how the communication need of different computer tasks changes with the type of the workload.
As can be easily seen from the figure, in the case of benchmark HP L, the initiating node must issue m communication messages and collect m returned results, i.e., the execution time is O(2m).In the case of benchmark HP CG this execution time is O(2N m) where N is the number of iterations (one cannot directly compare the execution times because of the different amounts of computations and the different amounts of sequential-only computations).Fig. 4.A displays the case of minimum communication, and Fig. 4.B a moderately increased one (corresponding to real-life supercomputer tasks).As nominal performance increases linearly and payload performance decreases inversely with the number of cores, at some critical value where an inflection point occurs, the resulting performance starts to fall.The resulting non-parallelizable fraction sharply decreases efficacy (in other words: performance gain or speedup) of the system [24].This effect was noticed early [40], under different technical conditions, but somewhat faded due to development of parallelization technology.
The non-parallelizable fraction (denoted in the figure by α X ef f ) of a computing task comprises components X of different origin.As already discussed, and was noticed decades ago, "the inherent communication-to-computation ratio in a parallel application is one of the important determinants of its performance on any architecture" [40], suggesting that communication can be a dominant contribution to the system's performance.
The workload in ANN systems comprises components of type "computation" and "communication" (this time also involving data access and synchronization, i.e., everything that is 'non-computation').As logical interdependence between those contributions is strictly defined, payload performance of the system is limited by both factors, and the same system (maybe even within the same workload, case by case) can be either computing-bound and communication bound, or both.
Notice that supercomputers showing the breakdown depicted in Fig. 4, are not included in history depicted in Fig. 3. Aurora'18 failed, Aurora'21 semifailed, Gyokou was withdrawn, "Chinese decision-makers decided to withhold the country's newest Shuguang supercomputers even though they operate more than 50 percent faster than the best current US machines" 10 .Also, F ugaku stalled [13] at some 40% of its planned capacity.
Notice that a similar hillside cannot be drawn for benchmark HPCG, because of two reasons.On one side, HPCG measurement started only a few years ago.On the other side, top supercomputers publish data measured with cores less than the number of cores used to measure HPL.Recall that efficiency is a two-parameter function (see Fig. 1).For "real-life" programs, represented by HPCG, this critical number is much lower than in the case of HPL.There is a real competition between the different contributions to dominate system's performance.As demonstrated in Fig. 6 in [45], before 2010, running both benchmarks HPL and HPCG on a top supercomputer was a communicationbound task, since 2010 HPL is a computing-bound task, while HPCG persisted to be a communication-bound task.This is why some supercomputers provide their HPCG efficiency measured only with a fragment of their cores: the HPCG "roofline" is reached at that lower number of cores.Adding more cores does not increase their payload performance, but decreases their efficiency.

Accelerators
As a side-effect of "weak scaling", it is usually presumed that decreasing the time needed for the payload contribution affects the efficiency of ANN systems linearly.However, it is not so.As discussed in detail in [49], we also change the non-payload to payload ratio that defines the system's efficiency.We mention two prominent examples here: using shorter operands (move less data and perform less bit manipulations) and to mimic the operation of a neuron in an entirely different way: using quick analog signal processing rather than slow digital calculation.
It is a common fallacy that benchmark HPL-AI is benchmark for AI systems.Actually, it means "The High Performance LINPACK for Accelerator Introspection" (HPL-AI), and that benchmark seeks to highlight the convergence of HPC and artificial intelligence (AI) workloads. 11It has not much to do with AI, except that it uses the operand length common in AI tasks.HPL, similarly to AI, is a workload type 12 .Researchers succeeded to achieve more than three times better performance gain (3.01 for Summit and 3.42 for F ugaku), that (as correctly stated in the announcement) "Achieving a 445 petaflops mixed-precision result on HPL (equivalent to our 148.6 petaflops DP result)" [20], i.e. the peak DP performance did not change.However, this naming convention suggests the illusion that when using supercomputers for AI tasks and using half-precision, one can expect this payload performance.
Unfortunately, this achievement comes from accessing less data in memory and using quicker operations on shorter operands rather than reducing communication intensity.For AI applications, limitations remain the same as described above; except that when using Mixed Precision, the power efficiency shall be better by a factor of nearly four, compared to the power efficiency measured using double precision operands. 13 We expect that when using half-precision (FP16) rather than double precision (FP64) operands in the calculations, four times less data are transferred and manipulated by the system.The measured power consumption data underpin the statement.However, the system's computing performance is only slightly more than three times higher than using 64-bit (FP64) operands.The non-linearity has its effect even in this simple case (recall that HPL uses minimum communication).In the benchmark, the housekeeping activity (data access, indexing, counting, addressing) also takes time.Concerning the temporal behaviour [46] of the operation, in this case, the data transmission time T t is the same, the data processing time (due to the shorter length of operands) changes, and so the apparent speed changes non-linearly.Even, the measured performance data enabled us to estimate execution time with zero precision 14 (FP0) operands, see [45].
Another plausible assumption is that if we use quick analog signal processing to replace the slow digital calculation, as proposed in [10,1,52], our com-11 https://www.icl.utk.edu/hpl-ai/ 12 Even https://www.top500.org/lists/top500/2020/06/mismatches operand length and workload: "In single or further reduced precision, which are often used in machine learning and AI applications, Fugaku's peak performance is over 1,000 petaflops (1 exaflops)". 13Similarly, exchanging data directly between processing units [58] (without using the global memory) also enhances α (and payload performance) [5], but it represents a (slightly) different computing paradigm. 14Without dedicated measurement, no more accurate estimations are possible puting system gets proportionally quicker.Presumably, on systems comprising just a few neurons, one can measure a considerable, but less than expected, speedup.The housekeeping becomes more significant than in the case of purely digital processing.In a hypothetical measurement, the speedup would be much less than the ratio of the corresponding analog/digital processing times, even in the case of HPL benchmark.Recall that here the workload is of AI type, with much worse parallelization (and non-linearity).As a consequence, one cannot expect a considerable speedup in large neuromorphic systems.For a detailed discussion of introducing new effect/technologies/materials, see [46].
It sounds good that "The analog memristor array is effectively the neural network laid out in the form of a crossbar, which can perform the entire operation in one clock cycle" [26].In brackets, however fairly added, that "(not counting the clock cycles that may be required to fetch and store the input and output data)".Yes, all operands of the memristor array must be transferred to its input section (and previously, they must be computed or otherwise produced), furthermore the results must be transferred from its output section to their destination.The effective computing time of the memristor-related operations shall be compared to conventional operations' effective time, from the beginning to the end of the computing operation, to make a fair comparison.(The problem persists even if continuous-time data representation [52] is used.) The temporal behavior of components and their materials can easily be misidentified in the time-unaware model.Five decades ago, even memristance has been introduced [41] as a fundamental electrical component, meaning that the memristor's electrical resistance is not constant but depends on the history of current that had previously flowed through the device.There are, however, some serious doubts as to whether a genuine memristor can actually exist in physical reality [2].In the light of our analysis, some temporal behavior definitely exists; the question is how much it is related to material or biological features, if our time-aware computing method is followed.
Besides, adding analog components to a digital processor has its price.Given that a digital processor cannot handle resources outside of its world, one must call the operating system (OS) for help.That help, however, is expensive in terms of execution time.The required context switching takes time in the order of executing 10 4 instructions [12,42], which dramatically increases the time of housekeeping and the total execution time.This effect makes the systems' non-payload to payload ratio much worse than it was before introducing that enhancement.

Timing of activities
In ANNs, the data transfer time must be considered seriously.In both biological and electronic systems, both the distance between entities of the network, and the signal propagation speed is finite.Because of this, in physically largesized and/or intensively communicating systems, the "idle time" of processors defines the final performance that a parallelized sequential system can achieve.In conventional computing systems, 'data dependence' limits achievable parallelism: we must compute data before using it as an argument for another computation.Although, of course, also in conventional computing, the computed data must be delivered to the place of their second utilization, thanks to "weak scaling" [18], this "communication time" is neglected.For example, scaling of matrix operations and "sparsity", mentioned in [26], work linearly only if data transfer time is neglected.
Timing plays an important role in all levels of computing, from gatelevel processing to clouds connected to the Internet.In [46,47], the example describing temporal operation of a one-bit adder provides a nice example, that although the line-by-line compiling (sequential programming, called also Neumann-style programming [6]), formally introduces only logical dependence, through its technical implementation it implicitly and inherently introduces a temporal behavior, too.
In neuromorphic computing, including ANNs, the transfer time is a vital part of information processing.A biological brain must deploy a "speed accelerator" to ensure that the control signals arrive at the target destination before the arrival of the controlled messages, despite that the former derived from a distant part of the brain [19].This aspect is so vital in biology that the brain deploys many cells with the associated energy investment to keep the communication speed higher for the control signal.Computer technology cannot speed up communication selectively, as in biology.It is also impossible to keep part of the system for a lower speed selectively: the propagation speed of electromagnetic waves is predefined.However, as discussed in [49], handling data timing adequately, is vital, especially for bio-mimicking ANNs.

The layer structure
The bottom part of Fig. 4 depicts, how ANNs are supposed to operate.The life begins in several input channels (rather than one as in HPL and HPCG cases), that would be advantageous.However, the system must communicate its values to all nodes in the top hidden layer: the more input nodes and the more nodes in the hidden layer(s), the many times more communication is required for the operation.The same situation also happens when the first hidden layer communicates data to the second one, except that here the square of the number of nodes is to be used as a weight factor of communication.
Initially, n input nodes issue messages, each one m messages (queuing#1) to nodes in the first hidden layer, i.e., altogether nm messages.If one uses a commonly used shared bus to transfer messages, these nm messages must be queued (queuing#2).Also, every single node in the hidden layer receives (and processes) m input messages (queuing#3).Between hidden layers, the same queuing is repeated (maybe several times) with mm messages, and finally, km messages are sent to the output nodes.During this process, the system queues messages (at least) three times.Notice that using a single high-speed bus, because of the needed arbitration, drastically increases the transfer time of the individual messages, furthermore changes their timing, see section 3.8.
To make a fair comparison with benchmarks HP L and HP CG, let us assume one input and one output node.In this case, the AI execution time is O(h×m 2 ), provided that the AI system has h hidden layers.(Here we assumed that messaging mechanisms between different layers are independent.It is not so if they share a global bus.) For a numerical example: let us assume that in supercomputers, 1M cores are used.In AI networks, 1K nodes are present in the hidden layers, and only one input and output nodes are used.In that case, all execution times are O(1M ) (again, the amount of computation is sharply different, so the scaling can be compared, but not the execution times).This communication intensity explains why in Fig. 3 the HP CG "roofline" falls hundreds of times lower than that of the HP L: the increased communication need strongly decreases the systems's achievable performance gain.
Notice that the number of computation operations increases with m, while number of communication operations with m 2 .In other words: the more nodes in the hidden layers, the higher is their communication intensity (communicationto-computation ratio), and because of this, the lower is the efficiency of the system.Recall, that since AI nodes perform simple computations, compared to the functionality of supercomputer benchmarks, their communication-tocomputation ratio is much higher, making their efficacy even worse.The conclusions are underpinned by experimental research [27]: -"strong scaling is stalling after only a few dozen nodes" -"The scalability stalls when the compute times drop below the communication times, leaving compute units idle.Hence becoming a communication bound problem."-"the network layout has a large impact on the crucial communication/compute ratio: shallow networks with many neurons per layer . . .scale worse than deep networks with less neurons."

Using high-speed bus(es)
As discussed in connection with the reasoning, why the internal temporal ratio between transporting and processing data has significantly changed [46]: Moore's observation is (was) valid for electronic density only, but not valid for connecting technology, such as buses.On one side, because of the smaller physical size and the quicker clock signal, on the other side the unchanged cmlong bus cable, using a serial bus means spending the overwhelming majority of apparent processing time with arbitration (see the temporal diagram and case study in [46]), so using a sequential bus is at least questionable in largescale systems: the transfer time is limited by the needed arbitration (increases with the number of neurons!) rather than by the bus speed."The idea of using the popular shared bus to implement the communication medium is no longer acceptable, mainly due to its high contention."[33] x The massively "bursty" nature of the data (when imitating biological neural network, the different nodes of the layer want to communicate simultaneously) also makes the case harder.Communication circuits receive the task of sending data to N other nodes.What is worse, bus arbitration, addressing and latency, prolong the transfer time (and decrease the system's efficacy).This type of communicational burst may easily lead to a "communicational collapse" [34], but it may also produce unintentional "neuronal avalanches" [7].
The fundamental issue is replacing the private communication channel between biological neurons with the mandatory use of some kind of shared media in technological neurons.As discussed in [46], at a large number of communicating units, sharing the medium becomes the dominant contributor to the time consumption of computing.Its effect can be mitigated using different technological implementations.Still, the conclusion persists: its technical implementation of neuronal communication defines the payload computational efficiency of an ANN, and the computing performance of its nodes has only marginal importance.The lengthy queueing also leads to irrealistic timings: as discussed in section 3.9, some physically (not biologically or logically) delayed signals must be dropped to provide a seemingly acceptable computing performance.Another (wrong) approach to solving the same problem is introducing conditional computation [9] as discussed in section 3.12.The origin of the issue is that using a shared medium makes the computing system's temporal behavior much more emphasized, and that the temporal behavior cannot be compensated using methods developed having a timeless behavior in mind.
The ANN type workload introduces an extra handicap: the neurons must communicate the result of an elementary operation after every single operation; in addition through a single shared bus.In Fig. 5 the inset shows a simple neuromorphic use case: one input neuron and one output neuron communi-cate through a hidden layer, comprising only two neurons.Fig. 5.A mostly shows the biological implementation: all neurons are directly wired to their partners, i.e., a system of "parallel buses" (axons) exists.Notice that the operating time also comprises two "non-payload" times (T t ): data input and data output, which coincide with the non-payload time of the other communication party.The diagram displays the logical and temporal dependencies of the neuronal functionality.The payload operation ("the computing") can only start after its data is delivered (by the, from this point of view, non-payload functionality: input-side communication), and the output communication can only begin when the computing finished.Importantly, communication and calculation mutually block each other.Two important points that neuromorphic systems must mimic noticed immediately: i/ the communication time is an integral part of the total execution time, and ii/ the ability to communicate is a native functionality of the system.In such a parallel implementation, the performance of the system, measured as the resulting total time (processing + transmitting), scales linearly with increasing either the non-payload communication speed or the payload processing speed.
Fig. 5.B shows a technical implementation of a high-speed shared bus for communication.To the right of the grid, the activity that loads the bus at the given time is shown.A double arrow illustrates the communication bandwidth, the length of which is proportional to the number of packages the bus can deliver in a given time unit.We assume that the input neuron can send its information in a single message to the hidden layer; furthermore, the processing by neurons in the hidden layer both starts and ends simultaneously.However, the neurons must compete for accessing the bus, and only one of them can send its message immediately, the other(s) must wait until the bus gets released.The output neuron can only receive the message when the first neuron completed its sending.Furthermore, the output neuron must first acquire the second message from the bus, and the processing can only begin after having both input arguments.This constraint results in sequential bus delays both during non-payload processing in the hidden layer and payload processing in the output neuron.Adding one more neuron to the layer introduces one more delay.
At this point, two wrong solutions can be taken: either the second neuron must wait until the second input arrives (in biology, a spike also carries a synchronization signal, and triggers its integration), or (in "technical neurons", using continuous levels rather than pulses, this synchronization facility is omitted) changes its output continuously, as the inputs arrive, and its processing speed enables.In the latter case, however, until the second input arrives (and gets processed) the neuron provides an output signal, differing from the one expected based on the mathematical dependence.As discussed in detail in [46], this, temporarily may be wrong, output signal is known in the electronics, and those "glitches" are eliminated via using a "worst-case" delay for the output signal.However, including a serial bus in that computation would enormously prolong the needed "worst-case" delay.
Using the formalism introduced above, T t = 2 • T B + T d + X, i.e., the bus must be reached in time T B (not only the operand delivered to the bus, but also waiting for arbitration: the right to use the shared bus), twice, plus the physical delivery T d through the bus.The X denotes "foreign contribution": if the bus is not dedicated for "neurons in this layer only", any other traffic also loads the bus: both messages from different layers and the general system messages may make processing slower (and add their contribution to faking the imitated biological effect).
Even if only one single neuron exists in the hidden layer, it must use the mechanisms of sharing the bus, case by case.The physical delivery to the bus takes more time than a transfer to a neighboring neuron (both the arbiter and the bus are in cm distance range, meaning several nsec transfer times, while the direct transfer between the connected gates may be in the psec range).If we have more neurons (such as a hidden layer) on the bus and work in parallel, they must all wait for the bus.The high-speed bus is very slightly loaded when only a couple of neurons are present.Its load increases linearly with the number of neurons in the hidden layer (or, maybe, all neurons in the system).The temporal behavior of the bus, however, is different.
Under a biology-mimicking workload, the second neuron must wait for all its inputs originating in the hidden layer.If we have L neurons in the hidden layer, the transmission time of the neuron behind the hidden layer is T t = L • 2 • T B + T d + X.This temporal behavior explains why "shallow networks with many neurons per layer . . .scale worse than deep networks with less neurons" [27]: the physical bus delivery time T d , as well as the processing time T p , become marginal if the layer forces to make many arbitrations to reach the bus: the number of the neurons in the hidden layer defines the transfer time.In deeper networks, the system sends its messages at different times in its different layers (and, even they may have independent buses between the layers), although the shared bus persists in limiting the communication.Notice that there is no way to organize the message traffic: only one bus exists.
At this point comes into picture the role of the workload on the system: the two neurons in the hidden layer want to use the single shared bus, at the same time, for communication.As a consequence, the apparent processing time is several times higher, than the physical processing time, and it increases linearly with the number of neurons in the hidden layer (and, maybe, with also the total number of neurons in the system, if a single high-speed bus is used).
The ratio of the time spent with forwarding data on the high-speed bus gradually decreases as the system's size increases.In vast systems, especially when attempting to mimic neuromorphic workload, the speed of the bus is getting marginal.Notice that the times shown in the figure are not proportional: the (temporal) distances between cores are in the several picoseconds range, while the bus (and the arbiter) are at a distance well above nanoseconds, so the actual temporal behavior (and the idle time stemming from it) is much worse than the figure suggests."The idea of using the popular shared bus to implement the communication medium is no longer acceptable, mainly due to its high contention."[33].The extraordinary workload of AI, makes it much harder to operate the systems.
3.9 The "quantal nature of computing time" One of the famous cases demonstrating existence and competition of those limitations in the fields of AI is the research published in [3].The systems used in the study were a HW simulator [17] explicitly designed to simulate 10 9 neurons (10 6 cores and 10 3 neurons per core) and many-thread simulation running on a supercomputer [29] able to simulate 2 • 10 8 neurons (the authors mention 2 • 10 3 neurons per core and supercomputers having 10 5 cores), respectively.The experience, however, showed [3] that scaling stalled at 8 • 10 4 neurons, i.e., about four orders of magnitude less than expected.They experienced stalling about the same number of neurons, for both the HW and the SW simulator.
Given that supercomputers have a performance limit [45], one can comprehend the former experience: the brain simulation needs massive communication (the authors estimated that ≈ 10% of the execution time was spent with non-payload activity), that sharply decreases their achievable performance, so their system reached the maximum payload performance that their (1 − α) enables: the sequential portion was too high.But why the purpose-built brain simulator cannot reach its maximum expected performance?Is it just an accident that they both stalled at the same value, or some other limiting factor came into play?Paper [44] gives the detailed explanation.
The short reply is that digital systems, including brain simulators, have a central clock signal representing an inherent performance limit: no action in the system can happen in a shorter time.The total time divided by the clock period's length defines maximum performance gain [45] of a system.If the size of the clock period is the commonly used 1 ns, and measurement time (in the case of supercomputers) is in the order of several hours, clocking does not mean a limitation.
Computational time and biological time are not only not equal, but they are also not proportional.To synchronize the neurons periodically, a "time grid", commonly with 1 ms integration time, was introduced.The systems use this grid time to put the free-running artificial neurons back to the biological time scale, i.e., they act as a clock signal: simulation of the next computation step can only start when this clock signal arrives.This action is analogous with introducing a clock signal for executing machine instructions: the processor, even when it is idle, cannot begin the execution of its next machine instruction until this clock signal arrives.That is, in this case, the clock signal is 10 6 times longer than the clock signal of the processor.Just because neurons must work on the same (biological) time scale, when using this method of synchronization, the (commonly used) 1 millisecond "grid time" has a noticeable effect on the payload performance. 15he brain simulation measurement [3] enables us to guess the efficacy of ANNs.Given that using more cores only increased the nominal performance (and, correspondingly, its power consumption) of their system, the authors decided to use only a small fragment of their resources, only 1% of the cores available in the HW simulator.In this way, we can place the efficiency of brain simulation on a supercomputer benchmarking scale, see Fig. 2.Under those circumstances, as witnessed by Fig. 2, a performance gain about 10 3 was guessed for brain simulation.Notice that the large-scale supercomputers use about 10% of their cores in HPCG measurement, see also Fig. 1.The difference in the efficiency values of HPL, HPCG, and the Brain comes from the different workloads, which is the reason for the issue.For ANNs the efficiency can be similar to that of brain simulation, somewhat above the performance gain of HPCG, because the measurement time is shorter than that of HPCG on supercomputers, so the corresponding inherent non-parallelizable portion is higher.
Recall also the "communicational collapse" from the previous section: even if communication packages are randomized in time, it represents a colossal peak traffic, mainly if a single global (although high speed) bus is used.This effect is so strong in large systems that emergency measures must have been introduced, see section 3.12.In smaller ANNs, it was found [3], that only a few dozens of thousands of neurons can be simulated on processor-based brain simulators.This experience includes both many-thread software simulators and a purpose-built brain simulator 16 .Recall also from [27], that "strong scaling is stalling after only a few dozen nodes".For a discussion on the effect of serial bus in ANNs, see [46].

Rooflines of ANNs
As all technical implementations, computing also has technological limitations.The "roofline" model [54] successfully describes, that until some needed resource exceeds its technical limitation, utilization of that resource shows a simple linear dependency.Exceeding that resource is impossible: usage of the resource is stalling at the maximum possible level; the two lines form a "roofline".In a complex system, such as a computing system, the computing process uses different resources, and under other conditions, various resources may dominate in defining the "roofline(s) of computing process", see Fig. 2.An example is running benchmarks HPL and HPCG: as discussed in [45], either computing or interconnection dominates the payload performance.
As section 3.9 discusses, in some cases, a third competitor can also appear on the scene, and even it can play a significant role.That is, it is not easy at all to describe an ANN system in terms of the "roofline" [54] model: depending on the actual conditions, the dominant player (the one that defines the top level of the roofline) may change.Anyhow: it is sure that the contribution of the component, representing the lowest roofline, shall dominate.Still, the competition of parts may result in unexpected issues (for example, see how computation and interconnection changed their dominating rule, in [45]).Because of this, Fig. 2 has limited validity.It provides, however, a feeling that 1/ for all workflow types a performance plateau exists and already reached; 2/ what value of payload performance gain value can be achieved for different workloads; 3/ where the payload efficiency of the particular kinds of ANNs, brain simulation on supercomputers, are located compared to those of the standard benchmarks (a reasoned guess).

Role of parameters of computing components
As discussed in [45], different components of computing systems mutually block each other's operation.Datasheet parameters of components represent a hard limit, valid for ideal, stand-alone measurements, where the utilization is 100%.When they must cooperate with other components, the way as they cooperate (aka the workload of the system) defines their soft limit (degrades utilization of units): until its operand(s) delivered, computing cannot start; until computed, result transmission cannot start, mainly when several computing units compete for the shared medium.This competition is the reason why ANNs represent a very specific workload, where weaknesses of principles of computing systems are even more emphasized.

Training ANNs
One of the most shocking features of ANNs is their weeks-long training time, even for (compared to the functionality of brain) simple tasks.The mathematical methods, of course, do not comprise time-dependence, the technical implementation of ANNs, however, does: as their time-dependence is discussed in details in [46], the delivery times of new neuronal outputs (that serve as new neuronal inputs at the same time) are only loosely coupled: the assumption that producing an output means at the same time producing an input for some other neuron, works only in the timeless "classic computing" (and in biology using parallel axons), see discussing the temporal behavior of the serial bus in [46].
To comprehend what change of considering temporal behavior means, consider the temporal diagram of a 1-bit adder in [46].When using adders, we have a fixed time when we read out the result.We are not interested in "glitches", so we set a maximum time until all bits relaxed, and (at the price of losing some performance), we will receive the final result only; the adder is synchronized.
The case of ANNs, however, is different.In the adder, outputs of bit n are the input of bit n + 1 (there is no feedback), and the bits are wired directly.In ANNs, the signals are delivered via a bus, and the interconnection type and sequence depends on many factors (ranging from the kind of task to the actual inputs).During training ANNs, their feedback complicates the case.The only fixed thing in timing is that the neuronal input arrives inevitably only after a partner produced it.The time ordering of delivered events, however, is not sure: it depends on technical parameters of delivery, rather than on the logic that generates them.Time stamping cannot help much.There are two bad choices.Option one is that neurons should have a (biological) sending time-ordered input queue and begin processing when all partner neurons have sent their message.That needs a synchronous signal and leads to severe performance loss (in parallel with the one-bit adder).Option two is that they have a (physical) arrival-time ordered queue, and they are processing the messages as soon as they arrive.This technical solution enables us to give feedback to a neuron that fired later (according to its timestamp), and set a new neuronal variable state; which is a "future state" when processing a message received physically later, but with a timestamp referring to a biologically earlier time.A third, maybe better, option would be to maintain a biological-time ordered queue, and either in some time slots (much shorter than the commonly used "grid time") send out output and feedback, individually process the received events, and send back feedback and output immediately.In both cases, it is worth to consider if their effect is significant (exceeds some tolerance level compared to the last state) and mitigate the need for communication also in this way.
We start showing an input during training, and the system begins to work, using the synaptic weights valid before showing that input.Those weights may be randomized, or maybe that they correspond to the previous input data.The signals that the system sends, are correct, but a receiver does not know the future: it processes a signal only after it was physically delivered 17 , meaning that it (and its dependents) may start to adjust their weights to a state that is still undefined.In [46], the first AND gate has quite a short indefinite time, but the OR has a long one.
When playing chess against its opponent, a faster computer can be advantageously used to analyze the forthcoming moves.Even it can compute all future moves before its opponent makes the next move.However, with publishing its next move, it must wait the next move made by its opponent, otherwise it may publish a wrong move.Sending the feedback to its opponent as soon as the next move is computed -i.e., without synchronization-results in a quickly computed but maybe wrong move.Without synchronization, the faster is the computer, the worse is its performance as chess player.
At the beginning of their operation, some component neurons of the network may have undefined states and weights.Their operation is essentially an iteration, where -without synchronizationthe actors mostly use mostly wrong input signals, and surely adjust their weights to false signals initially and with significant time delay at later times.If we are lucky (and consider that we are working with unstable states in the case of more complex systems), the system will converge, but painfully slowly.Or not at all.Not considering the temporal behavior of the network leads to painfully slow and doubtful convergence.
Synchronization is a must, even in ANNs.We must take care when using accelerators, feedback and recurrent networks.The time matters.Computing neuronal results faster, to provide feedback more quickly, cannot help much, if at all.Delivering feedback information also needs time and uses the same shared medium, with all its disadvantages.In biology, the "computing time" and the "communication time" are in the same order of magnitude.In computing, the communication time is very much longer than computation, that is, the received feedback refers to a time (and the related state variables) that was valid a very long time ago.In biology, spiking is also a "look at me" signal: the feedback shall be sent to that neuron, reflecting the change its output caused 18 .Without this, neurons receive feedback about "the effect of all fellow neurons, including me".Receiving a spike defines the time of the beginning of signal's validity; "leaking" also defines their "expiration time".When using spiking networks, their temporal behavior is vital.
In excessive systems, some result/feedback events must be dropped because of long queuing to provide seemingly higher performance.The logical dependence that the feedback is computed from the results of the neuron that receives the feedback, the physical implementation of the computing system converts to time dependence [46].Because of this time sequence, the feedback messages will arrive at the neuron later (even if at the same biological time, according to their time stamp they carry), so they stand at the end of the queue.Because of this, it is highly probable that they "are dropped if the receiving process is busy over several delivery cycles" [3].In vast systems, the feedback in the learning process involves results based on undefined inputs, and the calculated and (maybe correct) feedback may be neglected.
An excellent "experimental proof" of the claims above is provided in [9].With the words of that paper: "Yet the task of training such networks remains a challenging optimization problem.Several related problems arise: very long training time (several weeks on modern computers, for some problems), the potential for over-fitting (whereby the learned function is too specific to the training data and generalizes poorly to unseen data), and more technically, the vanishing gradient problem"."The immediate effect of activating fewer units is that propagating information through the network will be faster, both at training as well as at test time."This effect also means that the computed feedback, based maybe on undefined inputs, reaches the previous layer's neurons faster.A natural consequence is that (see their Fig.5): "As λ s increases, the running time decreases, but so does performance."Similarly, introducing the spatiotemporal behavior of ANNs, even in its simple form, using separated (i.e., not connected in the way proposed in [46]) time and space contributions to describe them, significantly improved the efficacy of video analysis [56].
The role of time (mismatching) is confirmed directly, via making investigations in the time domain."The CNN models are more sensitive to low- 18 See the Hebbian learning: the neuron uses its inputs and output, exclusively.

Fig. 2 :
Fig.2: Performance gain of supercomputers in function of their year of construction, under different workloads.The diagram lines display the measured values derived using HPL and HPCG benchmarks, for the TOP3 supercomputers in the gives years.The small black dots mark the performance data of supercomputers JU QU EEN and K as of 2014 June, for HPL and HPCG benchmarks, respectively.The big black dot denotes the payload performance of the system used by[3].The saturation effect can be observed for both HPL and HPCG benchmarks.

4 RPFig. 4 :
Fig. 4: Different communication/computation intensities of the applications lead to different payload performance values in the same supercomputer system.Left column: models of computing intensities for different benchmarks.Right column: the corresponding payload performances and α contributions in function of the nominal performance of a fictive supercomputer (P = 1Gf lop/s @ 1GHz).The blue diagram lines refer to the right hand scale (R M ax values), all others ((1 − α X ef f ) contributions) to the left hand scale.The figure is purely illustrating the concepts; the displayed numbers are somewhat similar to the real ones.

Fig. 5 :
Fig.5: Implementing neuronal communication in different technical approaches.For legend's details see text.A (the biological implementation): the parallel bus; B and C(the technical implementation): the shared serial bus, before and after reaching the communication "roofline"[54].