You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The computing demands of modern scientific experiments are growing at a faster rate than the performance improvements
31
-
of traditional processors (CPUs). This trend is driven by increasing data collection rates, tightening latency requirements,
31
+
of traditional general-purpose processors (CPUs). This trend is driven by increasing data collection rates, tightening latency requirements,
32
32
and rising complexity of algorithms, particularly those based on machine learning.
33
-
Such a computing landscape strongly motivates the adoption of specialized coprocessors, such as FPGAs, GPUs, and TPUs.
33
+
Such a computing landscape strongly motivates the adoption of specialized coprocessors, such as FPGAs, GPUs, and TPUs. However, this introduces new resource allocation and scaling issues.
34
34
35
35
.. container:: rightside
36
36
@@ -41,8 +41,8 @@ Why "inference-as-a-service"?
41
41
`Image source: A3D3 <https://a3d3.ai/about/>`_
42
42
43
43
44
-
In "inference-as-a-service" model, the data processing workflows ("clients") off-load computationally intensive steps,
45
-
such as neural network inference, to a remote "server" equipped with coprocessors. This design allows to optimize both
44
+
In the inference-as-a-service model, the data processing workflows ("clients") off-load computationally intensive steps,
45
+
such as neural network inference, to a remote "server" equipped with coprocessors. This design allows for optimization of both
46
46
data processing throughput and coprocessor utilization by dynamically balancing the ratio of CPUs to coprocessors.
47
47
Numerous R&D efforts implementing this paradigm in HEP and MMA experiments are grouped under the name
48
48
**SONIC (Services for Optimized Network Inference on Coprocessors)**.
@@ -54,16 +54,16 @@ Numerous R&D efforts implementing this paradigm in HEP and MMA experiments are g
54
54
55
55
-----
56
56
57
-
SuperSONIC: a case for shared server infrastructure
57
+
SuperSONIC: A Case for Shared Server Infrastructure
0 commit comments