Skip to content
BY 4.0 license Open Access Published by De Gruyter 2023

8 Communication Awareness

From the book Volume 1 Machine Learning under Resource Constraints - Fundamentals

  • Kuan-Hsun Chen , Junjie Shi , Henning Funke and Jens Teubner

Abstract

With the increasing demand for time-predictable machine learning applications, e.g., object detection in autonomous driving systems, such a trend poses several new challenges for resource synchronization in real-time systems, especially when hardware accelerators like Graphics Processing Units (GPUs) are considered as shared resources. When the shared resources have relatively high utilization, conventional synchronization mechanisms might result in performance downgrade. We thus propose the emerging Dependency Graph Approach (DGA), where the precedence constraints of all the computation segments are pre-proceeded. Such a non-workconserving approach can schedule long critical sections, which may be even longer than the period of another task. This is not the case in all the other work-conserving protocols typically in use. Throughout numerical experiments, we show that DGA outperforms all the other conventional protocols in all the evaluated configurations when shared resources are highly utilized. Additionally, a system does not always have sufficient resources at its disposal to perform the necessary operations fast enough if machine learning workloads are executed merely locally. One sound approach is to offload heavy workload to powerful remote servers and expect the inference outcome can be received in time. However, since this approach highly depends on network connectivity and responsiveness, typically only non-critical tasks are offloaded, whose timing requirements are less strict than those of critical tasks. Against such a pessimistic design, we present two novel offloading protocols that offload both critical and non-critical tasks. They handle uncertain connections while providing certain timing guarantees. To achieve a timing-predictable design, typical timing analyses always consider the worst-case execution pattern to derive timing guarantees. But this approach is often too restrictive for some machine learning applications with soft timing constraints. To mitigate the pessimism, we develop several timing analyses of the probability of deadline misses and the deadline miss rate, two important metrics considered in the literature to quantify timeliness.

In this section, we look at distributed processing on a smaller scale. Even a single-machine system today internally looks-and behaves-like a distributed system: multiple processing modules of different flavors (e.g., CPUs, GPUs, FPGAs) interact with memory modules, which are scattered over the system, through an interconnect that is comprised of, say, QPI, PCIe, or “real” network links. In such environments, communication quickly becomes the limiting factor-not only for observable performance, but also for other system aspects, such as energy consumption. We specifically look at communication patterns in heterogeneous CPU/GPU environments, and we illustrate how novel processing models can minimize communication overhead in such systems, which in turn results in significant performance improvements for real-world settings. In GPU-accelerated, data-intensive systems, the PCIe link is often perceived as the limiting factor. Oftentimes, successions of fine-granular GPU kernel invocations amplify the problem, since they tend to cause multiple round-trips over the bandwidth-limited link. As we see in this section, unnecessary round-trips can be avoided by fusing finegranular kernels into larger work units that can hide costly PCIe transfer times (query compilation can be a device to implement kernel fusion). Eliminating the PCIe bottleneck, however, only exposes the GPU’s on-chip communication as the new bottleneck to GPU-assisted data processing. Here, the data-parallel processing modes of graphics processors and the synchronization between parallel units are the cause for redundant round-trips over the precious on-chip communication interfaces. These bottlenecks can be avoided when processing models are deliberately designed to be communication aware. A specific example is a novel combination of pipelining/streaming processing models with the data-parallel nature of GPUs, which aligns particularly well with the semantics of database query execution. For real-world settings, this results in a reduction of memory access volumes by factors of up to 7.5× and shorter GPU kernel execution times by factors of up to 9.5×.

© 2022 Walter de Gruyter GmbH, Berlin/Boston
Downloaded on 16.11.2024 from https://www.degruyter.com/document/doi/10.1515/9783110785944-008/html
Scroll to top button