Akademska digitalna zbirka SLovenije - logo
E-viri
Recenzirano Odprti dostop
  • Towards extending the SWITC...
    Knight, Louise; Štefanič, Polona; Cigale, Matej; Jones, Andrew C.; Taylor, Ian

    Future generation computer systems, November 2019, 2019-11-00, Letnik: 100
    Journal Article

    SWITCH (Software Workbench for Interactive, Time Critical and Highly self-adaptive cloud applications) allows for the development and deployment of real-time applications in the cloud, but it does not yet support instances backed by Graphics Processing Units (GPUs). Wanting to explore how SWITCH might support CUDA (a GPU architecture) in the future, we have undertaken a review of time-critical CUDA applications, discovering that run-time requirements (which we call ‘wall time’) are in many cases regarded as the most important. We have performed experiments to investigate which parameters have the greatest impact on wall time when running multiple Amazon Web Services GPU-backed instances. Although a maximum of 8 single-GPU instances can be launched in a single Amazon Region, launching just 2 instances rather than 1 gives a 42% decrease in wall time. Also, instances are often wasted doing nothing, and there is a moderately-strong relationship between how problems are distributed across instances and wall time. These findings can be used to enhance the SWITCH provision for specifying Non-Functional Requirements (NFRs); in the future, GPU-backed instances could be supported. These findings can also be used more generally, to optimise the balance between the computational resources needed and the resulting wall time to obtain results. •Runtime is the most important QoS metric for time-critical CUDA applications.•One does not need to use the maximum number of instances available.•It is important that instances are not wasted due to too few input problems.•A balance must be struck between money spent on resources and performance gain.