InferLine: latency-aware provisioning and scaling for prediction serving pipelines
Serving ML prediction pipelines spanning multiple models and hardware accelerators is a
key challenge in production machine learning. Optimally configuring these pipelines to meet …
key challenge in production machine learning. Optimally configuring these pipelines to meet …
Joint task scheduling and containerizing for efficient edge computing
Container-based operation system (OS) level virtualization has been adopted by many edge-
computing platforms. However, for an edge server, inter-container communications, and …
computing platforms. However, for an edge server, inter-container communications, and …
Mashup: making serverless computing useful for hpc workflows via hybrid execution
This work introduces Mashup, a novel strategy to leverage serverless computing model for
executing scientific workflows in a hybrid fashion by taking advantage of both the traditional …
executing scientific workflows in a hybrid fashion by taking advantage of both the traditional …
Daydream: Executing dynamic scientific workflows on serverless platforms with hot starts
HPC applications are increasingly being designed as dynamic workflows for the ease of
development and scaling. This work demonstrates how the serverless computing model can …
development and scaling. This work demonstrates how the serverless computing model can …
How workflow engines should talk to resource managers: A proposal for a common workflow scheduling interface
Scientific workflow management systems (SWMSs) and resource managers together ensure
that tasks are scheduled on provisioned resources so that all dependencies are obeyed …
that tasks are scheduled on provisioned resources so that all dependencies are obeyed …
[PDF][PDF] Inferline: Ml inference pipeline composition framework
The dominant cost in production machine learning workloads is not training individual
models but serving predictions from increasingly complex prediction pipelines spanning …
models but serving predictions from increasingly complex prediction pipelines spanning …
Towards advanced monitoring for scientific workflows
Scientific workflows consist of thousands of highly parallelized tasks executed in a
distributed environment involving many components. Automatic tracing and investigation of …
distributed environment involving many components. Automatic tracing and investigation of …
Scsf: A scheduling simulation framework
High-throughput and data-intensive applications are increasingly present, often composed
as workflows, in the workloads of current HPC systems. At the same time, trends for future …
as workflows, in the workloads of current HPC systems. At the same time, trends for future …
E-hpc: a library for elastic resource management in hpc environments
Next-generation data-intensive scientific workflows need to support streaming and real-time
applications with dynamic resource needs on high performance computing (HPC) platforms …
applications with dynamic resource needs on high performance computing (HPC) platforms …
Starship: Mitigating i/o bottlenecks in serverless computing for scientific workflows
R Basu Roy, D Tiwari - Proceedings of the ACM on Measurement and …, 2024 - dl.acm.org
This work highlights the significance of I/O bottlenecks that data-intensive HPC workflows
face in serverless environments-an issue that has been largely overlooked by prior works …
face in serverless environments-an issue that has been largely overlooked by prior works …