EdgeAI & ML Model Management

Systems challenges in deploying and managing machine learning workloads across heterogeneous edge infrastructure

edge AIdistributed machine learningML model orchestrationedge inferencecontainer partitioningmodel deploymentheterogeneous infrastructure

Research Impact

Publications4
Active Projects1
Team Members5

Machine learning workloads are increasingly deployed at the network edge to meet latency requirements, preserve data privacy, and reduce bandwidth consumption from transmitting raw sensor data to cloud data centers. However, edge infrastructure presents fundamental challenges for ML deployment: devices span an enormous range of computational capabilities from powerful GPU-equipped servers to resource-constrained embedded systems; modern neural network models often exceed the memory and compute capacity of individual edge devices; and dynamic conditions including user mobility, network variability, and resource availability require adaptive deployment strategies. Traditional approaches either deploy full models to powerful edge servers (excluding resource-constrained devices) or run all inference in the cloud (reintroducing latency and bandwidth concerns that motivated edge deployment). Distributing ML workloads across edge-cloud infrastructure while maintaining acceptable inference latency and adapting to heterogeneous device capabilities remains a central systems challenge.

Our research addresses systems challenges in deploying and managing ML workloads at the edge. We develop automatic container partitioning techniques that analyze application structure, identify optimal partition points, and generate separate container images for distributed execution across edge and cloud without requiring manual application restructuring. This enables deploying models that exceed individual device capabilities while reducing wide-area network traffic by processing data locally and transmitting only intermediate representations. We explore orchestration strategies for heterogeneous edge infrastructure that consider device capabilities (CPU, GPU, specialized accelerators), network conditions, application latency requirements, and energy constraints for battery-powered devices. Our work integrates ML deployment with edge orchestration frameworks, providing ML-aware scheduling, model registry and versioning systems, and performance monitoring. We investigate model lifecycle management including deployment optimization (quantization, pruning, knowledge distillation), runtime adaptation to changing conditions, and continuous model updates. Research directions include energy-aware inference strategies, security and privacy in distributed ML deployments, model-hardware co-design for edge devices, and explainability techniques with limited computational resources. Our EdgeAI research connects closely with our edge orchestration work, with techniques integrated into open-source orchestration platforms such as Oakestra.

Publications

JournalEdge Computing

Revisiting Edge AI: Opportunities and Challenges

IEEE Internet Computing
Tobias Meuser
Lauri Lovén
Monowar Bhuyan
Shishir G Patil
Schahram Dustdar
Atakan Aral
Suzan Bayhan
Christian Becker
Eyal de Lara
Aaron Yi Ding
Janick Edinger
James Gross
Nitinder Mohan
Nitinder Mohan
Andy Pimentel
Etienne Rivière
Henning Schulzrinne
Pieter Simoens
Gürkan Solmaz
Michael Welzl
PDFDOICACM
Scholar
No citations yet
IC
Artifacts Available
Artifacts Evaluated Functional
Results Reproduced

Open Source Contributions

2DFS

2DFS

On-Demand Container Partitioning for Distributed Machine Learning

A novel container image distribution system that enables on-demand, file-level container partitioning for distributed machine learning workloads. 2DFS dramatically reduces container startup times and bandwidth consumption by intelligently fetching only the required files when needed.

edgeresearch-prototype

Awards & Recognition

2025EuroSysSystemsContainers

EuroSys Best Poster Award

European Conference on Computer Systems

Awarded at one of the premier systems conferences for novel approach to container image distribution through on-demand file partitioning.

EuroSys Best Poster Award

Project Funding

Active Funding

🇳🇱

6G Future Network Services Growth Fund

Organization: NWO Netherlands
Period: 2025-present
Role: Co-Principal Investigator

Research on next-generation 6G network services and edge computing architectures. Developing innovative solutions for future network infrastructure including intelligent edge orchestration, distributed AI workloads, and sustainable network design.

Learn More

Previous Funding

🇪🇺

EDGELESS - Cognitive Edge Computing

Organization: European Union Horizon 2020
Period: 2022-2024
Role: Co-Principal Investigator

Large-scale EU project on cognitive edge computing for next-generation networks. Developed Oakestra orchestration framework and advanced edge computing concepts. Led to multiple publications and open-source releases.

Learn More

Invited Talks & Panels

View all
ATC

On-Demand Container Partitioning for Distributed Machine Learning

USENIX Annual Technical Conference
Santa Clara, CA, USA
Edge ComputingMachine LearningContainer Orchestration+1 more
Dagstuhl

Edge-AI: Identifying Key Enablers in Edge Intelligence

Dagstuhl Seminar 23432
Edge ComputingAIEdge Intelligence
Dagstuhl

Identifying Key Enablers in Edge Intelligence

Dagstuhl Seminar 21342
Edge ComputingAIEdge Intelligence

Thesis Projects

View all

Available Theses

MASTER

Efficient AI pipeline management in distributed edge infrastructures

Develop an intelligent AI pipeline orchestration system integrated with Oakestra's hierarchical architecture to optimize deployment and execution of machine learning workloads across distributed edge infrastructures.

Learn more →

Recent Completed Theses

BACHELOR

Synthetic 5G Traffic Generation: A Machine Learning Approach

Student: Karsen Cedric van der Deijl

Year: 2025

MASTER

FLOps: Practical Federated Learning via Automated Orchestration (on the Edge)

Student: Alexander Malyuk

Year: 2024

MASTER

Agricultural AI Development on the Edge with NEVONEX

Student: Edward Waterman

Year: 2022