Projects

Active Projects

AI/ML Systems

ML Inference Optimization

Designing efficient scheduling strategies for (distributed) inference of large AI models across heterogeneous compute clusters.

Started 2024-07 · Ongoing

AI/ML Systems

Heterogeneous Inference Conputation

Developing optimal partitioning strategies for edge inference of transformer models across heterogeneous compute devices.

Started 2024-12 · Ongoing

Data Centers

Data Center Energy Optimization

Configuration and resource management strategies for achieving optimal energy-performance tradeoffs in large-scale data center networks.

Started 2023-07 · Ongoing

Past Projects

Cloud Computing

Cloud Pricing for AI Workloads

Revenue-maximizing pricing schemes for networked compute nodes serving large-scale AI training and inference jobs in public and private clouds.

2024-06 — 2026-04

Edge Computing

Edge Computing Scheduler

Optimal job scheduling and resource allocation for hierarchical edge-cloud architectures to minimize latency under energy constraints.

2024-07 — 2025-06