TrimCaching: Parameter-Sharing AI Model Caching in Wireless Edge Networks
Next-generation mobile networks are expected to facilitate fast AI model downloading to end users. By caching models on edge servers, mobile networks can deliver models to end users with low latency, resulting in a paradigm called edge model caching. In this paper, we develop a novel model placement...
Saved in:
Published in | Proceedings of the International Conference on Distributed Computing Systems pp. 36 - 46 |
---|---|
Main Authors | , , , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
23.07.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Be the first to leave a comment!