Ask a Question

Prefer a chat interface with context about you and your work?

HawkVision: Low-Latency Modeless Edge AI Serving

HawkVision: Low-Latency Modeless Edge AI Serving

The trend of modeless ML inference is increasingly growing in popularity as it hides the complexity of model inference from users and caters to diverse user and application accuracy requirements. Previous work mostly focuses on modeless inference in data centers. To provide low-latency inference, in this paper, we promote modeless …