Ask a Question

Prefer a chat interface with context about you and your work?

Sequoia: Scalable, Robust, and Hardware-aware Speculative Decoding

Sequoia: Scalable, Robust, and Hardware-aware Speculative Decoding

As the usage of large language models (LLMs) grows, performing efficient inference with these models becomes increasingly important. While speculative decoding has recently emerged as a promising direction for speeding up inference, existing methods are limited in their ability to scale to larger speculation budgets, and adapt to different hyperparameters …