[2412.03594] BatchLLM: Optimizing Large Batched LLM Inference with Global Prefix Sharing and Throughput-oriented Token Batching
Summary
Abstract page for arXiv paper 2412.03594: BatchLLM: Optimizing Large Batched LLM Inference with Global Prefix Sharing and Throughput-oriented Token Batching
Description
Abstract page for arXiv paper 2412.03594: BatchLLM: Optimizing Large Batched LLM Inference with Global Prefix Sharing and Throughput-oriented Token Batching
Original reporting
AFBytes is a read-only aggregator. Use the original source for full context and complete reporting.
Open original source