A surprising number of advanced artificial intelligence language models are straining under the immense computational demands of processing long user requests. To address this, Moonshot AI and researchers from Tsinghua University have jointly proposed PrfaaS, a novel architecture designed to serve these large language models (LLMs) more efficiently by distributing tasks across multiple data centers. This innovation aims to overcome current bottlenecks by intelligently separating different stages of the AI processing pipeline.
Breaking Down LLM Inference for Broader Reach
The PrfaaS architecture introduces a strategic division of labor for LLM operations. It selectively offloads the computationally intensive ‘prefill’ phase, which handles the initial processing of long contexts, to specialized, compute-dense clusters. The critical ‘KVCache’ data, essential for subsequent processing, is then transferred over standard Ethernet to separate clusters optimized for the ‘decode’ phase. According to technical documentation, this separation was demonstrated in a case study using an internal 1T-parameter hybrid model, which achieved 54% higher serving throughput than a homogeneous baseline setup.
This new approach significantly improves efficiency. PrfaaS achieved 32% higher throughput compared to a naive heterogeneous setup and, when comparing systems at equal hardware cost, demonstrated approximately 15% greater throughput. The underlying technology enabling this relies on hybrid attention models, which reduce the KVCache size by intelligently interleaving fully attentive layers with more efficient linear-complexity layers. This reduction in KVCache size is crucial for making inter-datacenter transfers feasible.
The Networked Intelligence Challenge
The core of PrfaaS is its sophisticated routing and transport system, designed to manage the KVCache transfer over commodity Ethernet. The architecture utilizes length-based threshold routing to direct incoming requests to either local Processing & Decoding (PD) clusters or remote PrfaaS clusters. To ensure reliability, PrfaaS employs layer-wise prefill pipelining, multi-connection TCP transport, and proactive congestion monitoring.
A dual-timescale scheduler plays a critical role, monitoring network traffic and queue depths at short intervals to adjust routing dynamically as links approach their bandwidth limits. This scheduler also handles cache-affine routing, optimizing transfers based on available cache prefixes. At longer timescales, it rebalances the prefill and decode node counts within the local PD cluster. In one configuration, a PrfaaS cluster of 32 H200 GPUs was paired with a local PD cluster of 64 H20 GPUs, utilizing approximately 100 Gbps of cross-cluster bandwidth, with an aggregate egress load of about 13 Gbps under optimal conditions. According to technical documentation, at the scale of a 10,000-GPU datacenter, the aggregate egress bandwidth required for KVCache transfer totals about 1.8 Tbps.
📊 Key Numbers
- Serving throughput increase (vs homogeneous baseline): 54%
- Serving throughput increase (vs naive heterogeneous setup): 32%
- Throughput gain (vs homogeneous baseline at equal hardware cost): ~15%
- Full PrfaaS-PD system throughput (vs homogeneous baseline on 1T-parameter hybrid model): 1.54x
- Naive heterogeneous configuration throughput (vs homogeneous baseline on 1T-parameter hybrid model): 1.16x
- Mean Time to First Token (TTFT) reduction (vs homogeneous baseline): 50%
- P90 TTFT reduction (vs homogeneous baseline): 64%
- Cross-cluster bandwidth: 100 Gbps
- Aggregate PrfaaS egress load (optimal configuration): ~13 Gbps
- Estimated aggregate egress bandwidth required at 10,000-GPU scale: 1.8 Tbps
🔍 Context
This announcement addresses the growing challenge of serving large language models (LLMs) with extended context windows, a limitation that has previously constrained their real-world applications. The PrfaaS architecture by Moonshot AI and Tsinghua University represents a significant shift towards disaggregated LLM inference, moving away from single-datacenter, tightly integrated systems. This development aligns with a broader industry trend of seeking more cost-effective and scalable AI infrastructure solutions, especially as model sizes continue to balloon.
A direct market rival to this approach could be considered NVIDIA’s strategies for unified, high-performance computing within single data centers, often leveraging RDMA networks for low-latency communication. NVIDIA’s advantage lies in its established ecosystem and specialized hardware like the Rubin CPX for prefill and LPU-style chips for decode, offering a potentially simpler, albeit less distributed, solution. The timeliness of PrfaaS is underscored by the rapid advancements in hybrid attention models and the increasing need to optimize inference costs, a critical concern for AI deployments over the last six months.
💡 AIUniverse Analysis
Our reading: The PrfaaS architecture offers a compelling vision for overcoming LLM serving bottlenecks by strategically distributing prefill and decode tasks across data centers. The genuine advance lies in the clever use of hybrid attention models to shrink the KVCache, making inter-datacenter transfers over commodity Ethernet a practical, albeit complex, possibility. The resulting performance gains, particularly the 50% reduction in TTFT, demonstrate the potential for significant efficiency improvements in handling long-context requests.
However, the shadow cast by this innovation is the inherent complexity and reliance on the stability of inter-datacenter networking. While Ethernet is ubiquitous, its latency and bandwidth can be more variable than dedicated, within-datacenter RDMA fabrics. The success of PrfaaS hinges on the robustness of its multi-connection TCP transport and congestion monitoring to mask these network uncertainties. Furthermore, while the system achieves higher throughput, the managerial overhead of orchestrating distributed prefill and decode clusters across potentially geographically dispersed locations might offset some of the economic benefits for many organizations, especially when compared to consolidated, highly optimized single-datacenter solutions.
For PrfaaS to truly matter in twelve months, there must be widespread evidence of its successful, large-scale deployment in production environments, demonstrating consistent reliability and clear cost advantages over existing alternatives. This would require robust tooling for deployment and management, alongside strong assurances regarding network performance and security between datacenters.
⚖️ AIUniverse Verdict
Promising. The proposed PrfaaS architecture demonstrates a clever method for improving LLM serving throughput by distributing prefill and decode across datacenters, evidenced by a 54% throughput gain in a case study, but its practical adoption hinges on overcoming the complexities of inter-datacenter networking.
🎯 What This Means For You
Founders & Startups: Founders can explore building specialized LLM serving infrastructure that leverages distributed datacenters for cost-effective scaling, targeting workloads with long contexts.
Developers: Developers need to understand and implement new routing logic and multi-connection transport mechanisms to utilize distributed prefill and decode architectures effectively.
Enterprise & Mid-Market: Enterprises can potentially reduce infrastructure costs and improve serving throughput by disaggregating LLM inference across multiple datacenters.
General Users: End-users may experience faster response times for applications requiring long context processing in LLM-powered services.
⚡ TL;DR
- What happened: Moonshot AI and Tsinghua University unveiled PrfaaS, an architecture that splits LLM processing across datacenters to improve efficiency.
- Why it matters: It promises higher serving throughput and faster responses for models handling long text inputs by intelligently distributing prefill and decode tasks.
- What to do: Watch for industry adoption and independent benchmarks on the reliability and cost-effectiveness of this distributed model serving approach.
📖 Key Terms
- PrfaaS
- A cross-datacenter architecture designed for serving large language models by separating prefill and decode phases.
- KVCache
- A crucial data structure used in LLM inference that stores key and value states from attention layers, essential for generating subsequent tokens.
- Prefill
- The initial phase of LLM inference where the model processes the entire input prompt to generate the first token.
- Decode
- The subsequent phase of LLM inference where the model generates output tokens one by one, conditioned on previously generated tokens.
- Hybrid Attention
- A technique that combines different attention mechanisms within an LLM, such as full attention and linear complexity attention, to optimize performance and reduce memory usage.
- Commodity Ethernet
- Standard networking hardware, such as Ethernet cables and switches, commonly used in data centers and local area networks, as opposed to more specialized high-performance interconnects.
- RDMA
- Remote Direct Memory Access, a high-performance networking technology that allows direct memory access between computers without involving the operating system, typically used for low-latency communication in supercomputing and high-performance clusters.
Analysis based on reporting by MarkTechPost. Original article here.

