Foresight News reports that the decentralized AI protocol Prime Intellect has released a preview of its inference stack. The inference stack aims to address challenges in computational utilization during autoregressive decoding, KV cache memory bottlenecks, and public network latency issues. It employs a pipeline parallel design, supporting high computational density and asynchronous execution. Additionally, Prime Intellect has released three open-source codebases: PRIME-IROH (peer-to-peer communication backend), PRIME-VLLM (vLLM integration supporting public network pipeline parallelism), and PRIME-PIPELINE (research sandbox). Users can run large models using GPUs such as the 3090 and 4090.