Loading video player...
Perplexity AI Multilingual Open-Weight Retrieval Models In February 2026, Perplexity released pplx-embed, a new family of open-source multilingual embedding models designed to optimize web-scale information retrieval. By converting Alibaba’s Qwen3 architecture from a standard causal decoder into a bidirectional encoder using diffusion-based pretraining, these models more effectively capture global context from noisy web data. The release features two specialized versions: pplx-embed-v1 for standalone queries and pplx-embed-context-v1 for document chunks in Retrieval-Augmented Generation (RAG) systems. These models outperform competitors like Google’s Gemini and Alibaba’s Qwen3-Embedding while using significantly less memory due to native INT8 and binary quantization.