All Stories

Simplify Search with Multilingual Embedding Models

This blog post presents and shows how to represent a robust multilingual embedding model of the E5 family in Vespa.

Vespa Newsletter, July 2023

Photo by Ilya Pavlov on Unsplash

Vespa Newsletter, July 2023

Advances in Vespa features and performance include Vector Streaming Search, GPU accelerated embeddings, Huggingface models and a solution to MIPS using a nearest neighbor search.

Leveraging frozen embeddings in Vespa with SentenceTransformers

How to implement frozen embeddings approach in Vespa using SentenceTransformers library and optimize your search application at the same time.

Announcing Maximum Inner Product Search

Vespa can now solve Maximum Inner Product Search problems using an internal transformation to a Nearest Neighbor search. This is enabled by the new dotproduct distance metric.

Announcing vector streaming search: AI assistants at scale without breaking the bank

With personal data, you need complete results at low cost, something vector databases cannot provide. Vespa's new vector streaming search delivers complete results at a fraction of the cost.

Vespa at Berlin Buzzwords 2023

Summarizing Berlin Buzzwords 2023, Germany’s most exciting conference on storing, processing, streaming and searching large amounts of digital data.

Enhancing Vespa’s Embedding Management Capabilities

Photo by vnwayne fan on Unsplash

Enhancing Vespa’s Embedding Management Capabilities

We are thrilled to announce significant updates to Vespa’s support for inference with text embedding models that maps texts into vector representations.

Vespa Newsletter, May 2023

Photo by Scott Graham on Unsplash

Vespa Newsletter, May 2023

Advances in Vespa features and performance include multi-vector HNSW Indexing, global-phase re-ranking, LangChain support, improved bfloat16 throughput, and new document feed/export features in the Vespa CLI.

High performance feeding with Vespa CLI

Photo by Shiro hatori on Unsplash

High performance feeding with Vespa CLI

Vespa CLI can now feed large sets of documents to Vespa efficiently.

Vespa support in langchain

Langchain now comes with a Vespa retriever.

Minimizing LLM Distraction with Cross-Encoder Re-Ranking

Announcing global-phase re-ranking support in Vespa, unlocking efficient re-ranking with precise cross-encoder models. Cross-encoder models minimize distraction in retrieval-augmented completions generated by Large Language Models.

Customizing Reusable Frozen ML-Embeddings with Vespa

Photo by fabio on Unsplash

Customizing Reusable Frozen ML-Embeddings with Vespa

Deep-learned embeddings are popular for search and recommendation use cases. This post introduces the concept of using reusable frozen embeddings and tailoring them with Vespa.