BT

Facilitating the Spread of Knowledge and Innovation in Professional Software Development

Write for InfoQ

Topics

Choose your language

InfoQ Homepage News Amazon MemoryDB Provides Fastest Vector Search on AWS

Amazon MemoryDB Provides Fastest Vector Search on AWS

AWS recently announced the general availability of vector search for Amazon MemoryDB, the managed in-memory database with Multi-AZ availability. The new capability provides ultra-low latency and the fastest vector search performance at the highest recall rates among vector databases on AWS.

Launched in 2021, Amazon MemoryDB is a Redis-compatible, durable, in-memory database. It is now the recommended managed choice for vector search on AWS in scenarios like generative AI applications where peak performance is the most important selection criterion. Channy Yun, principal developer advocate at AWS, writes:

With vector search for Amazon MemoryDB, you can use the existing MemoryDB API to implement generative AI use cases such as Retrieval Augmented Generation (RAG), anomaly (fraud) detection, document retrieval, and real-time recommendation engines. You can also generate vector embeddings using artificial intelligence and machine learning (AI/ML) services like Amazon Bedrock and Amazon SageMaker and store them within MemoryDB.

Developers can generate vector embeddings using managed services like Amazon Bedrock and SageMaker and store them within MemoryDB for real-time semantic search for RAG, low-latency durable semantic caching, and real-time anomaly detection.

Vector search for MemoryDB supports storing millions of vectors with single-digit millisecond queries and provides update latencies at the highest throughput levels with over 99% recall. Yun adds:

With vector search for MemoryDB, you can detect fraud by modeling fraudulent transactions based on your batch ML models, then loading normal and fraudulent transactions into MemoryDB to generate their vector representations through statistical decomposition techniques such as principal component analysis (PCA).

Source: AWS blog

The new capability was released in preview at re:Invent 2023 and the recent general availability introduces new features and improvements. These include VECTOR_RANGE, which allows the database to operate as a low-latency, durable semantic cache, and SCORE, which better filters on similarity. Vector fields support K-nearest neighbor searching (KNN) of fixed-sized vectors using the flat search (FLAT) and hierarchical navigable small worlds (HNSW) algorithms.

MemoryDB is not the only managed database on AWS supporting vector search. Among the different services targeting generative AI workloads, OpenSearch, Aurora PostgreSQL, RDS PostgreSQL, Neptune, and DocumentDB have introduced vector-related functionalities in the past year. Vinod Goje, software engineering manager at Bank of America, comments:

I've been watching the vector database market, which has been growing rapidly with numerous new products emerging (...) Experts believe the market is becoming overcrowded, making it difficult for new products to stand out amidst the plethora of existing options.

Shayon Sanyal and Graham Kutchek, database specialist solutions architects at AWS, detail the key considerations when choosing a database for generative AI applications. They suggest:

If you’re already using OpenSearch Service, Aurora PostgreSQL, RDS for PostgreSQL, DocumentDB or MemoryDB, leverage their vector search capabilities for your existing data. For graph-based RAG applications, consider Amazon Neptune. If your data is stored in DynamoDB, OpenSearch can be an excellent choice for vector search using zero-ETL integration. If you are still unsure, use OpenSearch Service.

All cloud providers provide have recently introduced vector search capabilities to compete with vector databases like Pinecone and serverless Momento Cache. For example, InfoQ previously reported on Google BigQuery and Microsoft Vector Search.

Vector search is available for Amazon MemoryDB version 7.1 and a single shard configuration in all regions where the database is available.

About the Author

Rate this Article

Adoption
Style

BT