Spice.ai
A federated AI/data runtime that combines embedded DuckDB compute with native delegation to Amazon S3 Vectors for similarity search. Configured via a single declarative `spicepod.yaml` file. Also serves as Vortex's launch home before its Linux Foundation transition.
Summary
A federated AI/data runtime that combines embedded DuckDB compute with native delegation to Amazon S3 Vectors for similarity search. Configured via a single declarative `spicepod.yaml` file. Also serves as Vortex's launch home before its Linux Foundation transition.
Spice.ai is a "data plane in a binary" — pulls hot data into DuckDB caches while delegating cold semantic search to S3 Vectors. Removes the persistent vector DB infrastructure layer for RAG and federated AI applications. Multimodal embedding support (Bedrock Nova, Titan) lets agents query text + images over a single S3 index.
- Spice.ai is not a vector database. It delegates vector search to S3 Vectors and does not host its own vector index.
- The federated model assumes S3 Vectors as the cold tier; non-AWS deployments require adaptation.
- DuckDB's caching is materialized; cache invalidation behavior must be understood before treating Spice.ai as a real-time layer.
depends_onDuckDB — analytical compute layerdepends_onAmazon S3 Vectors — vector similarity tieraugmentsDuckDB — adds vector search delegationscoped_toVector Indexing on Object Storage, S3
Definition
A lightweight, federated AI/data runtime that combines **embedded DuckDB compute** with native delegation to **Amazon S3 Vectors** for similarity search. Configuration lives in a single declarative `spicepod.yaml` file. Materializes hot data locally via DuckDB's caching layer while pushing semantic search down to the S3 Vectors tier; supports multimodal embeddings (Amazon Bedrock Nova, Titan) directly over the S3 index. Also serves as the launch home of **Vortex** — Spice.ai authored the format before it transitioned to the Linux Foundation.
RAG and federated AI applications conventionally require a heavyweight stack — application server + vector DB + cache + query engine. Spice.ai collapses this to a single declaratively-configured runtime that delegates analytical compute to DuckDB and vector search to S3 Vectors, eliminating the persistent infrastructure layer between model and data.
Federated AI/RAG applications without persistent vector DB infrastructure, edge inference with hot-data caching, multimodal embedding search over S3, prototype-to-production AI agents needing zero-ops data plane.
Connections 6
Outbound 6
scoped_to2depends_on2augments1solves1Resources 3
AWS Storage Blog walkthrough of Spice.ai's federated DuckDB + S3 Vectors architecture with worked spicepod.yaml configuration.
Spice.ai's own writeup explaining the Vortex columnar format integration that preceded the Linux Foundation transition.
Source repository with examples covering federated AI runtime configuration, S3 Vectors delegation, and DuckDB caching.