Our team is building retrieval over 10+ TB of multimodal data — text, images, PDFs, Excel, and more. More and more, I'm convinced that an S3-like storage first architecture is the only scalable path forward here. I didn't think we'd have to tackle this scale this early in our journey, but here we are! Using object storage as the source of truth, stateless compute layers, and intelligent caching would let you deliver near real-time results without the enormous costs of in-memory solutions. Ecosystem-wise, Turbopuffer is the only one I found that aligns closely with this approach; other popular options, such as Pinecone, Milvus, or Weaviate, lean more heavily on in-memory or block storage. Opensearch was not even on the radar for me this time around! Curious how others have handled this challenge.
We've been using Turbopuffer as well. The cost and api have been great but wish their dev console was more polished.
Decentralized storage via blockchain 🙂
Genuinely curious to understand the power of turbo puffer. I have heard a lot of folks recommend it
Building LMS integrations for Grammarly Authorship.
1moI've been building a lot with s3 as a database. I believe databases have their place but for many problems they are an unnecessary architectural bottleneck.