Scaling data retrieval with S3-like storage and caching

View profile for Kunal Singhal

Builder | ex Grammarly, Meta | 2x ICPC world finalist

Our team is building retrieval over 10+ TB of multimodal data — text, images, PDFs, Excel, and more. More and more, I'm convinced that an S3-like storage first architecture is the only scalable path forward here. I didn't think we'd have to tackle this scale this early in our journey, but here we are! Using object storage as the source of truth, stateless compute layers, and intelligent caching would let you deliver near real-time results without the enormous costs of in-memory solutions. Ecosystem-wise, Turbopuffer is the only one I found that aligns closely with this approach; other popular options, such as Pinecone, Milvus, or Weaviate, lean more heavily on in-memory or block storage. Opensearch was not even on the radar for me this time around! Curious how others have handled this challenge.

Jason Sperske

Building LMS integrations for Grammarly Authorship.

1mo

I've been building a lot with s3 as a database. I believe databases have their place but for many problems they are an unnecessary architectural bottleneck.

Like
Reply

We've been using Turbopuffer as well. The cost and api have been great but wish their dev console was more polished.

Robert Schultz

Senior TPM, AI @ Meta

1mo

Decentralized storage via blockchain 🙂

Vivek Nayyar

Senior Engineering Manager at Qoala

1mo

Genuinely curious to understand the power of turbo puffer. I have heard a lot of folks recommend it

Like
Reply
See more comments

To view or add a comment, sign in

Explore content categories