-
Notifications
You must be signed in to change notification settings - Fork 11
Description
I’m currently using version 1.2.5.290. I’ve created an index, and the folder size is around 800-900 MB. The dataset consists of roughly 10,000 video transcripts, each averaging about 10 minutes in length, so substantial, but not massive.
When I start FastPlaid using search.FastPlaid(index_path), the memory usage exceeds 5 GB of RAM, eventually causing my current EC2 instance to crash. I can try running it on a larger instance, but I’m starting to worry about FastPlaid’s scalability.
Does memory usage scale linearly with data size? Ultimately, I’ll need to index more than 30 times this amount of data, and running an instance with 100 GB of RAM isn’t feasible due to cost constraints.
The embedding dimension is 128.
Do you have any recommendations or advice?