Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Save and load compressed index #18

Open
xudongguan202 opened this issue Sep 5, 2023 · 2 comments
Open

Save and load compressed index #18

xudongguan202 opened this issue Sep 5, 2023 · 2 comments

Comments

@xudongguan202
Copy link

Hi,

I got a total 755G index saved in my disk after encoding the whole wiki passage. The large index takes huge storage and long time to load to GPU. However, it requires less than 100G after loading to GPU, which could be the index compression mentioned in your paper. Is it possible to save and load the compressed index for better time and storage consumption?

@Vincent-ch99
Copy link

Yeh, I met the same situation like u
I loaded the embeddings and passages of precomputing, but when I run evaluate.py, it always shows cuda out of memory while loading. I have an A100 80G GPU memory. How much GPU memory is needed at least to load the precomputing embedding?
And I also encountered the problem of extremely slowly loading. Is there any way to optimize it?

@xudongguan202
Copy link
Author

@Vincent-ch99
At least 2x A100 80G GPUs are required to run the evaluation using the default configurations from my experiment.

The extremely slow loading could be due to index compression of saved index, as mentioned in the paper. Therefore, I am currently trying to find a way to save and load the compressed index for faster loading and light storage.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants