Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LSQ++ in 16x4 (nbits=4) - Does NOT scale up to large training sets #43

Open
k-amara opened this issue Aug 12, 2021 · 1 comment
Open
Labels

Comments

@k-amara
Copy link

k-amara commented Aug 12, 2021

Hello,

I have run LSQ++ for codebook of size M=16 (number of subspaces) and codes encoded in nbits=4 for BigANN1M and Deep1M. When increasing the size of the training set, I observe a drop in recall (@1, @10, @100) for both datasets. Please find attached graphics that illustrate the problem.
Screenshot 2021-08-12 at 09 46 04
Screenshot 2021-08-12 at 09 50 06

I have used for LSQ++ the FAISS implementation (faiss.LocalSearchQuantizer(d, M, nbits)). @mdouze

Have you experienced this issue when testing LSQ++16x4?
I did a gridsearch on niter_train and niter_ils_train but have observed no difference in the drop...

Cheers
@k-amara

@una-dinosauria
Copy link
Owner

Hello @k-amara,

This seems very strange. I remember trying out larger training sets during my PhD, and I did not observe drops in recall -- and definitely not dramatic ones like the ones you've shared.

Does this happen with the implementation in Rayuela too? If it doesn't, then it's probably a bug in the FAISS implementation.

Cheers,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants