Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEA] General batch knn multi-gpu algorithm implementation #727

Open
cjnolet opened this issue Feb 25, 2025 · 0 comments
Open

[FEA] General batch knn multi-gpu algorithm implementation #727

cjnolet opened this issue Feb 25, 2025 · 0 comments
Assignees
Labels
feature request New feature or request

Comments

@cjnolet
Copy link
Member

cjnolet commented Feb 25, 2025

nn-descent was recently given a batching feature that used clustering + spilling to scale the algorithm to datasets that were much larger than the GPU. This implementation has proven to be very useful for scaling algorithms for data analysis, such as UMAP, TSNE and HDBSCAN.

While nn-descent directly approximates the knn graph well, it is sensitive to CPU threads, and so scaling it directly to multi-GPU introduces additional challenges. One way to handle this is to break apart the "batches" into isolated tasks, each with their own number of threads. However, since the batching algorithm is really agnostic of the ANN algorithm, it would be preferable to host a top-level API for this to more easily allow the ann algorithmt to be plugged in.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request
Projects
Status: No status
Development

No branches or pull requests

2 participants