bm.benchmark()
fails to de-allocate GPU memory. garbage collection fail?
#186
Labels
bug
Something isn't working
Report
I'm running
bm.benchmark()
using threeembedding_obsm_keys
on 2.7m cell dataset. It previously ran, albiet slowly on a smaller dataset using only two keys.I can reproduce it on a local 4080, and on dual remote v100s.
CODE: ( almost directly from tutoral)
FAILURE:
The
bm.prepare(neighbor_computer=faiss_brute_force_nn)
works fine and speed scales nicely with resources. (Thanks Jax!). Which exact metric it fails on seems to vary, but in general for each itteration of the embedding key:More corroborative evidence of issue:
If I subsample the adata to be small enough to not hit the ceiling on the three iterations, I can follow the steps of GPU memory utilization jumping at each iteration and get the bm.benchmark() to finish. But the memory is not freed from the GPU when
bm.benchmark()
returns. Even after deleting the instance or reassigning bm, GPU garbage collection doesn't happen. E.g. If i run in an interactive python the memory allocation the memory is held until the python is killed.Version information
cat requirements.txt
The text was updated successfully, but these errors were encountered: