Reputation: 1
Hi I have some CUDA memory issue even though I am using multiple GPUs. I am calling coreference resolution API on long document (aorund 2000words). It seems that the memory is not paralleled. How can I solve this issue? (I am currently using the API as here https://demo.allennlp.org/coreference-resolution)
Upvotes: -2
Views: 80
Reputation: 2627
The coref model uses a lot of memory. It does not automatically take advantage of multiple GPUs. The best thing you can do is reduce the maximum sequence length you send to the model until it fits.
Upvotes: 0