Last updated:
0 purchases
preallocatecudamemory 1.0.0
Preallocate CUDA memory for pytorch
This is a code that helps preallocate memory for PyTorch, used for competing for computational resources with others.
You can use the following command directly on the command line:
python -m preallocate_cuda_memory
Or you can use in python file:
import preallocate_cuda_memory as pc
mc = pc.MemoryController(0) # 0 is the GPU index
mc.occupy_all_available_memory()
mc.free_memory()
If you find any issues, please feel free to contact the author by raising an issue on GitHub.
For personal and professional use. You cannot resell or redistribute these repositories in their original state.
There are no reviews.