summaryrefslogtreecommitdiff
path: root/training/gpu_manager.py
diff options
context:
space:
mode:
authorGustaf Rydholm <gustaf.rydholm@gmail.com>2021-04-05 20:47:55 +0200
committerGustaf Rydholm <gustaf.rydholm@gmail.com>2021-04-05 20:47:55 +0200
commit9ae5fa1a88899180f88ddb14d4cef457ceb847e5 (patch)
tree4fe2bcd82553c8062eb0908ae6442c123addf55d /training/gpu_manager.py
parent9e54591b7e342edc93b0bb04809a0f54045c6a15 (diff)
Add new training loop with PyTorch Lightning, remove stale files
Diffstat (limited to 'training/gpu_manager.py')
-rw-r--r--training/gpu_manager.py62
1 files changed, 0 insertions, 62 deletions
diff --git a/training/gpu_manager.py b/training/gpu_manager.py
deleted file mode 100644
index ce1b3dd..0000000
--- a/training/gpu_manager.py
+++ /dev/null
@@ -1,62 +0,0 @@
-"""GPUManager class."""
-import os
-import time
-from typing import Optional
-
-import gpustat
-from loguru import logger
-import numpy as np
-from redlock import Redlock
-
-
-GPU_LOCK_TIMEOUT = 5000 # ms
-
-
-class GPUManager:
- """Class for allocating GPUs."""
-
- def __init__(self, verbose: bool = False) -> None:
- """Initializes Redlock manager."""
- self.lock_manager = Redlock([{"host": "localhost", "port": 6379, "db": 0}])
- self.verbose = verbose
-
- def get_free_gpu(self) -> int:
- """Gets a free GPU.
-
- If some GPUs are available, try reserving one by checking out an exclusive redis lock.
- If none available or can not get lock, sleep and check again.
-
- Returns:
- int: The gpu index.
-
- """
- while True:
- gpu_index = self._get_free_gpu()
- if gpu_index is not None:
- return gpu_index
-
- if self.verbose:
- logger.debug(f"pid {os.getpid()} sleeping")
- time.sleep(GPU_LOCK_TIMEOUT / 1000)
-
- def _get_free_gpu(self) -> Optional[int]:
- """Fetches an available GPU index."""
- try:
- available_gpu_indices = [
- gpu.index
- for gpu in gpustat.GPUStatCollection.new_query()
- if gpu.memory_used < 0.5 * gpu.memory_total
- ]
- except Exception as e:
- logger.debug(f"Got the following exception: {e}")
- return None
-
- if available_gpu_indices:
- gpu_index = np.random.choice(available_gpu_indices)
- if self.verbose:
- logger.debug(f"pid {os.getpid()} picking gpu {gpu_index}")
- if self.lock_manager.lock(f"gpu_{gpu_index}", GPU_LOCK_TIMEOUT):
- return int(gpu_index)
- if self.verbose:
- logger.debug(f"pid {os.getpid()} could not get lock.")
- return None