Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
M
mhc_rank
Manage
Activity
Members
Labels
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Package Registry
Model registry
Operate
Environments
Terraform modules
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Terms and privacy
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
Patrick Skillman-Lawrence
mhc_rank
Commits
32bec557
Commit
32bec557
authored
5 years ago
by
Tim O'Donnell
Browse files
Options
Downloads
Patches
Plain Diff
fix
parent
ced5ddc3
No related branches found
Branches containing commit
No related tags found
Tags containing commit
No related merge requests found
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
mhcflurry/parallelism.py
+21
-7
21 additions, 7 deletions
mhcflurry/parallelism.py
with
21 additions
and
7 deletions
mhcflurry/parallelism.py
+
21
−
7
View file @
32bec557
import
traceback
import
sys
import
os
import
functools
from
multiprocessing
import
Pool
,
Queue
,
cpu_count
from
six.moves
import
queue
from
multiprocessing.util
import
Finalize
...
...
@@ -46,6 +47,10 @@ def add_worker_pool_args(parser):
default
=
None
,
help
=
"
Restart workers after N tasks. Workaround for tensorflow memory
"
"
leaks. Requires Python >=3.2.
"
)
group
.
add_argument
(
"
--worker-log-dir
"
,
default
=
None
,
help
=
"
Write worker stdout and stderr logs to given directory.
"
)
def
worker_pool_with_gpu_assignments_from_args
(
args
):
...
...
@@ -54,7 +59,8 @@ def worker_pool_with_gpu_assignments_from_args(args):
num_gpus
=
args
.
gpus
,
backend
=
args
.
backend
,
max_workers_per_gpu
=
args
.
max_workers_per_gpu
,
max_tasks_per_worker
=
args
.
max_tasks_per_worker
max_tasks_per_worker
=
args
.
max_tasks_per_worker
,
worker_log_dir
=
args
.
worker_log_dir
,
)
...
...
@@ -63,7 +69,8 @@ def worker_pool_with_gpu_assignments(
num_gpus
=
0
,
backend
=
None
,
max_workers_per_gpu
=
1
,
max_tasks_per_worker
=
None
):
max_tasks_per_worker
=
None
,
worker_log_dir
=
None
):
num_workers
=
num_jobs
if
num_jobs
else
cpu_count
()
...
...
@@ -72,7 +79,7 @@ def worker_pool_with_gpu_assignments(
set_keras_backend
(
backend
)
return
None
worker_init_kwargs
=
None
worker_init_kwargs
=
[{}
for
_
in
range
(
num_workers
)]
if
num_gpus
:
print
(
"
Attempting to round-robin assign each worker a GPU.
"
)
if
backend
!=
"
tensorflow-default
"
:
...
...
@@ -82,8 +89,7 @@ def worker_pool_with_gpu_assignments(
gpu_assignments_remaining
=
dict
((
(
gpu
,
max_workers_per_gpu
)
for
gpu
in
range
(
num_gpus
)
))
worker_init_kwargs
=
[]
for
worker_num
in
range
(
num_workers
):
for
(
worker_num
,
kwargs
)
in
enumerate
(
worker_init_kwargs
):
if
gpu_assignments_remaining
:
# Use a GPU
gpu_num
=
sorted
(
...
...
@@ -97,13 +103,17 @@ def worker_pool_with_gpu_assignments(
# Use CPU
gpu_assignment
=
[]
worker_init_
kwargs
.
append
({
kwargs
.
update
({
'
gpu_device_nums
'
:
gpu_assignment
,
'
keras_backend
'
:
backend
})
print
(
"
Worker %d assigned GPUs: %s
"
%
(
worker_num
,
gpu_assignment
))
if
worker_log_dir
:
for
kwargs
in
worker_init_kwargs
:
kwargs
[
"
worker_log_dir
"
]
=
worker_log_dir
worker_pool
=
make_worker_pool
(
processes
=
num_workers
,
initializer
=
worker_init
,
...
...
@@ -208,7 +218,11 @@ def worker_init_entry_point(
init_function
(
**
kwargs
)
def
worker_init
(
keras_backend
=
None
,
gpu_device_nums
=
None
):
def
worker_init
(
keras_backend
=
None
,
gpu_device_nums
=
None
,
worker_log_dir
=
None
):
if
worker_log_dir
:
sys
.
stderr
=
sys
.
stdout
=
open
(
os
.
path
.
join
(
worker_log_dir
,
"
LOG-worker.%d.txt
"
,
"
w
"
))
# Each worker needs distinct random numbers
numpy
.
random
.
seed
()
random
.
seed
()
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment