Skip to content

Add multi-GPU worker assignment#292

Merged
L-M-Sherlock merged 1 commit intomainfrom
feature/multi-gpu-training
Feb 1, 2026
Merged

Add multi-GPU worker assignment#292
L-M-Sherlock merged 1 commit intomainfrom
feature/multi-gpu-training

Conversation

@L-M-Sherlock
Copy link
Member

No description provided.

@L-M-Sherlock L-M-Sherlock merged commit 553dce5 into main Feb 1, 2026
0 of 2 checks passed
@L-M-Sherlock L-M-Sherlock deleted the feature/multi-gpu-training branch February 1, 2026 15:16
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces multi-GPU support for worker processes, allowing for round-robin assignment of CUDA devices. The implementation includes command-line argument parsing for GPU selection, validation of device IDs, and configuration of worker processes. The changes are well-structured. I have one suggestion to improve error handling in the device configuration logic to prevent silent failures.

Comment on lines +220 to +221
except Exception:
pass
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

Using a broad except Exception: pass is risky as it can hide important errors, such as ImportError if reptile_trainer is not found or AttributeError if DEVICE is not a member. This could lead to the model silently running on the wrong device. It's better to catch specific exceptions and log a warning to make debugging easier.

Suggested change
except Exception:
pass
except (ImportError, AttributeError) as e:
print(f"Warning: Could not configure device for reptile_trainer: {e}")

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant

Comments