Skip to content

Version based GPU configuration and QoS addition #1092

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged

Conversation

kvenkateshan-meta
Copy link
Contributor

Summary:
Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call gpus-per-node in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

Changes here

  1. Introduced Slurm Version based GPU request configuration

  2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jul 23, 2025
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 23, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304
@kvenkateshan-meta kvenkateshan-meta force-pushed the export-D78778304 branch 2 times, most recently from f67d4b6 to d06b519 Compare July 23, 2025 20:25
kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 23, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 23, 2025
Summary:
Pull Request resolved: pytorch#1092

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Differential Revision: D78778304
kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 28, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 28, 2025
Summary:
Pull Request resolved: pytorch#1092

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
kvenkateshan-meta added a commit to kvenkateshan-meta/torchx that referenced this pull request Jul 29, 2025
Summary:

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.


https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
Summary:
Pull Request resolved: pytorch#1092

 Slurm 24.11.0rc1 and beyond do not suport GRES per task. So we need to call `gpus-per-node` in sbatch to ensure failure free allocation.

https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-24.11.md

# Changes here

1. Introduced Slurm Version based GPU request configuration

2. Introduced an option QoS parameter which can be used to control priority of jobs.

Reviewed By: kiukchung

Differential Revision: D78778304
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D78778304

@facebook-github-bot facebook-github-bot merged commit ae55901 into pytorch:main Jul 29, 2025
23 of 24 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants