Skip to content

Commit 91889da

Browse files
authored
Bump cuQuantum 25.09 -> 25.09.1 (#351)
Requires: NVIDIA/cuda-quantum#3586
1 parent b7f6d8d commit 91889da

File tree

6 files changed

+8
-8
lines changed

6 files changed

+8
-8
lines changed

.cudaq_version

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
{
22
"cudaq": {
33
"repository": "NVIDIA/cuda-quantum",
4-
"ref": "051eed70cb1426a5522140183c8e16cf72d51835"
4+
"ref": "4f284af95b252e69ab08ba4f1ef7fa1ea69dd66a"
55
}
66
}

.github/workflows/all_libs.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ jobs:
109109
# Install the correct torch first.
110110
cuda_no_dot=$(echo ${{ matrix.cuda_version }} | sed 's/\.//')
111111
pip install torch==2.9.0 --index-url https://download.pytorch.org/whl/cu${cuda_no_dot}
112-
pip install numpy pytest cupy-cuda${{ steps.config.outputs.cuda_major }}x cuquantum-cu${{ steps.config.outputs.cuda_major }} lightning ml_collections mpi4py transformers quimb opt_einsum nvidia-cublas cuquantum-python-cu${{ steps.config.outputs.cuda_major }}==25.09
112+
pip install numpy pytest cupy-cuda${{ steps.config.outputs.cuda_major }}x cuquantum-cu${{ steps.config.outputs.cuda_major }} lightning ml_collections mpi4py transformers quimb opt_einsum nvidia-cublas cuquantum-python-cu${{ steps.config.outputs.cuda_major }}==25.09.1
113113
# The following tests are needed for docs/sphinx/examples/qec/python/tensor_network_decoder.py.
114114
if [ "$(uname -m)" == "x86_64" ]; then
115115
# Stim is not currently available on manylinux ARM wheels, so only

.github/workflows/all_libs_release.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -133,7 +133,7 @@ jobs:
133133
# Install the correct torch first.
134134
cuda_no_dot=$(echo ${{ matrix.cuda_version }} | sed 's/\.//')
135135
pip install torch==2.9.0 --index-url https://download.pytorch.org/whl/cu${cuda_no_dot}
136-
pip install numpy pytest cupy-cuda${{ steps.config.outputs.cuda_major }}x cuquantum-cu${{ steps.config.outputs.cuda_major }} lightning ml_collections mpi4py transformers quimb opt_einsum nvidia-cublas cuquantum-python-cu${{ steps.config.outputs.cuda_major }}==25.09
136+
pip install numpy pytest cupy-cuda${{ steps.config.outputs.cuda_major }}x cuquantum-cu${{ steps.config.outputs.cuda_major }} lightning ml_collections mpi4py transformers quimb opt_einsum nvidia-cublas cuquantum-python-cu${{ steps.config.outputs.cuda_major }}==25.09.1
137137
# The following tests are needed for docs/sphinx/examples/qec/python/tensor_network_decoder.py.
138138
if [ "$(uname -m)" == "x86_64" ]; then
139139
# Stim is not currently available on manylinux ARM wheels, so only

.github/workflows/lib_qec.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -106,7 +106,7 @@ jobs:
106106
# Install the correct torch first.
107107
cuda_no_dot=$(echo ${{ matrix.cuda_version }} | sed 's/\.//')
108108
pip install torch==2.9.0 --index-url https://download.pytorch.org/whl/cu${cuda_no_dot}
109-
pip install numpy pytest cupy-cuda${{ steps.config.outputs.cuda_major }}x cuquantum-cu${{ steps.config.outputs.cuda_major }} quimb opt_einsum nvidia-cublas cuquantum-python-cu${{ steps.config.outputs.cuda_major }}==25.09
109+
pip install numpy pytest cupy-cuda${{ steps.config.outputs.cuda_major }}x cuquantum-cu${{ steps.config.outputs.cuda_major }} quimb opt_einsum nvidia-cublas cuquantum-python-cu${{ steps.config.outputs.cuda_major }}==25.09.1
110110
# The following tests are needed for docs/sphinx/examples/qec/python/tensor_network_decoder.py.
111111
if [ "$(uname -m)" == "x86_64" ]; then
112112
# Stim is not currently available on manylinux ARM wheels, so only

libs/qec/pyproject.toml.cu12

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ tensor_network_decoder = [
5656
"quimb",
5757
"opt_einsum",
5858
"torch",
59-
"cuquantum-python-cu12==25.09"
59+
"cuquantum-python-cu12==25.09.1"
6060
]
6161
trt_decoder = [
6262
"tensorrt-cu12"
@@ -65,6 +65,6 @@ all = [
6565
"quimb",
6666
"opt_einsum",
6767
"torch",
68-
"cuquantum-python-cu12==25.09",
68+
"cuquantum-python-cu12==25.09.1",
6969
"tensorrt-cu12; platform_machine == 'x86_64'"
7070
]

libs/qec/pyproject.toml.cu13

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ tensor_network_decoder = [
5656
"quimb",
5757
"opt_einsum",
5858
"torch>=2.9.0",
59-
"cuquantum-python-cu13==25.09"
59+
"cuquantum-python-cu13==25.09.1"
6060
]
6161
trt_decoder = [
6262
"tensorrt-cu13"
@@ -65,6 +65,6 @@ all = [
6565
"quimb",
6666
"opt_einsum",
6767
"torch>=2.9.0",
68-
"cuquantum-python-cu13==25.09",
68+
"cuquantum-python-cu13==25.09.1",
6969
"tensorrt-cu13"
7070
]

0 commit comments

Comments
 (0)