Skip to content

Commit 646a581

Browse files
authored
Renamed internally used runai names (#189)
1 parent c0d5e20 commit 646a581

File tree

27 files changed

+79
-91
lines changed

27 files changed

+79
-91
lines changed

pkg/binder/binding/default_binder_test.go

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -83,18 +83,18 @@ func TestBindApplyResourceReceivedType(t *testing.T) {
8383
Name: common.NvidiaVisibleDevices,
8484
ValueFrom: &v1.EnvVarSource{
8585
ConfigMapKeyRef: &v1.ConfigMapKeySelector{
86-
Key: common.RunaiVisibleDevices,
86+
Key: common.VisibleDevices,
8787
LocalObjectReference: v1.LocalObjectReference{
8888
Name: "my-config",
8989
},
9090
},
9191
},
9292
},
9393
{
94-
Name: common.RunaiNumOfGpus,
94+
Name: common.NumOfGpusEnvVar,
9595
ValueFrom: &v1.EnvVarSource{
9696
ConfigMapKeyRef: &v1.ConfigMapKeySelector{
97-
Key: common.RunaiNumOfGpus,
97+
Key: common.NumOfGpusEnvVar,
9898
LocalObjectReference: v1.LocalObjectReference{
9999
Name: "my-config",
100100
},

pkg/binder/binding/fraction_binder_test.go

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ var happyFlowObjectsBc = []runtime.Object{
4242
Containers: []v1.Container{{
4343
Env: []v1.EnvVar{
4444
{
45-
Name: common.RunaiNumOfGpus,
45+
Name: common.NumOfGpusEnvVar,
4646
ValueFrom: &v1.EnvVarSource{
4747
ConfigMapKeyRef: &v1.ConfigMapKeySelector{
4848
LocalObjectReference: v1.LocalObjectReference{
@@ -217,8 +217,8 @@ var _ = Describe("FractionBinder", func() {
217217
if err := fakeClient.Get(context.TODO(), client.ObjectKeyFromObject(configMap), configMap); err != nil {
218218
Fail(fmt.Sprintf("Failed to read configmap: %v", err))
219219
} else {
220-
Expect(configMap.Data[common.RunaiVisibleDevices]).To(Equal(testData.gpuIndexByGroupIndex))
221-
Expect(configMap.Data[common.RunaiNumOfGpus]).To(Equal("0.5"))
220+
Expect(configMap.Data[common.VisibleDevices]).To(Equal(testData.gpuIndexByGroupIndex))
221+
Expect(configMap.Data[common.NumOfGpusEnvVar]).To(Equal("0.5"))
222222
}
223223
})
224224
}

pkg/binder/binding/resourcereservation/resource_reservation.go

Lines changed: 3 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,6 @@ const (
4040
numberOfGPUsToReserve = 1
4141
reservationPodRandomCharacters = 5
4242
unknownGpuIndicator = "-1"
43-
nodeIndex = "runai-node"
4443
)
4544

4645
type service struct {
@@ -231,20 +230,15 @@ func (rsc *service) updatePodGPUGroup(
231230
"namespace", pod.Namespace, "name", pod.Name, "node", nodeName,
232231
"gpu-group", gpuGroup)
233232
originalPod := pod.DeepCopy()
234-
if pod.Annotations == nil {
235-
pod.Annotations = map[string]string{}
233+
if pod.Labels == nil {
234+
pod.Labels = map[string]string{}
236235
}
237-
pod.Annotations[nodeIndex] = nodeName
238-
239236
isMultiFraction, err := resources.IsMultiFraction(pod)
240237
if err != nil {
241238
return fmt.Errorf(
242-
"failed to determine is the pod <%s/%s> is a multi fractional pod while setting gpu group label. %w",
239+
"failed to determine if pod <%s/%s> is a multi fractional pod while setting gpu group label. %w",
243240
pod.Namespace, pod.Name, err)
244241
}
245-
if pod.Labels == nil {
246-
pod.Labels = map[string]string{}
247-
}
248242
if isMultiFraction {
249243
labelKey, labelValue := resources.GetMultiFractionGpuGroupLabel(gpuGroup)
250244
pod.Labels[labelKey] = labelValue

pkg/binder/common/configmap.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ func getConfigMapByMounts(pod *v1.Pod, container *v1.Container) string {
4343

4444
func getConfigMapNameByEnvVar(container *v1.Container) string {
4545
for _, envVar := range container.Env {
46-
if envVar.Name == RunaiNumOfGpus && envVar.ValueFrom != nil &&
46+
if envVar.Name == NumOfGpusEnvVar && envVar.ValueFrom != nil &&
4747
envVar.ValueFrom.ConfigMapKeyRef != nil {
4848
return envVar.ValueFrom.ConfigMapKeyRef.Name
4949
}

pkg/binder/common/constants.go

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,8 +5,8 @@ package common
55

66
const (
77
NvidiaVisibleDevices = "NVIDIA_VISIBLE_DEVICES"
8-
RunaiVisibleDevices = "RUNAI-VISIBLE-DEVICES"
9-
RunaiNumOfGpus = "RUNAI_NUM_OF_GPUS"
8+
VisibleDevices = "RUNAI-VISIBLE-DEVICES"
9+
NumOfGpusEnvVar = "RUNAI_NUM_OF_GPUS"
1010
ReceivedTypeFraction = "Fraction"
1111
ReceivedTypeRegular = "Regular"
1212
)

pkg/binder/common/gpu_access.go

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ func AddVisibleDevicesEnvVars(container *v1.Container, sharedGpuConfigMapName st
1616
Name: NvidiaVisibleDevices,
1717
ValueFrom: &v1.EnvVarSource{
1818
ConfigMapKeyRef: &v1.ConfigMapKeySelector{
19-
Key: RunaiVisibleDevices,
19+
Key: VisibleDevices,
2020
LocalObjectReference: v1.LocalObjectReference{
2121
Name: sharedGpuConfigMapName,
2222
},
@@ -25,10 +25,10 @@ func AddVisibleDevicesEnvVars(container *v1.Container, sharedGpuConfigMapName st
2525
})
2626

2727
AddEnvVarToContainer(container, v1.EnvVar{
28-
Name: RunaiNumOfGpus,
28+
Name: NumOfGpusEnvVar,
2929
ValueFrom: &v1.EnvVarSource{
3030
ConfigMapKeyRef: &v1.ConfigMapKeySelector{
31-
Key: RunaiNumOfGpus,
31+
Key: NumOfGpusEnvVar,
3232
LocalObjectReference: v1.LocalObjectReference{
3333
Name: sharedGpuConfigMapName,
3434
},
@@ -41,7 +41,7 @@ func SetNvidiaVisibleDevices(ctx context.Context, kubeClient client.Client, pod
4141
container *v1.Container, visibleDevicesValue string) error {
4242
nvidiaVisibleDevicesDefinedInSpec := false
4343
for _, envVar := range container.Env {
44-
if envVar.Name == RunaiNumOfGpus && envVar.ValueFrom != nil &&
44+
if envVar.Name == NumOfGpusEnvVar && envVar.ValueFrom != nil &&
4545
envVar.ValueFrom.ConfigMapKeyRef != nil {
4646
nvidiaVisibleDevicesDefinedInSpec = true
4747
}
@@ -53,7 +53,7 @@ func SetNvidiaVisibleDevices(ctx context.Context, kubeClient client.Client, pod
5353
}
5454
err = UpdateConfigMapEnvironmentVariable(ctx, kubeClient, pod, capabilitiesMapName,
5555
func(data map[string]string) error {
56-
data[RunaiVisibleDevices] = visibleDevicesValue
56+
data[VisibleDevices] = visibleDevicesValue
5757
return nil
5858
})
5959
return err
@@ -75,7 +75,7 @@ func SetNumOfGPUDevices(
7575
ctx context.Context, kubeClient client.Client, pod *v1.Pod, container *v1.Container, numOfGPUs string,
7676
) error {
7777
updateFunc := func(data map[string]string) error {
78-
data[RunaiNumOfGpus] = numOfGPUs
78+
data[NumOfGpusEnvVar] = numOfGPUs
7979
return nil
8080
}
8181
configMapName, err := GetConfigMapName(pod, container)

pkg/binder/common/gpusharingconfigmap/config_map.go

Lines changed: 0 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -23,8 +23,6 @@ const (
2323
maxVolumeNameLength = 63
2424
configMapNameNumRandomChars = 7
2525
configMapNameExtraChars = configMapNameNumRandomChars + 6
26-
runaiVisibleDevices = "RUNAI-VISIBLE-DEVICES"
27-
runaiNumOfGpus = "RUNAI_NUM_OF_GPUS"
2826
)
2927

3028
func UpsertJobConfigMap(ctx context.Context,
@@ -219,13 +217,6 @@ func setConfigMapNameAnnotation(pod *v1.Pod, name string) {
219217
pod.Annotations[DesiredConfigMapPrefixKey] = name
220218
}
221219

222-
func GenerateCapabilitiesConfigMapData() map[string]string {
223-
data := make(map[string]string)
224-
data[runaiVisibleDevices] = ""
225-
data[runaiNumOfGpus] = ""
226-
return data
227-
}
228-
229220
// ownerReferencesDifferent compares two OwnerReferences and returns true if they are not the same
230221
func ownerReferencesDifferent(lOwnerReferences, rOwnerReferences metav1.OwnerReference) bool {
231222
return !(lOwnerReferences.APIVersion == rOwnerReferences.APIVersion &&

pkg/common/resources/gpu_sharing.go

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -49,6 +49,9 @@ func GetGPUMemory(pod *v1.Pod) (int64, error) {
4949
}
5050

5151
func GetNumGPUFractionDevices(pod *v1.Pod) (int64, error) {
52+
if pod.Annotations == nil {
53+
return 0, fractionDevicesAnnotationNotFound
54+
}
5255
mumDevicesStr, found := pod.Annotations[constants.GpuFractionsNumDevices]
5356
if !found {
5457
_, foundFraction := pod.Annotations[constants.GpuFraction]

pkg/podgrouper/podgrouper/plugins/job/job_grouper_test.go

Lines changed: 17 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -72,10 +72,10 @@ func TestGetPodGroupMetadata_Hpo(t *testing.T) {
7272
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects().Build()
7373

7474
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
75-
runaiJobGrouper := NewK8sJobGrouper(client, defaultGrouper, false)
75+
jobGrouper := NewK8sJobGrouper(client, defaultGrouper, false)
7676

77-
podGroupMetadata, err := runaiJobGrouper.GetPodGroupMetadata(owner, pod)
78-
podGroupMetadata2, err2 := runaiJobGrouper.GetPodGroupMetadata(owner, pod2)
77+
podGroupMetadata, err := jobGrouper.GetPodGroupMetadata(owner, pod)
78+
podGroupMetadata2, err2 := jobGrouper.GetPodGroupMetadata(owner, pod2)
7979

8080
assert.Nil(t, err)
8181
assert.Nil(t, err2)
@@ -122,7 +122,7 @@ func TestGetPodGroupMetadata_LegacyPodGroup(t *testing.T) {
122122
},
123123
}
124124

125-
var runaiTestResources = []runtime.Object{
125+
var testResources = []runtime.Object{
126126
&schedulingv2.PodGroup{
127127
TypeMeta: metav1.TypeMeta{
128128
Kind: "PodGroup",
@@ -146,12 +146,12 @@ func TestGetPodGroupMetadata_LegacyPodGroup(t *testing.T) {
146146
t.Fail()
147147
}
148148

149-
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(runaiTestResources...).Build()
149+
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(testResources...).Build()
150150

151151
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
152-
runaiJobGrouper := NewK8sJobGrouper(client, defaultGrouper, true)
152+
jobGrouper := NewK8sJobGrouper(client, defaultGrouper, true)
153153

154-
podGroupMetadata, err := runaiJobGrouper.GetPodGroupMetadata(owner, pod)
154+
podGroupMetadata, err := jobGrouper.GetPodGroupMetadata(owner, pod)
155155

156156
assert.Nil(t, err)
157157
assert.Equal(t, "pg-test_name-4kgrb-1234-5678", podGroupMetadata.Name)
@@ -195,7 +195,7 @@ func TestGetPodGroupMetadata_LegacyDisabledPodGroup(t *testing.T) {
195195
},
196196
}
197197

198-
var runaiTestResources = []runtime.Object{
198+
var testResources = []runtime.Object{
199199
&schedulingv2.PodGroup{
200200
TypeMeta: metav1.TypeMeta{
201201
Kind: "PodGroup",
@@ -219,12 +219,12 @@ func TestGetPodGroupMetadata_LegacyDisabledPodGroup(t *testing.T) {
219219
t.Fail()
220220
}
221221

222-
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(runaiTestResources...).Build()
222+
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(testResources...).Build()
223223

224224
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
225-
runaiJobGrouper := NewK8sJobGrouper(client, defaultGrouper, false)
225+
jobGrouper := NewK8sJobGrouper(client, defaultGrouper, false)
226226

227-
podGroupMetadata, err := runaiJobGrouper.GetPodGroupMetadata(owner, pod)
227+
podGroupMetadata, err := jobGrouper.GetPodGroupMetadata(owner, pod)
228228

229229
assert.Nil(t, err)
230230
assert.Equal(t, "pg-test_name-4kgrb-1234-5678", podGroupMetadata.Name)
@@ -268,20 +268,20 @@ func TestGetPodGroupMetadata_LegacyNotFound(t *testing.T) {
268268
},
269269
}
270270

271-
var runaiTestResources = []runtime.Object{}
271+
var testResources = []runtime.Object{}
272272

273273
scheme := runtime.NewScheme()
274274
err := schedulingv2.AddToScheme(scheme)
275275
if err != nil {
276276
t.Fail()
277277
}
278278

279-
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(runaiTestResources...).Build()
279+
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(testResources...).Build()
280280

281281
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
282-
runaiJobGrouper := NewK8sJobGrouper(client, defaultGrouper, true)
282+
jobGrouper := NewK8sJobGrouper(client, defaultGrouper, true)
283283

284-
podGroupMetadata, err := runaiJobGrouper.GetPodGroupMetadata(owner, pod)
284+
podGroupMetadata, err := jobGrouper.GetPodGroupMetadata(owner, pod)
285285

286286
assert.Nil(t, err)
287287
assert.Equal(t, "pg-test_name-4kgrb-1234-5678", podGroupMetadata.Name)
@@ -334,9 +334,9 @@ func TestGetPodGroupMetadata_RegularPodGroup(t *testing.T) {
334334
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects().Build()
335335

336336
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
337-
runaiJobGrouper := NewK8sJobGrouper(client, defaultGrouper, false)
337+
jobGrouper := NewK8sJobGrouper(client, defaultGrouper, false)
338338

339-
podGroupMetadata, err := runaiJobGrouper.GetPodGroupMetadata(owner, pod)
339+
podGroupMetadata, err := jobGrouper.GetPodGroupMetadata(owner, pod)
340340

341341
assert.Nil(t, err)
342342
assert.Equal(t, "pg-test_name-4kgrb-1234-5678", podGroupMetadata.Name)

pkg/podgrouper/podgrouper/plugins/runaijob/runaijob_grouper_test.go

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -120,7 +120,7 @@ func TestGetPodGroupMetadata_LegacyPodGroup(t *testing.T) {
120120
},
121121
}
122122

123-
var runaiTestResources = []runtime.Object{
123+
var testResources = []runtime.Object{
124124
&schedulingv2.PodGroup{
125125
TypeMeta: metav1.TypeMeta{
126126
Kind: "PodGroup",
@@ -144,7 +144,7 @@ func TestGetPodGroupMetadata_LegacyPodGroup(t *testing.T) {
144144
t.Fail()
145145
}
146146

147-
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(runaiTestResources...).Build()
147+
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(testResources...).Build()
148148

149149
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
150150
runaiJobGrouper := NewRunaiJobGrouper(client, defaultGrouper, true)
@@ -193,7 +193,7 @@ func TestGetPodGroupMetadata_LegacyDisabledPodGroup(t *testing.T) {
193193
},
194194
}
195195

196-
var runaiTestResources = []runtime.Object{
196+
var testResources = []runtime.Object{
197197
&schedulingv2.PodGroup{
198198
TypeMeta: metav1.TypeMeta{
199199
Kind: "PodGroup",
@@ -217,7 +217,7 @@ func TestGetPodGroupMetadata_LegacyDisabledPodGroup(t *testing.T) {
217217
t.Fail()
218218
}
219219

220-
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(runaiTestResources...).Build()
220+
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(testResources...).Build()
221221

222222
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
223223
runaiJobGrouper := NewRunaiJobGrouper(client, defaultGrouper, false)
@@ -266,15 +266,15 @@ func TestGetPodGroupMetadata_LegacyNotFound(t *testing.T) {
266266
},
267267
}
268268

269-
var runaiTestResources = []runtime.Object{}
269+
var testResources = []runtime.Object{}
270270

271271
scheme := runtime.NewScheme()
272272
err := schedulingv2.AddToScheme(scheme)
273273
if err != nil {
274274
t.Fail()
275275
}
276276

277-
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(runaiTestResources...).Build()
277+
client := fake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(testResources...).Build()
278278

279279
defaultGrouper := defaultgrouper.NewDefaultGrouper(queueLabelKey, nodePoolLabelKey)
280280
runaiJobGrouper := NewRunaiJobGrouper(client, defaultGrouper, true)

0 commit comments

Comments
 (0)