Skip to content

Conversation

@AdnanElAssadi56
Copy link
Contributor

If you add a model or a dataset, please add the corresponding checklist:

self.clip = CLIPModel.from_pretrained(model_name, revision=revision).to(device)
# text side (CLIP) - we use the standard OpenAI CLIP model as mentioned in paper
clip_model_name = "openai/clip-vit-base-patch32"
self.clip = CLIPModel.from_pretrained(clip_model_name).to(device)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

add revision as well

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You should be able to load it directly onto the device using the device_map

# Pad with zeros
pad_length = max_length - wav.shape[-1]
padded_wav = torch.nn.functional.pad(wav, (0, pad_length))
padded_wav = np.pad(wav, (0, pad_length), mode="constant")
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

torch is fine - no need to change

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Changed back. I think some dataset gave an error when using torch but can't remember which or if it was because of something else. For now, changed back to torch. They are equivalent in 1D audio arrays.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you remove link to submodule?

@Samoed
Copy link
Member

Samoed commented Jan 6, 2026

I think this the same as #3781

@AdnanElAssadi56
Copy link
Contributor Author

Yup closing the other one.

@AdnanElAssadi56
Copy link
Contributor Author

@Samoed Anything else needed here?

@AdnanElAssadi56 AdnanElAssadi56 changed the title Add proper wav2clip text encoder [MAEB] Add proper wav2clip text encoder Jan 6, 2026
@Samoed Samoed merged commit 782ce45 into embeddings-benchmark:maeb Jan 7, 2026
9 of 10 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants