Can't load model

#3
by Samoed - opened

When I try to download the model, it gives the following error:

from transformers import AutoTokenizer, AutoModel

tokenizer = AutoTokenizer.from_pretrained("zeta-alpha-ai/Zeta-Alpha-E5-Mistral")
model = AutoModel.from_pretrained("zeta-alpha-ai/Zeta-Alpha-E5-Mistral")
OSError: Could not locate config.json inside intfloat/e5-mistral-7b-instruct.

transformers==4.47.1
Example code

Zeta Alpha org

Hi there,
I'm looking into it now. I can load it just fine using sentence_transformers. I think the issue is that, for some reason, the loader is defaulting to try to get the PEFT adapters instead of the full model. I will take a look and get back to you in a bit.

Zeta Alpha org

@Samoed I think I've solved the problem. MTEB was having issues loading the PEFT using Sentence-Transformers. I'm testing the fix now (essentially moving the adapters to another folder) and will let you know when its live.
I've also found an issue where the non-PEFT model (i.e., the safetensor weights) are pointing to the non-ft model. So I will also fix that in a bit.

Awesome! Thank you

Zeta Alpha org

I've tested the "fixed" version and the performance is a bit lower than expected. I think the issue is due to BF16 vs TF32, but I'm not sure. @Samoed do you want me to upload the not-so-fixed version ASAP? I will check what is happening and upload the correct version when I can fix it.

No, I can wait!

Sign up or log in to comment