When a GGUF would be available?
It would be nice to have a gguf of the model
say not supported by llama.cpp when I tried to convert it
WARNING:hf-to-gguf:Trying to load config.json instead
INFO:hf-to-gguf:Model architecture: MaincoderForCausalLM
ERROR:hf-to-gguf:Model MaincoderForCausalLM is not supported
and https://huggingface.co/spaces/ggml-org/gguf-my-repo errored out
I know, i tried that too. I am guessing that you need to modify llama.cpp to be compatible with MaincoderForCausalLM
Someone from our team is already working on it so there should be an update soon : )
Okay! Hope everything goes well!
Any update on this?
Hi.. We've been taking a break after the release so this could some time but rest assured it will up updated here the moment it is done and released....
But again, in the meantime, to encourage community involvement and participation too, I think it would be great to see someone from the community do that if they're up for the task... The model's license permits that too so you're all cool...