You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am facing an issue when using the transformers backend with huggingface models and was hoping that someone can provide additional insights. I am quite sure it is just a small thing on my end, but I already tried quite a bunch of combinations and none of them worked.
LocalAI version:
v2.20.1 with localai/localai:latest-gpu-nvidia-cuda-12
Environment, CPU architecture, OS, and Version:
TBD
Describe the bug
When trying to interact with the model downloaded by the backend, I get the following error:
{
"error": {
"code": 500,
"message": "could not load model (no success): Unexpected err=OSError(\"Can't load the model for 'facebook/opt-125m'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'facebook/opt-125m' is the correct path to a directory containing a file named pytorch_model.bin, tf_model.h5, model.ckpt or flax_model.msgpack.\"), type(err)=<class 'OSError'>",
"type": ""
}
The model is apparently successfully downloaded into the models directory. The model path is also correctly configured (e.g. manually downloaded .gguf models can be used with llama-cpp backend). It is just that the docs on GPT with transformers backend seem to be out-of-date, at least I can not reproduce their results.
To Reproduce
Work through the docs on GPT for the transformers backend and try to implement the example.
Make sure the backend is configured and correctly loaded:
EXTRA_BACKENDS: backend/python/transformersPreparing backend: backend/python/transformersmake: Entering directory '/build/backend/python/transformers'bash install.shInitializing libbackend for transformersvirtualenv activatedactivated virtualenv has been ensuredstarting requirements install for /build/backend/python/transformers/requirements.txtAudited 4 packages in 73msfinished requirements install for /build/backend/python/transformers/requirements.txtmake: Leaving directory '/build/backend/python/transformers'
Expected behavior
As per the docs, I would have expected the model to be downloaded (which apparently happens) and be correctly detected, loaded, and used by the respective backend.
Logs
None
Additional context
I also carefully read this issue (and posted a detailed comment there) but it didn't solve my problem.
The text was updated successfully, but these errors were encountered:
I am facing an issue when using the transformers backend with huggingface models and was hoping that someone can provide additional insights. I am quite sure it is just a small thing on my end, but I already tried quite a bunch of combinations and none of them worked.
LocalAI version:
v2.20.1 with localai/localai:latest-gpu-nvidia-cuda-12
Environment, CPU architecture, OS, and Version:
TBD
Describe the bug
When trying to interact with the model downloaded by the backend, I get the following error:
The model is apparently successfully downloaded into the models directory. The model path is also correctly configured (e.g. manually downloaded
.gguf
models can be used with llama-cpp backend). It is just that the docs on GPT with transformers backend seem to be out-of-date, at least I can not reproduce their results.To Reproduce
Work through the docs on GPT for the transformers backend and try to implement the example.
Make sure the backend is configured and correctly loaded:
Expected behavior
As per the docs, I would have expected the model to be downloaded (which apparently happens) and be correctly detected, loaded, and used by the respective backend.
Logs
None
Additional context
I also carefully read this issue (and posted a detailed comment there) but it didn't solve my problem.
The text was updated successfully, but these errors were encountered: