runtime error
There was a problem when trying to write in your cache folder (/.cache/huggingface/hub). You should set the environment variable TRANSFORMERS_CACHE to a writable directory. None of PyTorch, TensorFlow >= 2.0, or Flax have been found. Models won't be available and only tokenizers, configuration and file/data utilities can be used. Traceback (most recent call last): File "/app/inference.py", line 14, in <module> server = LLMInferenceServer(model_name) File "/app/inference.py", line 5, in __init__ self.model = transformers.AutoModelForCausalLM.from_pretrained(model_name) File "/usr/local/lib/python3.9/site-packages/transformers/utils/import_utils.py", line 1124, in __getattribute__ requires_backends(cls, cls._backends) File "/usr/local/lib/python3.9/site-packages/transformers/utils/import_utils.py", line 1112, in requires_backends raise ImportError("".join(failed)) ImportError: AutoModelForCausalLM requires the PyTorch library but it was not found in your environment. Checkout the instructions on the installation page: https://pytorch.org/get-started/locally/ and follow the ones that match your environment. Please note that you may need to restart your runtime after installation.
Container logs:
Fetching error logs...