unable download

#97
by luckyrain67777 - opened

always get“The requested URL returned error: 403” in my terminal
i don't know why

kind of same problem
OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like meta-llama/Meta-Llama-3-8B-Instruct is not the path to a directory containing a file named config.json.
and Internet connection is good (curl -I https://huggingface.co is fine)

Facing same issue .Is there any solution

I also encountered the same problem, after getting the link again or prompt 403: Forbidden, how to solve this problem?

+1 Same issue here...

it is more easier to download the model from the official website https://llama.meta.com/llama-downloads/ . It would be the original pt model. There is a way to convert the original model to hf models. Just use the script 'convert_llama_weights_to_hf.py' in the newest transformers. It must be the newest version transformers, because for older version, the 'convert_llama_weights_to_hf.py' only supports llama or llama2

does anyone got their error resolved? @JiayinWang , @luckyrain67777

A less than ideal solution is to use the path to the snapshot, i.e. $HF_HOME/models--meta-llama--Meta-Llama-3-8B-Instruct/snapshots/e5e23bbe8e749ef0efcf16cad411a7d23bd23298

Meta Llama org
osanseviero changed discussion status to closed

If anyone is having fine grained access token, make sure to add the repo in the "Repositories permissions" session.

Hi there! Please make sure to read https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/discussions/130

I have been granted the model access , but I still got the error: OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like meta-llama/Meta-Llama-3-8B-Instruct is not the path to a directory containing a file named config.json.
Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.

Sign up or log in to comment