About 7,620,000 results
Open links in new tab
  1. python - Cannot load a gated model from hugginface despite having ...

    Nov 21, 2024 · I am training a Llama-3.1-8B-Instruct model for a specific task. I have request the access to the huggingface repository, and got access, confirmed on the huggingface webapp dashboard. I …

  2. huggingface hub - ImportError: cannot import name …

    Jan 21, 2025 · ImportError: cannot import name 'cached_download' from 'huggingface_hub' Asked 10 months ago Modified 8 months ago Viewed 24k times

  3. How to download a model from huggingface? - Stack Overflow

    May 19, 2021 · How about using hf_hub_download from huggingface_hub library? hf_hub_download returns the local path where the model was downloaded so you could hook this one liner with …

  4. Load a pre-trained model from disk with Huggingface Transformers

    Sep 22, 2020 · Load a pre-trained model from disk with Huggingface Transformers Asked 5 years, 2 months ago Modified 2 years, 7 months ago Viewed 290k times

  5. Facing SSL Error with Huggingface pretrained models

    Mar 31, 2022 · huggingface.co now has a bad SSL certificate, your lib internally tries to verify it and fails. By adding the env variable, you basically disabled the SSL verification.

  6. How to do Tokenizer Batch processing? - HuggingFace

    Jun 7, 2023 · 9 in the Tokenizer documentation from huggingface, the call fuction accepts List [List [str]] and says: text (str, List [str], List [List [str]], optional) — The sequence or batch of sequences to be …

  7. Where does hugging face's transformers save models?

    Update 2023-05-02: The cache location has changed again, and is now ~/.cache/huggingface/hub/, as reported by @Victor Yan. Notably, the sub folders in the hub/ directory are also named similar to the …

  8. Huggingface: How do I find the max length of a model?

    Jun 24, 2023 · Given a transformer model on huggingface, how do I find the maximum input sequence length? For example, here I want to truncate to the max_length of the model: tokenizer (examples …

  9. Offline using cached models from huggingface pretrained

    Nov 9, 2023 · HuggingFace includes a caching mechanism. Whenever you load a model, a tokenizer, or a dataset, the files are downloaded and kept in a local cache for further utilization.

  10. How can I download a HuggingFace dataset via HuggingFace CLI while ...

    Apr 5, 2024 · I downloaded a dataset hosted on HuggingFace via the HuggingFace CLI as follows: pip install huggingface_hub [hf_transfer] huggingface-cli download huuuyeah/MeetingBank_Audio --repo …