ML | huggingface

Conda env

1
2
3
4
5
6
7
8
9
# check CUDA Version:
$ nvidia-smi

micromamba env create -n huggingface
micromamba activate huggingface
micromamba install python==3.10
micromamba install pytorch torchvision torchaudio pytorch-cuda=12.4 -c pytorch -c nvidia
micromamba install conda-forge::transformers
micromamba install git-lfs

Where does hugging face's transformers save models?

The cache location has changed again, and is now ~/.cache/huggingface/hub/

1
2
3
4
5
6
7
8
9
10
$ du -sh ~/.cache/huggingface/hub/*
6.3G /home/wpsze/.cache/huggingface/hub/models--deepseek-ai--deepseek-vl2-tiny
414M /home/wpsze/.cache/huggingface/hub/models--dslim--bert-base-NER
961M /home/wpsze/.cache/huggingface/hub/models--ecmwf--aifs-single
1.7G /home/wpsze/.cache/huggingface/hub/models--microsoft--DialoGPT-medium
172K /home/wpsze/.cache/huggingface/hub/models--microsoft--Florence-2-large
6.7G /home/wpsze/.cache/huggingface/hub/models--stabilityai--stable-diffusion-xl-base-1.0
4.4G /home/wpsze/.cache/huggingface/hub/models--stabilityai--stable-diffusion-xl-refiner-1.0
4.0K /home/wpsze/.cache/huggingface/hub/version.txt
4.0K /home/wpsze/.cache/huggingface/hub/version_diffusers_cache.txt

Example


ML | huggingface
https://waipangsze.github.io/2025/01/08/ML-huggingface-LLM/
Author
wpsze
Posted on
January 8, 2025
Updated on
January 20, 2025
Licensed under