| transformers==4.45.2 | |
| datasets==2.19.2 | |
| transformer-tricks>=0.2.3 | |
| accelerate>=0.9.0 | |
| safetensors>=0.4.5 | |
| huggingface-hub>=0.25.1 | |
| torch>=2.4.1 | |
| tqdm>=4.64.1 | |
| jupytext>=1.16.4 | |
| autopep8>=2.3.1 | |
| twine>=5.1.1 | |
| build>=1.2.2 | |
| # pip list # see all versions | |
| # | |
| # Phi-3 needs flash-attn, but this requires CUDA | |
| # flash-attn==2.5.8 | |