EADST

Sharding and SafeTensors in Hugging Face Transformers

In the Hugging Face transformers library, managing large models efficiently is crucial, especially when working with limited disk space or specific file size requirements. Two key features that help with this are sharding and the use of SafeTensors.

Sharding

Sharding is the process of splitting a large model's weights into smaller files or "shards." This is particularly useful when dealing with large models that exceed file size limits or when you want to manage storage more effectively.

Usage

To shard a model during the saving process, you can use the max_shard_size parameter in the save_pretrained method. Here's an example:

# Save the model with sharding, setting the maximum shard size to 1GB
model.save_pretrained('./model_directory', max_shard_size="1GB")

In this example, the model's weights will be divided into multiple files, each not exceeding 1GB. This can make storage and transfer more manageable, especially when dealing with large-scale models.

SafeTensors

The safetensors library provides a new format for storing tensors in a safe and efficient way. Unlike traditional formats like PyTorch's .pt files, SafeTensors ensures that the tensor data cannot be accidentally executed as code, offering an additional layer of security. This is particularly important when sharing models across different systems or with the community.

Usage

To save a model using SafeTensors, simply specify the safe_serialization parameter when saving:

# Save the model using SafeTensors format
model.save_pretrained('./model_directory', safe_serialization=True)

This will create files with the .safetensors extension, ensuring the saved tensors are stored safely.

Combining Sharding and SafeTensors

You can combine both sharding and SafeTensors to save a large model securely and efficiently:

# Save the model with sharding and SafeTensors
model.save_pretrained('./model_directory', max_shard_size="1GB", safe_serialization=True)

This setup splits the model into shards, each in the SafeTensors format, offering both manageability and security.

Conclusion

By leveraging sharding and SafeTensors, Hugging Face transformers users can handle large models more effectively. Sharding helps manage file sizes, while SafeTensors ensures the safe storage of tensor data. These features are essential for anyone working with large-scale models, providing both practical and security benefits.

相关标签
About Me
XD
Goals determine what you are going to be.
Category
标签云
Qwen2.5 FP32 Ubuntu Conda 财报 LoRA uwsgi 版权 Permission Django VGG-16 飞书 OpenAI 第一性原理 阿里云 LLAMA Python tar Review Password GoogLeNet Distillation Datetime Bipartite Claude Linux Sklearn Paper 域名 Translation Llama Quantization Miniforge Pandas FlashAttention Cloudreve 多线程 Michelin Diagram SQL Knowledge JSON AI Tensor Logo 证件照 DeepStream PDB Domain Jetson FP64 v0.dev Baidu BTC QWEN Tiktoken Pickle Zip Animate 多进程 BF16 FP8 Qwen Hilton SAM Transformers VSCode Docker Nginx Freesound Food 搞笑 Land GPTQ Math Clash HuggingFace Gemma Jupyter Dataset XGBoost WAN DeepSeek Random Video YOLO printf Plotly Quantize HaggingFace Search Disk SPIE Tracking 算法题 Web SQLite logger Interview Website llama.cpp EXCEL XML Use Base64 OpenCV Firewall git Vim CTC GGML Hotel Anaconda Algorithm Excel CUDA PyCharm TensorRT transformers WebCrawler Github SVR ChatGPT Bin Color Agent scipy Git Template 腾讯云 LLM LaTeX 签证 顶会 diffusers RAR Pillow C++ Windows FastAPI GIT TSV 报税 CEIR Pytorch Input Safetensors Paddle 图形思考法 Magnet NameSilo Numpy RGB PDF FP16 音频 GPT4 BeautifulSoup API Qwen2 MD5 OCR Streamlit 公式 PIP Data TTS 强化学习 继承 COCO Augmentation Plate Bert Statistics Vmess ResNet-50 PyTorch Mixtral git-lfs Ptyhon Hungarian Proxy 关于博主 Heatmap Crawler CV LeetCode mmap Google Breakpoint IndexTTS2 v2ray Bitcoin ModelScope Card tqdm UI TensorFlow hf CLAP Markdown NLP CSV InvalidArgumentError uWSGI News CAM Attention 净利润 Shortcut torchinfo CC VPN UNIX ONNX 递归学习法 NLTK Image2Text
站点统计

本站现有博文320篇,共被浏览756549

本站已经建立2420天!

热门文章
文章归档
回到顶部