Skip to content

GeeeekExplorer/transformers-patch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

13 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Transformers-Patch πŸ› οΈ

Memory optimization patches for HuggingFace Transformers.

Features ✨

  • Memory Reduction - Significantly lowers memory usage in Transformers models

  • Zero Configuration - Works automatically after import

Installation ⚑

pip install git+https://github.com/GeeeekExplorer/transformers-patch.git

Quick Start πŸš€

Just import the patch before loading any Transformers models:

import transformers_patch
from transformers import AutoModel

Benchmark πŸ“Š

Test Configuration:

  • 8x GPU machine
  • Micro batch size: 1
  • Sequence length: 4096
  • Gradient checkpointing: Disabled
  • Model: Qwen3-8B
Memory Component Fixed Allocation Before Patch After Patch
Model + Gradients 30.5 GB - -
ZeRO Optimizer States 11.4 GB - -
Activations - 35.4 GB 17.8 GB

50% reduction in activation memory!

Example Usage πŸ“‹

See complete example in train.py.

Acknowledgements πŸ™

About

patches for huggingface transformers to save memory

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages